Job Title: ETL Developer – Python & SQL (On-Premise Data Pipeline)
Job Summary:
We are seeking a skilled ETL Developer to design, develop, and maintain on-premise data pipelines. The ideal candidate will have strong experience in Python programming and SQL, with a deep understanding of data extraction, transformation, and loading (ETL) processes. This role will focus on integrating data from multiple sources and ensuring data quality, performance, and reliability without relying on cloud platforms.
Key Responsibilities:
· Design and implement scalable ETL pipelines using Python and SQL
· Develop connectors/scripts to ingest data from structured and semi-structured sources
· Optimize queries and data workflows for performance and reliability
· Ensure data validation, quality checks, and logging across the pipeline
· Collaborate with business analysts and data stakeholders to understand requirements
· Maintain documentation for ETL workflows, schemas, and dependencies
Required Skills:
· Proficiency in Python (with experience using libraries like Pandas, SQLAlchemy, etc.)
· Strong SQL skills (preferably on platforms like PostgreSQL, MySQL, or Oracle)
· Experience building ETL pipelines without cloud-based tools
· Knowledge of data integration, transformation logic, and performance tuning
· Familiarity with scheduling tools (e.g., cron, Airflow – optional if run locally)
· Good understanding of file formats like CSV, JSON, XML
Preferred Qualifications:
· Experience in shell scripting and working in Linux environments
· Exposure to data warehousing concepts
· Knowledge of version control tools (e.g., Git)