Design, develop, and optimize ETL workflows and data pipelines for extracting, transforming, and loading data from various sources into data warehouses or data lakes.
Ensure seamless integration of data from multiple sources (e.g., databases, APIs, cloud platforms) and maintain data quality.
Perform data cleansing, aggregation, and normalization to ensure consistent and accurate reporting.
Monitor and improve ETL process performance by optimizing SQL queries, indexing, and resource utilization.
Collaborate with data architects to design and implement star/snowflake schemas, dimensional models, and other data warehouse structures.
Implement robust error handling, logging, and notification mechanisms for ETL processes.
Create and maintain detailed technical documentation for ETL processes, data flows, and data models.
Ensure ETL processes adhere to company security policies and data governance standards.
Required Skills & Qualifications:
8+ years of experience as an ETL Developer or in a similar role.
Strong proficiency with ETL tools (e.g., Informatica, Talend, SSIS, Apache Nifi, or AWS Glue).
Expert-level SQL skills and experience with database platforms (e.g., Oracle, SQL Server, PostgreSQL, MySQL).
Hands-on experience with data warehousing concepts and best practices.
Strong understanding of data quality, data governance, and data integration principles.
Experience with cloud-based data platforms (AWS, Azure, or Google Cloud).
Knowledge of scripting languages (e.g., Python, Shell, or Bash).
Excellent problem-solving and analytical skills.
Strong communication and collaboration skills.
Preferred Skills:
Experience with big data technologies (e.g., Hadoop, Spark, Databricks).
Familiarity with real-time data streaming and integration tools (e.g., Kafka, AWS Kinesis).
Knowledge of data visualization tools (e.g., Power BI, Tableau).
Exposure to DevOps and CI/CD practices for ETL pipelines.