Are you ready to join a cutting-edge Digital Solution Company and help to shape the future of business IT solutions?
Our client is a leading global provider of IT solutions and services, known for their customer-centric approach to digital transformation. With a rich history dating back to 1996, they have continually evolved to meet the changing needs of their customers. Their services encompass consulting, technology, and outsourcing, delivering innovative solutions to complex challenges. They have also been honored multiple times as a top employer, including being named a Great Place To Work from 2015 to 2024.
You will be working on a project for a major international company operating in the energy and industrial engineering sector.
Your future responsibilities will include:
Design and maintain ETL pipelines using AWS Glue and PySpark, ensuring code standardization and scalability across workflows
Collect, process, and integrate time series, structured, and unstructured data (e.g., images).
Standardize code across workflows and validate data accuracy through unit testing.
Implement layered data architecture in AWS and Snowflake.
Create and maintain mapping documentation for processes and data assets.
Develop additional views and logic to support reporting and analytics needs.
Support deployment to QA/UAT and production environments using CI/CD pipelines.
Requirements:
5+ years of experience in data engineering, focusing on data collection, processing, and integration.
Strong proficiency in ETL tools and processes, particularly AWS Glue and PySpark.