We’re always looking for the ones
truly passionate about their work. If you are amongst them, you can rest assured there is a place for you in eMAG. We’ve grown very fast and are determined to keep doing so. What brought us here is our desire for continuous evolution and practical results.
More than 6000 colleagues are part of
eMAG Teams . We strongly believe in people's development and therefore every year we invest more and more energy and resources to remain an organization that is constantly learning. We want to ensure that you’ll have the most talented colleagues and the proper environment to grow and achieve great results, to become what you desire on a personal and professional level.
Join us, grow faster !
Big Data Engineer (Data&AI, Data Platform Team)
The
Data Platform team at eMAG empowers engineers, data scientists, and everyone else in the company with massive amounts of data and state-of-the-art data infrastructure. We leverage open-source technologies like Kafka, Hadoop, Spark, and other in-house tools to provide the capability to ingest, process, analyze, and store large-scale datasets. We are migrating towards a hybrid cloud environment, leveraging AWS services alongside our existing on-premise infrastructure.
As a
Big Data Engineer, you will play a critical role in designing, building, and optimizing our data platform. You will focus on ensuring the performance, reliability, and scalability of our platform, enabling data engineers and data scientists to build and deploy efficient data pipelines. You will contribute to our cloud migration strategy, selecting and integrating appropriate AWS services.
What You’ll Have To Do
- Performance Tuning and Optimization: Deeply analyze and optimize the performance of the Cloudera on-premises platform and AWS cloud services used for big data processing. This includes tuning Hadoop/Spark configurations, database performance (Impala, Hive, etc.), query optimization as well as resource contention analysis and resolution;
- Platform Reliability and Stability: Ensure the high availability and stability of the big data platform. Implement proactive monitoring, alerting, and incident response procedures;
- Capacity Planning and Management: Forecast future platform capacity needs based on data growth and usage patterns. Plan and implement capacity upgrades, ensuring scalability and cost-effectiveness. Manage resource allocation across the platform;
- Cloud Migration and Integration: Contribute to the gradual migration of the big data platform to AWS. Ensure seamless integration between the on-premise Cloudera platform and AWS cloud services. Evaluate and select appropriate AWS services (e.g., EMR, Glue, Athena, Redshift) and design hybrid cloud solutions;
- Technology Evaluation and Research: Stay up-to-date with the latest big data technologies and trends. Evaluate new tools and techniques for improving platform performance, reliability, and scalability. Conduct proof-of-concept projects to assess the feasibility of new technologies;
- Collaboration and Knowledge Sharing: Work closely with data engineers, data scientists, and other stakeholders to understand their platform requirements and provide technical guidance. Share knowledge and best practices with the team;
- Documentation: Create and maintain comprehensive documentation for the big data platform, including architecture diagrams, configuration settings, and operational procedures.
What Makes You a Good Fit
- Have 5+ years of full-time, relevant industry experience;
- Working experience with data at the petabyte scale;
- Design and operation of large-scale data ingestion or any other distributed systems;
- Have experience with open-source technologies like Kafka, Hadoop, Hive, Airflow and Spark;
- Have experience with column-oriented data files like Parquet or high-performance formats like Iceberg;
- Have experience with Java, Scala and/or Python is preferred;
- Familiar with profiling tools, performance testing frameworks, and query optimization techniques;
- Have bachelor’s or Master’s degree in Computer Science, Engineering, or a related field.
What Makes You Stand Out
- Have experience with AWS big data services (EMR, Glue, Athena, Redshift, S3);
- Experience with relational databases, such as SQL Server, Oracle or MySQL;
- DevOps skills: Kubernetes;
- Monitoring/Logging/Operations: Prometheus, Grafana.
What We’ve Prepared For You
- Medical subscription: Medicover, MedLife or Regina Maria.
- A flexible budget that you can invest in yourself as you wish: meal tickets, holiday tickets, cultural vouchers, private pension, foreign language classes, eMAG, Fashion Days, Therme & Genius, membership to different gyms or even professional development classes.
- Different discounts from our partners: banking, mobile, dental medicine or wellness.
- Access to the Bookster library and free credits on the Hilio psycho-emotional health platform.
- An accelerated learning environment, with access to over 100.000 curated online resources and platforms, learning academies and development programs.
- New headquarters, where sleek design, natural light, and versatile spaces create an energizing and comfortable environment for hybrid work.
Curious to find out more about the next step in your career?
Apply now and if your experience is relevant for the role you wish, we will give you a call for more details!