Quark Technologies

Big Data Engineer - Quark Technologies - Cluj-Napoca

Romania

almost 5 years ago
Save Job

Summary

We strongly believe the future belongs to those who know how to use data. Understanding and innovating with data has the potential to change the way we do almost anything for the better. Our aim is to become a Centre of Excellence for Big Data & Analytics, Cloud solutions, Internet of Things and Business Intelligence, enabling our customers to use the best technologies and tools to accelerate their business, shorten their time to market, be agile and innovative. At Quark Technologies, we value open-minded people who take pride in their work and want to make a difference. If you share our passion for big data and want to be part of a growing team where you can grow both professionally and personally, please come talk to us. Roles & Responsibilities Develop big data solutions for near real-time stream processing, as well as batch processing on the Big Data platform Work with business domain experts, data scientists and application developers to identify data that is relevant for analysis and develop a scalable solution Keep up to date with new developments in the big data ecosystem and learn new technologies Ensure on-time quality deliverable with ability to coordinate effectively with the rest of the team Requirements Good understanding of (one or more): system design, distributed systems, database systems, cloud infrastructure, security, performance optimizations Ability to create simple, elegant design Able to pick up a new language or framework in very short time Qualifications Love data and care about data security, correctness, and completeness Excellent problem analysis and solving skills Excellent communication skills and team player Strong foundation in algorithms and data structures and their real world use cases Production quality coding standards and patterns Passion for developing amazing customer experiences 4+ years of general software programming experience; at least one year of industry experience Bachelor degree in Computer Science, Math, related technical field or equivalent practical experience Pluses 4+ years of modern, production level Python/Scala/JavaScript Experience with large scale distributed computing, API design for distributed services Experience with developing and maintaining production data pipelines, Kafka, Apache Spark, etc. Experience with Spark Streaming, Spark SQL, Pyspark in a production setting Knowledge of SQL and conventional databases Experience in NOSQL technologies like Hbase, Cassandra, MongoDB Experience with data preparation, transformation and analytical model development Understanding of cloud, and distributed systems principles, including load balancing, networks, scaling, in-memory vs. Disk Experience with cloud services like Lambda and Kinesis Bash/Shell scripting Continuous Integration / Continuous Deployment

How strong is your resume?

Upload your resume and get feedback from our expert to help land this job