We strongly believe the future belongs to those who know how to use data. Understanding and innovating with data has the potential to change the way we do almost anything for the better.
Our aim is to become a Centre of Excellence for Big Data & Analytics, Cloud solutions, Internet of Things and Business Intelligence, enabling our customers to use the best technologies and tools to accelerate their business, shorten their time to market, be agile and innovative.
At Quark Technologies, we value open-minded people who take pride in their work and want to make a difference. If you share our passion for big data and want to be part of a growing team where you can grow both professionally and personally, please come talk to us.
Roles & Responsibilities
Develop big data solutions for near real-time stream processing, as well as batch processing on the Big Data platform
Work with business domain experts, data scientists and application developers to identify data that is relevant for analysis and develop a scalable solution
Keep up to date with new developments in the big data ecosystem and learn new technologies
Ensure on-time quality deliverable with ability to coordinate effectively with the rest of the team
Requirements
Good understanding of (one or more): system design, distributed systems, database systems, cloud infrastructure, security, performance optimizations
Ability to create simple, elegant design
Able to pick up a new language or framework in very short time
Qualifications
Love data and care about data security, correctness, and completeness
Excellent problem analysis and solving skills
Excellent communication skills and team player
Strong foundation in algorithms and data structures and their real world use cases
Production quality coding standards and patterns
Passion for developing amazing customer experiences
4+ years of general software programming experience; at least one year of industry experience
Bachelor degree in Computer Science, Math, related technical field or equivalent practical experience
Pluses
4+ years of modern, production level Python/Scala/JavaScript
Experience with large scale distributed computing, API design for distributed services
Experience with developing and maintaining production data pipelines, Kafka, Apache Spark, etc.
Experience with Spark Streaming, Spark SQL, Pyspark in a production setting
Knowledge of SQL and conventional databases
Experience in NOSQL technologies like Hbase, Cassandra, MongoDB
Experience with data preparation, transformation and analytical model development
Understanding of cloud, and distributed systems principles, including load balancing, networks, scaling, in-memory vs. Disk
Experience with cloud services like Lambda and Kinesis
Bash/Shell scripting
Continuous Integration / Continuous Deployment