Company Overview
Viraaj HR Solutions is dedicated to delivering top-tier HR services and talent acquisition strategies to help companies throughout India thrive. Our mission is to connect skilled professionals with excellent opportunities, fostering a culture of collaboration, integrity, and innovation. We pride ourselves on understanding the unique needs of both our clients and candidates, ensuring a perfect fit for every role. At Viraaj HR Solutions, we prioritize our people's growth and development, making us a dynamic and rewarding workplace.
Role Responsibilities
- Design and implement scalable Big Data solutions using Hadoop technologies.
- Develop and maintain ETL processes to manage and process large data sets.
- Collaborate with data architects and analysts to gather requirements and deliver solutions.
- Optimize existing Hadoop applications for maximizing efficiency and performance.
- Write, test, and maintain complex SQL queries to extract and manipulate data.
- Implement data models and strategies that accommodate large-scale data processing.
- Conduct data profiling and analysis to ensure data integrity and accuracy.
- Utilize MapReduce frameworks to execute data processing tasks.
- Work closely with data scientists to facilitate exploratory data analysis.
- Ensure compliance with data governance and privacy regulations.
- Participate in code reviews and maintain documentation for all development processes.
- Troubleshoot and resolve performance bottlenecks and other technical issues.
- Stay current with technology trends and tools in Big Data and Cloud platforms.
- Train junior developers and assist in their professional development.
- Contribute to team meeting discussions regarding project status and ideas for improvement.
Qualifications
- Bachelor’s degree in Computer Science, Information Technology, or related field.
- Proven experience as a Big Data Developer or similar role.
- Strong foundational knowledge of the Hadoop ecosystem (HDFS, Hive, Pig, etc.).
- Proficient in programming languages such as Java, Python, or Scala.
- Experience with database management systems (SQL and NoSQL).
- Familiar with data processing frameworks like Apache Spark.
- Understanding of data pipeline architectures and data integration techniques.
- Knowledge of cloud computing services (AWS, Azure, or Google Cloud).
- Exceptional problem-solving skills and attention to detail.
- Strong communication skills and ability to work in a team environment.
- Experience with data visualization tools (Tableau, Power BI, etc.) is a plus.
- Ability to work under pressure and meet tight deadlines.
- Adaptability to new technologies and platforms as they emerge.
- Certifications in Big Data technologies would be an advantage.
- Willingness to learn and grow in the field of data sciences.
Skills: data visualization,spark,java,cloud computing (aws, azure, google cloud),sql proficiency,apache spark,data warehousing,data visualization tools,nosql,scala,python,etl,sql,hadoop,mapreduce,big data analytics,big data