BigData and Hadoop Developer

Hyderabad, TS, IN

6 days ago
Save Job

Summary

Company Overview

Viraaj HR Solutions is dedicated to delivering top-tier HR services and talent acquisition strategies to help companies throughout India thrive. Our mission is to connect skilled professionals with excellent opportunities, fostering a culture of collaboration, integrity, and innovation. We pride ourselves on understanding the unique needs of both our clients and candidates, ensuring a perfect fit for every role. At Viraaj HR Solutions, we prioritize our people's growth and development, making us a dynamic and rewarding workplace.

Role Responsibilities

  • Design and implement scalable Big Data solutions using Hadoop technologies.
  • Develop and maintain ETL processes to manage and process large data sets.
  • Collaborate with data architects and analysts to gather requirements and deliver solutions.
  • Optimize existing Hadoop applications for maximizing efficiency and performance.
  • Write, test, and maintain complex SQL queries to extract and manipulate data.
  • Implement data models and strategies that accommodate large-scale data processing.
  • Conduct data profiling and analysis to ensure data integrity and accuracy.
  • Utilize MapReduce frameworks to execute data processing tasks.
  • Work closely with data scientists to facilitate exploratory data analysis.
  • Ensure compliance with data governance and privacy regulations.
  • Participate in code reviews and maintain documentation for all development processes.
  • Troubleshoot and resolve performance bottlenecks and other technical issues.
  • Stay current with technology trends and tools in Big Data and Cloud platforms.
  • Train junior developers and assist in their professional development.
  • Contribute to team meeting discussions regarding project status and ideas for improvement.

Qualifications

  • Bachelor’s degree in Computer Science, Information Technology, or related field.
  • Proven experience as a Big Data Developer or similar role.
  • Strong foundational knowledge of the Hadoop ecosystem (HDFS, Hive, Pig, etc.).
  • Proficient in programming languages such as Java, Python, or Scala.
  • Experience with database management systems (SQL and NoSQL).
  • Familiar with data processing frameworks like Apache Spark.
  • Understanding of data pipeline architectures and data integration techniques.
  • Knowledge of cloud computing services (AWS, Azure, or Google Cloud).
  • Exceptional problem-solving skills and attention to detail.
  • Strong communication skills and ability to work in a team environment.
  • Experience with data visualization tools (Tableau, Power BI, etc.) is a plus.
  • Ability to work under pressure and meet tight deadlines.
  • Adaptability to new technologies and platforms as they emerge.
  • Certifications in Big Data technologies would be an advantage.
  • Willingness to learn and grow in the field of data sciences.

Skills: data visualization,spark,java,cloud computing (aws, azure, google cloud),sql proficiency,apache spark,data warehousing,data visualization tools,nosql,scala,python,etl,sql,hadoop,mapreduce,big data analytics,big data

How strong is your resume?

Upload your resume and get feedback from our expert to help land this job