Total IT / development experience of 3+ years
- Experience in Spark (Scala-Spark or PySpark) developing Big Data applications on Hadoop, Hive and/or Kafka, HBase, MongoDB+B4+A3:B5d technology strategies
- Exposure to deploying on Cloud platforms
- At least 2 years of development experience on designing and developing Data Pipelines for Data Ingestion or Transformation using Spark-Scala/PySpark
- At least 2 years of development experience in the following Big Data frameworks: File Format (Parquet, AVRO, ORC), Resource Management, Distributed Processing and RDBMS
- At least 2 years of developing applications in Agile with Monitoring, Build Tools, Version Control, Unit Test, Unix Shell Scripting, TDD, CI/CD, Change Management to support DevOps
GOOD-TO-HAVE
- Banking domain knowledge
- Hands-on experience in SAS toolset / statistical modelling migrating to Machine Learning models
- Banking Risk, Fraud or Digital Marketing Machine Learning models and use cases
- ETL / Data Warehousing, SQL and Data Modelling experience prior to Big Data experience Location Chennai / Bangalore / Pune Experience 4-7 Yrs
Skills: pyspark,ci/cd,machine learning,mongodb,data pipelines,unit test,data,unix shell scripting,hadoop,hive,kafka,learning,agile,spark,management,scala-spark,change management,machine learning models,a3,hbase,tdd,scala,big data