Desired Candidate Profile
Job Responsibilities :
- Responsible for Requirements gathering and assessment
- Breakdown complexity and translate requirements to specification artifacts and story boards to build towards, using a test-driven approach
- Engineer scalable data pipelines using big data technologies including but not limited to Hadoop, HDFS, Kafka, HBase, Elastic
- Implement the pipelines using execution frameworks including but not limited to MapReduce, Spark, Hive, using Java/Scala/Python for application design.
- Mentoring juniors in a dynamic team setting
- Manage stakeholders with proactive communication upholding our brand and values
Must Have The Following Skills :
- Strong problem solving ability
- Excellent software design and implementation ability
- Exposure and commitment to agile methodologies
- Detail oriented with willingness to proactively own software tasks as well as management tasks, and see them to completion with minimal guidance
- Minimum 7 years of experience
- Should have experience in full life-cycle of one big data application
- Strong understanding of various storage formats (ORC/Parquet/Avro)
- Should have hands on experience in one of the Hadoop distributions (Hortoworks/Cloudera/MapR)
- Should be well versed with at least one database (Mysql/Oracle/MongoDB/Postgres)
- Should have good code review and debugging skills
Should have experience in azure
Additional skills(Good to have) :
- Experience in Containerization (docker/Heroku)
- Exposure to microservices
- Exposure to DevOps practices
- Experience in Performance tuning of big data applications
Education:
UG: B.Tech/B.E. - Computers
PG: M.Tech - Computers, MCA - Computers, MS/M.Sc(Science) - Computers
Doctorate: Doctorate Not Required
Contact Details:
Keyskills:
Hadoop
Hive
Java
Spark
Mapreduce
Cloudera
SCALA
Hdfs
MongoDB
MySQL