Role Proficiency:
This role requires proficiency in data pipeline development including coding and testing data pipelines for ingesting wrangling transforming and joining data from various sources. Must be skilled in ETL tools such as Informatica Glue Databricks and DataProc with coding expertise in Python PySpark and SQL. Works independently and has a deep understanding of data warehousing solutions including Snowflake BigQuery Lakehouse and Delta Lake. Capable of calculating costs and understanding performance issues related to data solutions.
Outcomes:
Measures of Outcomes:
Outputs Expected:
Code Development:
Documentation:
Configuration:
Testing:
Domain Relevance:
Project Management:
Defect Management:
Estimation:
Knowledge Management:
Release Management:
Design Contribution:
Customer Interface:
Team Management:
Certifications:
Skill Examples:
Knowledge Examples:
Knowledge Examples
Additional Comments:
Role(s): Data Engineer Role Location(s): India Planned Start Date: 11/3/2025 Role Scope / Deliverables: Create new data pipelines in DataBricks. Support existing data pipelines in DataBricks. Create DAG setup in in airflow. Resolve day to day job failures/performance issues. Key Skills: Minimum 4+ years of total experience. Proficient in Databricks, Pyspark/Python (minimum 3+ year of experience) Proficient in SQL (minimum 3+ year of experience) Hands on experience in could AWS/Azure (preferably AWS) Good to have analytical experience Excellent communication. Should be able to work independently.
Databricks,Pyspark,SQL,Python

Keyskills: glue pyspark data warehousing data pipeline sql analytics apache etl tool gcp awsazure design bigquery etl programming communication skills snowflake python development airflow talend microsoft azure dataproc nosql data bricks aws glue optimization techniques aws informatica