Job Description
We are Hiring Data Engineer | Datastore Migration (AWS Lakehouse) at Coforge Ltd.
Experience: 4 to 6 Years
Employment Type: Full-time
Work Location: Hyderabad & Bangalore (Only)
Apply Now: Ga***********r@co****e.com
WhatsApp: 9667427***
Role Overview
We are seeking a skilled Data Engineer to join our Datastore Migration Factory Team, responsible for the end-to-end migration of enterprise data platforms from On-Prem DataLake environments to AWS-hosted Lakehouse architectures.
The ideal candidate will have strong hands-on data engineering experience, a sound understanding of modern data platforms, and the ability to work closely with stakeholders to deliver reliable, production-grade data solutions.
Key Responsibilities:-
1. Pipeline Migration
- Refactor and migrate extraction logic and scheduling from legacy frameworks to modern Lakehouse architectures.
- Execute physical migration of large-scale datasets while ensuring data accuracy, integrity, and consistency.
- Act as a technical liaison to internal stakeholders, enabling smooth hand-off and formal sign-off with data owners.
2. Consumption Pattern Migration
- Convert and optimize legacy SQL and Spark-based data consumption patterns (raw and curated layers) for Snowflake and Apache Iceberg.
- Analyze usage patterns to design and deliver fit-for-purpose data products.
- Perform data reconciliation and validation to ensure migrated datasets meet business and functional expectations.
- Engage stakeholders for validation, sign-off, and post-migration assurance.
3. Data Validation & Quality Assurance
- Apply a rigorous data reconciliation framework to validate functional equivalency between legacy and migrated datasets.
- Ensure high standards of data quality, reliability, and production readiness.
4. Collaboration & Platform Enablement
- Work closely with internal data platform and data management teams.
- Demonstrate the ability to quickly learn new workflows, tools, and language constructs as required.
Required Technical Skills:
Basic Qualifications:-
- Education: Bachelors or masters degree in computer science, Engineering, Applied Mathematics, or a related field.
- Experience: 4 to 6 years of hands-on development experience in a collaborative environment.
- Strong troubleshooting skills in SQL and experience with basic scripting.
- Programming Languages: Proficiency in Python or Java.
- Strong understanding of SDLC, CI/CD best practices, and exposure to Kubernetes (K8s)deployments.
Core Data Engineering Competencies:-
- Temporal Data Modeling: Experience handling historical data (e.g., SCD Type 2).
- Schema Management: Expertise in schema evolution and governance (Apache Iceberg).
- Performance Optimization: Advanced knowledge of partitioning, clustering, and query optimization.
- Data Architecture: Strong understanding of normalization vs. denormalization and natural vs. surrogate keys.
Technology Stack (Exposure Required):-
- Extraction & Processing: Kafka, Apache Spark, ANSI SQL, FTP
- Data Formats: JSON, Avro, Parquet
- Platforms: Hadoop (HDFS/Hive), Snowflake, Apache Iceberg, Sybase IQ
Core Competencies:-
- Demonstrates integrity, ethical conduct, and professional accountability.
- Strong team player with the ability to collaborate across cross-functional and global teams.
- Clear, confident communicator with strong stakeholder management skills.
- Delivery-oriented mindset with a strong sense of ownership and commitment.
- High energy, proactive approach, and results-driven attitude.
- Curious, analytical thinker who identifies risks early and drives continuous improvement.
Job Classification
Industry: IT Services & Consulting
Functional Area / Department: Data Science & Analytics
Role Category: Data Science & Analytics - Other
Role: Data Science & Analytics - Other
Employement Type: Full time
Contact Details:
Company: Coforge
Location(s): Hyderabad
Keyskills:
SQL
AWS
Python
Kafka
Lamda
Ci/Cd
EMR
Aws Glue
Devops
Hive
S
Snowflake
Datastore Migration
Data Lake
Spark