Job Description
We specialize in delivering high-quality human-curated data and AI-first scaled operations servicesBased in San Francisco and Hyderabad, we are a fast-moving team on a mission to build AI for Good, driving innovation and societal impactRole Overview:We are seeking a Data Engineer / Data Architect who will be responsible for designing, building, and maintaining scalable data infrastructure and systems for a clientYoull play a key role in enabling efficient data flow, storage, transformation, and access across our organization or client ecosystemsWhether youre just beginning or already an expert, we value strong technical skills, curiosity, and the ability to translate complex requirements into reliable data pipelinesResponsibilities:Design and implement scalable, robust, and secure data pipelinesBuild ETL/ELT frameworks to collect, clean, and transform structured and unstructured dataCollaborate with data scientists, analysts, and backend engineers to enable seamless data access and model integrationMaintain data integrity, schema design, lineage, and quality monitoringOptimize performance and ensure reliability of data workflows in production environmentsDesign and manage data warehousing and lakehouse architectureSet up and manage infrastructure using IaC (Infrastructure as Code) when applicableRequired Skills:Strong programming skills in Python, SQL, and Shell scriptingHands-on experience with ETL tools and orchestration frameworks (e g, Airflow, Luigi, dbt)Proficiency in relational databases (e g , PostgreSQL, MySQL) and NoSQL databases (e g
, MongoDB, Redis)Experience with big data technologies: Apache Spark, Kafka, Hive, Hadoop, etcDeep understanding of data modeling, schema design, and data warehousing conceptsProficient with cloud platforms (AWS/GCP/Azure) and services like Redshift, BigQuery, S3, Dataflow, or DatabricksKnowledge of DevOps and CI/CD tools relevant to data infrastructureNice to Have:Experience working in real-time streaming environmentsFamiliarity with containerization and KubernetesExposure to MLOps and collaboration with ML teamsExperience with security protocols, data governance, and compliance frameworksEducational Qualifications:Bachelors or Masters in Computer Science, Data Engineering, Information Systems, or a related technical field
Location - Mumbai, Delhi / NCR, Bengaluru , Kolkata, Chennai, Hyderabad, Ahmedabad, Pune, India
Job Classification
Industry: IT Services & Consulting
Functional Area / Department: Engineering - Software & QA
Role Category: DBA / Data warehousing
Role: Data warehouse Architect / Consultant
Employement Type: Full time
Contact Details:
Company: Soul Ai
Location(s): Mumbai
Keyskills:
Data Engineering
Hive
PostgreSQL
Hadoop
MySQL
Kafka
Apache Spark