We are seeking a highly skilled and motivated Data Engineer with expertise in Apache Spark and Hadoop eco systems. The ideal candidate must have strong programming skills in Java or Python and possess a deep understanding of distributed data processing, pipeline development, and big data technologies.
Required Skills:
Big Data Frameworks: Extensive experience in PySpark , Spark-Scala , Hadoop , Hive , and HBase .
Programming: Proficiency in Python , SQL, and Linux shell scripting; familiarity with Scala is a plus.
Cloud Services: In-depth experience with AWS S3 , EMR , RDS , and MWAA (Airflow) .
Data Storage and Integration: Strong knowledge of Oracle, Redshift, and other relational and NoSQL databases.
Data Processing: Expertise in handling diverse file formats such as CSV, Parquet, JSON, and Avro.
Job Classification
Industry: IT Services & Consulting Functional Area / Department: Engineering - Software & QA Role Category: Software Development Role: Software Development - Other Employement Type: Full time