-Candidate will understand business data in source systems and Migrate the same as-is to Big data lake, without losing the value. -Create a transient layer in Hive, Perform fundamental structural & symantic transformations needed for the data to be complete and consumable -Design the Hub layer with the right structure based on the business needs / demands, and Load the data from transient to Hub layer. -Migrate the existing ETL transformation scripts in traditional datawarehouse to DataLake using HiveQL -Merge and optimize for high performance in datalake various scripts as required for performance and ease of management -Build data pipelines to output the final datasets and load it back to DataWarehouse for reporting layer to consume using Scoop -Schedule and automate these data pipelines in Oozie