Hyderabad Jobs |
Banglore Jobs |
Chennai Jobs |
Delhi Jobs |
Ahmedabad Jobs |
Mumbai Jobs |
Pune Jobs |
Vijayawada Jobs |
Gurgaon Jobs |
Noida Jobs |
Hyderabad Jobs |
Banglore Jobs |
Chennai Jobs |
Delhi Jobs |
Ahmedabad Jobs |
Mumbai Jobs |
Pune Jobs |
Vijayawada Jobs |
Gurgaon Jobs |
Noida Jobs |
Oil & Gas Jobs |
Banking Jobs |
Construction Jobs |
Top Management Jobs |
IT - Software Jobs |
Medical Healthcare Jobs |
Purchase / Logistics Jobs |
Sales |
Ajax Jobs |
Designing Jobs |
ASP .NET Jobs |
Java Jobs |
MySQL Jobs |
Sap hr Jobs |
Software Testing Jobs |
Html Jobs |
Job Location | Chennai |
Education | Not Mentioned |
Salary | Not Disclosed |
Industry | IT - Software |
Functional Area | DBA / DatawarehousingGeneral / Other Software |
EmploymentType | Full-time |
Data Engineer - ETL with Spark - Merit Group Data Engineer ETL with Spark DescriptionRole: Data Engineer Job Description We need someone with 3- 5 years of extensive experience in Data Warehousing, ETL and Big data technologies(Hadoop, Hive, Sqoop.etc) and 2+ years of mandatory experience in Spark with Python/ Scala with more than one end- to- end implementation experience. Roles and Responsibilities To develop Scala or Python scripts, UDFs using both Data frames/ SQL/ Data sets and RDD in Spark 2.3+ for Data Aggregation, queries and writing data back into the OLTP system through Sqoop. Should have a very good understanding of Partitions, Bucketing concepts and designed both Managed and external tables, ORC files in Hive to optimize performance. Wrote and Implemented Spark and Scala scripts to load data from and to store data into Cassandra/ Hbase/ any NoSQL Implementing SCD Type 1 and Type 2 model using Spark Developed Oozie workflow for scheduling and orchestrating the ETL process Experienced in performance tuning of Spark Applications for setting right Batch Interval time, the correct level of Parallelism and memory tuning Streaming data into Elastic search for visualization using Kibana Should have implemented the mapping parameters/ variables in the mapping and the session level to increase the reusability of the code and parameterize the hardcoded values. Additional skills: Knowledge in AWS stacks AWS Glue, S3, SQS Exposure to Elastic Search, Solr is a plus Exposure to NoSQL Databases Cassandra, MongoDB Exposure to Serverless computing Please enter the following information. Upload your CV Type: Full time Category: Non IT LEARNING & DEVELOPMENT Merit takes pride in offering extensive inhouse training programmes focused on developing employees for career progression including specific competence in technology and processes, as well as interpersonal skills. From intense finishing schools that shape the next generation of team leaders to internal training and placement programs that identify and train staff with potential in tech to help them reach their potential, we are committed to bringing the best out of the individuals that make up Merit.
Keyskills :
oltp python oozie hive solr orc glue java sql etl aws spark informatica lasticsearch datawarehousing bigdata performancetuning dataaggregation