hireejobs
Hyderabad Jobs
Banglore Jobs
Chennai Jobs
Delhi Jobs
Ahmedabad Jobs
Mumbai Jobs
Pune Jobs
Vijayawada Jobs
Gurgaon Jobs
Noida Jobs
Oil & Gas Jobs
Banking Jobs
Construction Jobs
Top Management Jobs
IT - Software Jobs
Medical Healthcare Jobs
Purchase / Logistics Jobs
Sales
Ajax Jobs
Designing Jobs
ASP .NET Jobs
Java Jobs
MySQL Jobs
Sap hr Jobs
Software Testing Jobs
Html Jobs
IT Jobs
Logistics Jobs
Customer Service Jobs
Airport Jobs
Banking Jobs
Driver Jobs
Part Time Jobs
Civil Engineering Jobs
Accountant Jobs
Safety Officer Jobs
Nursing Jobs
Civil Engineering Jobs
Hospitality Jobs
Part Time Jobs
Security Jobs
Finance Jobs
Marketing Jobs
Shipping Jobs
Real Estate Jobs
Telecom Jobs

Data Engineer

3.00 to 5.00 Years   Pune   01 Dec, 2022
Job LocationPune
EducationNot Mentioned
SalaryRs 16 - 28 Lakh/Yr
IndustryIT - Software
Functional AreaDBA / Datawarehousing
EmploymentTypeFull-time

Job Description

    Position : Data Engineer (Python, Spark, Hive/ SQL)Experience : 3 to 5 yrs.Location : The candidates need to travel to either of the offices (either HYD/ Pune) for onboarding and stay for 1 week, post which they can work from home. Look for candidates around Pune.Notice Period : look for candidates who can join in15 days ONLYMust Have:
    • Software Engineer with experience (3+ years) in designing / implementing ETL / ELT processes for on-prem databases / cloud warehouses using combination of Python scripting (Spark), SQL scripting & Data Engineering Tools (Cloud services / Open-Source Tools)
    • Solid background in database concepts, ETL architectures / strategies / administration to be able to design compute-efficient data pipelines
    • Prior experience working with large scale data (big data) processing pipelines / engines using PySpark
    Good to Have but not necessary:
    • Ability to design pipelines for analytics / ML / AI workflows and deploy models / solutions to production
    • Familiarity / Experience in SWE best practices for unit-testing, code modularization, QA
    • Coursework / Past Projects / Github Repos illustrating familiarity with data warehousing best practices
    • Experience working with open-source pipeline orchestration frameworks like Airflow, Prefect, Luigi
    • Familiarity with application containerization tools like Docker, Kubernetes
    Job Responsibilities
    • Responsible for the design, development, and implementation of dataintegration processes using SQL / Python Scripting / Azure Data Flow / Other Cloud Platforms & Tools
    • Responsible for setting up end to end data pipeline which includes importing, cleaning, transforming, validating and analyzing data with the purpose of understanding or making conclusions from the data for data modeling, data integration and decision making purposes.
    • Collaborate with business users, analyze user requirements, translate & apply business rules to data transformations
    • Lead and mentor junior developers in information integration standards, best practices, etc.
    • Create functional & technical documentation e.g. data integration architecture flows, source to target mappings, ETL specification documents, run books, test plans
    • Performs data collection, profiling, validation, cleansing, analysis, and reporting. Tests, debugs, and documents ETL processes, SQL queries, and stored procedures
    • Initiates analysis to identify data inconsistencies and resolve in order to raise the level of data integrity
    • Analyzes data volumes, data types and content to support the design of data architecture solutions
    • Work with Data Warehouse Architects performing source system analysis, identification of key data issues, data profiling and development of normalized and star-snowflake physical schemas

Keyskills :
pythonsql

Data Engineer Related Jobs

© 2019 Hireejobs All Rights Reserved