Pyspark Data Engineer

Full Time | PAN India | India

Industry : Information Technology and Services
Experience5 - 12 years
Compensation700,000 - 3,200,000
Openings2

Role Overview

Designation / Role1 Development in a programming language on GCP

Desired Competencies (Technical/Behavioral Competency)

Must-Have · Good Hands on knowledge on GCP · Should have worked on Data Migration projects, from On-prem to Cloud · Should have Cloud Storage Knowledge, Big Query, Cluster Knowledge · Sound programming knowledge on PySpark & SQL in terms of processing large amount of semi structured & unstructured data · Ability to design data pipelines in end to end manner · Knowledge on Avro, Parquet format · Knowledge on working on Hadoop Big Data platform and ecosystem · Strong debugging and troubleshooting capabilities. · Have experience to guide the Technical Team for attaining the delivery milestones

Good-to-Have 1. Knowledge on Jira, Agile, Sonar, Team city & CICD 2. Any exposure / experience for an international Banking client / multi-vendor / multi geography teams 3. Knowledge on Dataproc, Pyspark

 

SN Responsibility of / Expectations from the Role

1 Development in a programming language on GCP

2 Testing & Validation

3 Debugging & troubleshootin

Skill Set

GCP pyspark hadoop python hdfs sql
Application

Apply for this role

PDF, DOC, or DOCX up to your server upload limit.