Pyspark Data Engineer
Full Time | PAN India | India
Role Overview
Desired Competencies (Technical/Behavioral Competency)
Must-Have · Good Hands on knowledge on GCP · Should have worked on Data Migration projects, from On-prem to Cloud · Should have Cloud Storage Knowledge, Big Query, Cluster Knowledge · Sound programming knowledge on PySpark & SQL in terms of processing large amount of semi structured & unstructured data · Ability to design data pipelines in end to end manner · Knowledge on Avro, Parquet format · Knowledge on working on Hadoop Big Data platform and ecosystem · Strong debugging and troubleshooting capabilities. · Have experience to guide the Technical Team for attaining the delivery milestones
Good-to-Have 1. Knowledge on Jira, Agile, Sonar, Team city & CICD 2. Any exposure / experience for an international Banking client / multi-vendor / multi geography teams 3. Knowledge on Dataproc, Pyspark
SN Responsibility of / Expectations from the Role
1 Development in a programming language on GCP
2 Testing & Validation
3 Debugging & troubleshootin