Role overview
The Data Engineer (Spark) role entails working on complex data pipelines by leveraging new age Cloud and BigData technologies, to feed data to AI and analytical applications.
What you'll work on
-
Implement scalable solutions for ever-increasing data volumes, using big data/cloud technologies like Pyspark, Kafka, etc.
-
Implement real-time data ingestion and processing solutions
-
Work closely with Data scientists for data extraction, analysis for machine learning
-
Work closely with the development teams to develop and maintain scalable data platform architecture and pipelines across our working environments
-
Utilize the latest technologies to ensure ease of data integrations and reusability across various teams