Offer summary
Qualifications:
5+ years in data engineering, 3+ years with Apache Spark and Amazon EMR, Strong programming skills in Python and Scala, Proficient in SQL for data management, Experience with AWS services such as S3, EMR.
Key responsabilities:
- Design and optimize scalable ETL pipelines
- Collaborate to maintain high-performance data infrastructure
- Develop automated workflows for data ingestion
- Implement best practices for monitoring data pipelines
- Drive innovation by integrating new tools