Offer summary
Qualifications:
Proven experience in data engineering with distributed systems like Hadoop or GCP, Familiarity with Scio, GCP Dataflow, AWS EMR, and Apache Beam, Strong programming skills in Python, Java, or Scala, Commitment to agile processes, data-driven development, reliability, Hands-on experience with orchestration tools like Airflow.
Key responsabilities:
- Maintain & enhance Spotify's measurement pipelines for efficient data collection
- Build large-scale batch and real-time data pipelines for processing
- Enhance data models & ETL code to improve efficiency and quality
- Collaborate with stakeholders to translate business requirements into technical solutions
- Write scalable, testable, and high-quality code