Job Name : Data Engineer Job Location :SFO, CA (Remote) Job Type : Contract
Client : Airbnb Rate : Depend on Experience Job Authorization :US Citizen/ GC /EAD (H4/L2/TN) preferred C2c are open only for own corp not from any employers
Job Description:
Responsibilities:
Develop and automate large scale, high-performance data processing systems (batch and/or streaming) to drive Airbnb business growth and improve the product experience.
Build scalable Spark data pipelines leveraging Airflow scheduler/executor framework
Minimum Requirements:
8+ years of relevant industry experience
Demonstrated ability to analyze large data sets to identify gaps and inconsistencies, provide data insights, and advance effective product solutions
Working knowledge of relational databases and query authoring (SQL).
Good communication skills, both written and verbal
Strong experience using ETL framework (ex: Airflow, Flume, Oozie etc.) to build and deploy production-quality ETL pipelines.
Experience building batch data pipelines in Spark Scala.
Strong understanding of distributed storage and compute (S3, Hive, Spark)
General software engineering skills (Java or Python, Github)