Ability to analyze large data sets to identify gaps, provide data insights, and drive effective product solutions
Working knowledge of relational databases and SQL
Experience with ETL frameworks (Airflow, Flume, Oozie) to build and deploy production-quality pipelines
Requirements:
Develop and automate large-scale, high-performance data processing systems (batch and/or streaming)
Drive Airbnb business growth and improve the product experience through data processing and analytics
Build scalable Spark data pipelines leveraging Airflow scheduler/executor framework
Job description
Job Name : Data Engineer Job Location :SFO, CA (Remote) Job Type : Contract
Client : Airbnb Rate : Depend on Experience Job Authorization :US Citizen/ GC /EAD (H4/L2/TN) preferred C2c are open only for own corp not from any employers
Job Description:
Responsibilities:
Develop and automate large scale, high-performance data processing systems (batch and/or streaming) to drive Airbnb business growth and improve the product experience.
Build scalable Spark data pipelines leveraging Airflow scheduler/executor framework
Minimum Requirements:
8+ years of relevant industry experience
Demonstrated ability to analyze large data sets to identify gaps and inconsistencies, provide data insights, and advance effective product solutions
Working knowledge of relational databases and query authoring (SQL).
Good communication skills, both written and verbal
Strong experience using ETL framework (ex: Airflow, Flume, Oozie etc.) to build and deploy production-quality ETL pipelines.
Experience building batch data pipelines in Spark Scala.
Strong understanding of distributed storage and compute (S3, Hive, Spark)
General software engineering skills (Java or Python, Github)