Leadership: 3–5+ years in data engineering or backend data systems, with experience leading data pipeline projects and mentoring junior engineers.
Data Pipeline Expertise: Proficient in building and optimizing ETL/ELT pipelines using tools like Apache Airflow, Spark, dbt, Kafka, Databricks, or Flink.
Infrastructure Ops: Familiar with cloud platforms (AWS, GCP, Azure), containerization (Docker/Kubernetes), and orchestration, with experience in CI/CD and data versioning (e.g., DVC).
Requirements:
Lead data pipeline projects and mentor junior engineers in a data engineering team.
Build and optimize ETL/ELT pipelines using Apache Airflow, Spark, dbt, Kafka, Databricks, or Flink.
Work with SQL, Python or Scala to manage data warehouses (Snowflake, BigQuery, Redshift) and data lakes (Delta Lake, Lakehouse).
Collaborate with analytics, ML, or BI teams; contribute to data governance, privacy/compliance (GDPR, HIPAA); support real-time data processing.
Job description
This is a remote position.
Job Title: Data Analyst
Location: Remote
Duration: Full-Time
NEED IRS MBI Clearance.
• Experience & Leadership: 3–5+ years in data engineering or backend data systems, with experience leading data pipeline projects and mentoring junior engineers.
• Data Pipeline Expertise: Proficient in building and optimizing ETL/ELT pipelines using tools like Apache Airflow, Spark, dbt, Kafka, Databricks, or Flink.
• Tech Stack & Tools: Strong skills in SQL, Python or Scala, and working with data warehouses (Snowflake, BigQuery, Redshift) and data lakes (e.g., Delta Lake, Lakehouse architecture).
• Infrastructure & Ops: Familiar with cloud platforms (AWS, GCP, Azure), containerization (Docker/Kubernetes), and orchestration, with experience in CI/CD and data versioning (e.g., DVC).
• Bonus/Preferred: Experience with real-time data processing, cataloging, data governance, privacy/compliance (GDPR, HIPAA), and cross-functional collaboration with analytics, ML, or BI teams.