Senior Data Specialist R01553679

Work set-up: 
Full Remote
Contract: 
Experience: 
Senior (5-10 years)
Work from: 

Offer summary

Qualifications:

Bachelor's degree in Computer Science, Data Science, or related field., At least 3 years of experience with Python and PySpark for data transformation., Strong knowledge of Google Cloud Platform services such as BigQuery, GCS, and Pub/Sub., Experience with real-time data ingestion and API integration..

Key responsibilities:

  • Design and develop data pipelines using ETL and data modeling techniques.
  • Manage and orchestrate workflows with Airflow and Kubernetes.
  • Implement real-time data ingestion and processing solutions.
  • Collaborate with domain experts to ensure data quality and reliability.

Brillio logo
Brillio XLarge https://www.brillio.com
5001 - 10000 Employees
See all jobs

Job description

Senior Data Specialist

Primary Skills
  • ETL Fundamentals, SQL, BigQuery, Dataproc, SQL (Basic + Advanced), Python, Data Catalog, Data Warehousing, Composer, Dataflow, Cloud Trace, Cloud Logging, Cloud Storage, Datafusion, Modern Data Platform Fundamentals, Data Modelling Fundamentals, PLSQL, TSQL, Stored Procedures

  • Job requirements
  • Azure Data Engineer

  • Strong with PySpark (esp. Spark SQL) for complex transformation pipelines
  • Handson with Airflow for orchestration and BigQuery SQL for querying and data modeling
  • Good experience in GCP (BigQuery, GCS, PubSub, Cloud Run, Functions,Cloud SQL)
  • Comfortable with realtime ingestion: Kafka, webhooks, filebased triggers
  • Solid API integration skills (RESTwebhooks), with ability to handle payloaddriven workflows
  • Experience working in Kubernetes (GKE) for deploying and scaling pipelines
  • Comfortable handling UIdriven configuration, YAMLbased setups, and modular frameworks
  • Exposure to schema transformation, data validation, hashing, and DQF logic
  • Domain familiarity with logisticshealthcare (CVS context is a big plus)
  • Strong ownership mindset — able to work under pressure and balance speed with reliability
  • Basic understanding of BASH, SFTP transfers, Networking, Access Management
  • Bonus: PostgreSQL, CICD, monitoring, dashboarding, or lightweight UI development
  • =================================
  • Python & PySpark (Spark SQL) – 3+ years
  • Airflow (or any orchestration tool) – 2+ years
  • Google Cloud Platform (BigQuery, GCS, PubSub, Cloud Run, Functions, Cloud SQL) – 3+ years
  • Realtime data ingestion (Kafka, webhooks, filebased) – 2+ years
  • API integration (RESTwebhooks) – 2+ years
  • Kubernetes (GKE preferred) – 1–2 years
  • BigQuery SQL & PostgreSQL – 2+ years
  • YAMLconfigdriven pipeline design – 2+ years
  • Schema transformation, hashing, DQF – 2+ years
  • CICD, observability, lightweight dashboards (GrafanaStreamlitFlask UI) – 1+ year
  • Logistics or healthcare domain exposure – nice to have
  • Required profile

    Experience

    Level of experience: Senior (5-10 years)
    Spoken language(s):
    English
    Check out the description to know which languages are mandatory.

    Other Skills

    • Calmness Under Pressure
    • Reliability

    Data Analyst Related jobs