5+ years of experience in data engineering with strong SQL and ETL skills
Experience with SSIS and legacy SQL Server–based data marts
Proficiency in Google BigQuery and/or similar cloud data warehouses, with hands-on experience using DBT or other modern transformation frameworks
Experience migrating on-prem pipelines to cloud platforms and familiarity with GCP-native services (Cloud Storage, Pub/Sub, Dataflow, Composer, IAM)
Requirements:
Lead the migration of the existing SSIS-based ETL workflows to cloud-native pipelines using DBT and/or GCP tools such as Dataflow, Dataform, or Cloud Composer (Airflow)
Design and implement scalable data models in BigQuery, following best practices for dimensional modeling
Optimize and maintain existing SQL transformations, ensuring correctness and performance in the cloud
Collaborate with BI developers and analysts to ensure data marts align with Tableau reporting needs; ensure data integrity, security, and lineage through testing, documentation, and observability practices
Job description
Job: Sr. Data engineer Location: Remote (USA) Years of experience: 5+yrs
Our client is seeking Senior Data Engineer to help modernize and scale our Business Intelligence data infrastructure. This role will be critical in transitioning our existing on-prem SQL Server–based dimensional data mart and SSIS pipelines into a cloud-native solution on Google Cloud Platform (GCP) using tools like DBT, BigQuery, and other GCP-native services.
This is a hands-on engineering role focused on building robust, scalable data pipelines and enabling performant data models that power Tableau dashboards used throughout the organization.
Responsibilities:
· Lead the migration of the existing SSIS-based ETL workflows to cloud-native pipelines using DBT and/or GCP tools such as Dataflow, Dataform, or Cloud Composer (Airflow).
Design and implement scalable, efficient data models in BigQuery, following best practices for dimensional modeling.
Optimize and maintain existing SQL transformations, ensuring correctness and performance in the cloud.
Collaborate with BI developers and analysts to ensure data marts align with Tableau reporting needs.
Ensure data integrity, security, and lineage through testing, documentation, and observability practices.
Work with on-prem teams to phase out legacy systems and design transitional architectures where needed.
Establish best practices and mentor junior engineers on cloud-Financial engineering patterns.
Qualifications: Required:
5+ years of experience in data engineering with strong SQL and ETL skills.
Experience with SSIS and legacy SQL Server–based data marts.
Proficiency in Google BigQuery and/or similar cloud data warehouses.
Hands-on experience with DBT or modern transformation frameworks.
Strong knowledge of dimensional modeling and data warehousing principles.
Experience migrating on-prem pipelines to cloud platforms.
Familiarity with GCP-native services such as Cloud Storage, Pub/Sub, Dataflow, Composer, and IAM.
Strong knowledge of Healthcare Information Systems.
Preferred:
Experience supporting or integrating with Tableau-based BI solutions.
Exposure to infrastructure-as-code tools like Terraform for GCP.
Knowledge of data observability tools and practices.
Comfortable with Git-based CI/CD for data pipeline deployments.
Nice to Have:
Familiarity with GCP networking and cost optimization strategies.
Experience with data validation or automated testing frameworks for pipelines.
Knowledge of metadata management or cataloging tools (e.g., Data Catalog, DataPlex).
What You'll Bring:
A builder's mindset with a bias for simplification and automation.
A collaborative approach to working with BI and application teams.
The ability to balance long-term platform scalability with short-term deliverables.
A passion for cloud innovation and data platform modernization.