Senior Data Engineer (GCPDatabricks)

extra holidays - extra parental leave - work from anywhere - fully flexible
Work set-up: 
Full Remote
Contract: 
Experience: 
Mid-level (2-5 years)
Work from: 

Offer summary

Qualifications:

Bachelor's degree in Computer Science, Data Engineering, or related field., At least 3 years of professional experience in data engineering, with a focus on cloud-based solutions., Strong proficiency in Python, Java, or Scala for data pipeline development., Experience with Google Cloud Platform services such as BigQuery, Dataflow, Pub/Sub, and GCP security best practices..

Key responsibilities:

  • Design and implement scalable data architectures on GCP.
  • Build and optimize ETL/ELT pipelines using Apache Airflow, dbt, Dataflow, and Pub/Sub.
  • Ensure data quality and governance using frameworks like Great Expectations.
  • Collaborate with cross-functional teams to support analytics and machine learning initiatives.

Leadtech Group logo
Leadtech Group Internet Scaleup https://leadtech.com/
501 - 1000 Employees
See all jobs

Job description

We are looking for a Senior Data Engineer to design, develop, and optimize our data infrastructure on Google Cloud Platform (GCP). You will architect scalable pipelines using Databricks, BigQuery, Google Cloud Storage, Apache Airflow, dbt, Dataflow, and PubSub, ensuring high availability and performance across our ETLELT processes. You will leverage great expectations to enforce data quality standards. The role also involves building our Data Mart (Data Mach) environment, containerizing services with Docker and Kubernetes (K8s), and implementing CICD best practices.

A successful candidate has extensive knowledge of cloudnative data solutions, strong proficiency with ETLELT frameworks (including dbt), and a passion for building robust, costeffective pipelines.

Key Responsibilities
Data Architecture & Strategy
  • Define and implement the overall data architecture on GCP, including data warehousing in BigQuery, data lake patterns in Google Cloud Storage, and Data Mart (Data Mach) solutions.
  • Integrate Terraform for Infrastructure as Code to provision and manage cloud resources efficiently.
  • Establish both batch and realtime data processing frameworks to ensure reliability, scalability, and cost efficiency.
    • Pipeline Development & Orchestration
      • Design, build, and optimize ETLELT pipelines using Apache Airflow for workflow orchestration.
      • Implement dbt (Data Build Tool) transformations to maintain versioncontrolled data models in BigQuery, ensuring consistency and reliability across the data pipeline.
      • Use Google Dataflow (based on Apache Beam) and PubSub for largescale streamingbatch data processing and ingestion.
      • Automate job scheduling and data transformations to deliver timely insights for analytics, machine learning, and reporting.
        • EventDriven & Microservices Architecture
          • Implement eventdriven or asynchronous data workflows between microservices.
          • Employ Docker and Kubernetes (K8s) for containerization and orchestration, enabling flexible and efficient microservicesbased data workflows.
              • Implement CICD pipelines for streamlined development, testing, and deployment of data engineering components.
                • Data Quality, Governance & Security
                  • Enforce data quality standards using Great Expectations or similar frameworks, defining and validating expectations for critical datasets.
                  • Define and uphold metadata management, data lineage, and auditing standards to ensure trustworthy datasets.
                  • Implement security best practices, including encryption at rest and in transit, Identity and Access Management (IAM), and compliance with GDPR or CCPA where applicable.
                    • BI & Analytics Enablement
                      • Integrate with Looker (or similar BI tools) to provide data consumers with intuitive dashboards and realtime insights.
                      • Collaborate with Data Science, Analytics, and Product teams to ensure the data infrastructure supports advanced analytics, including machine learning initiatives.
                      • Maintain Data Mart (Data Mach) environments that cater to specific business domains, optimizing access and performance for key stakeholders.
                        • Requirements

                          • 3+ years of professional experience in data engineering, with at least 1 year in mobile data.
                          • Proven track record building and maintaining BigQuery environments and Google Cloud Storagebased data lakes.
                          • Deep knowledge of Apache Airflow for schedulingorchestration and ETLELT design.
                          • Experience implementing dbt for data transformations, RabbitMQ for eventdriven workflows, and PubSub + Dataflow for streamingbatch data pipelines.
                          • Familiarity with designing and implementing Data Mart (Data Mach) solutions, as well as using Terraform for IaC.
                          • Strong coding capabilities in Python, Java, or Scala, plus scripting for automation.
                          • Experience with Docker and Kubernetes (K8s) for containerizing datarelated services.
                          • Handson with CICD pipelines and DevOps tools (e.g., Terraform, Ansible, Jenkins, GitLab CI) to manage infrastructure and deployments.
                          • Proficiency in Great Expectations (or similar) to define and enforce data quality standards.
                          • Expertise in designing systems for data lineage, metadata management, and compliance (GDPR, CCPA).
                          • Strong understanding of OLTP (Online Transaction Processing) and OLAP (Online Analytical Processing) systems.
                          • Excellent communication skills for both technical and nontechnical audiences.
                          • High level of organization, selfmotivation, and problemsolving aptitude.
                            • Will be a plus
                              • Machine Learning (ML) Integration: Familiarity with endtoend ML workflows and model deployment on GCP (e.g., Vertex AI).
                              • Advanced Observability: Experience with Prometheus, Grafana, Datadog, or New Relic for system health and performance monitoring.
                              • Security & Compliance: Advanced knowledge of compliance frameworks such as HIPAA, SOC 2, or relevant regulations.
                              • RealTime Data Architectures: Additional proficiency in Kafka, Spark Streaming, or other streaming solutions.
                              • Certifications: GCPspecific certifications (e.g., Google Professional Data Engineer) are highly desirable.
                                • Benefits

                                  Why should you join us?

                                  Growth and career development

                                  • At Leadtech, we prioritize your growth. Enjoy a flexible career path with personalized internal training and an annual budget for external learning opportunities.
                                    • WorkLife balance

                                      • Benefit from a flexible schedule with flextime (7 9:30 a.m. start, 3:30 6 p.m. end) and the option of working full remote or from our Barcelona office. Enjoy free Friday afternoons with a 7hour workday, plus a 35hour workweek in July and August so you can savor summer!
                                        • Comprehensive benefits

                                          • Competitive salary, fulltime permanent contract, and toptier private health insurance (including dental and psychological services).
                                          • 25 days of vacation plus your birthday off, with flexible vacation options—no blackout days!
                                            • Unique Perks

                                              • If you wish to come, in our office in Barcelona you’ll find it coplete with free coffee, fresh fruit, snacks, a game room, and a rooftop terrace with stunning Mediterranean views.
                                              • Additional benefits include ticket restaurant and nursery vouchers, paid directly from your gross salary.
                                                • Join us in an environment where you’re free to innovate, learn, and grow alongside passionate professionals. At Leadtech, you’ll tackle exciting challenges and be part of a vibrant team dedicated to delivering exceptional user experiences

                                                  Equal Employment Opportunity Employer:

                                                  Leadtech is an Equal Employment Opportunity (EEO) Employer, which means we encourage applications from people with different backgrounds, interests, and personal circumstances. Our team welcomes applicants regardless of their race, gender, age, religion, nationality, sexual orientation, andor disabilities. All we need is your high energy, skills, and willingness to be a part of a great project!

                                                  Location

                                                  Youll have the flexibility to choose whether youd like to come to the office every day, from time to time, or work fully remote. We want you to find the best combination for you.

                                                  If you prefer to be surrounded with amazing people, our exceptional office is in Barcelona’s Blue Building, located right on the citys seafront. Besides our stunning views, you’ll enjoy our office perks such as free fruit, snacks, and coffee and you’ll also be able to take part in our Mario Kart and table tennis competitions.

                                                  The personal data you provide will be processed in order to manage your candidacy for the corporate selection processes that fit your profile. If you wish, you can exercise your rights of access, rectification or cancellation by writing to our address (Avenida Litoral, 1214, 5ta planta. Barcelona. 08005) or to the email address protecciondedatos@LeadTech.com, attaching to your request a document that can validate your identity.

Required profile

Experience

Level of experience: Mid-level (2-5 years)
Industry :
Internet
Spoken language(s):
English
Check out the description to know which languages are mandatory.

Other Skills

  • Self-Motivation
  • Communication
  • Problem Solving

Data Engineer Related jobs