Logo for Aledade, Inc.

Senior Data Platform Engineer II (Databricks)

Roles & Responsibilities

  • 6+ years experience building and optimizing highly scalable distributed data systems (e.g., Databricks, Spark, or Snowflake).
  • 3+ years of experience working with SQL and data modeling on large multi-table data sets.
  • 3+ years of experience acting as a trusted technical decision-maker in a team setting, solving for short-term and long-term business value.
  • 3+ years of experience coaching other engineers.

Requirements:

  • Architect and manage scalable, high-performance distributed data environments powering healthcare analytics (Databricks Lakehouse and Snowflake), ensuring health, optimization, and security.
  • Develop and implement scalable and performant data solutions.
  • Partner with Engineering Managers, Product Managers, and stakeholders across Aledade to develop and execute technical roadmaps using Agile processes.
  • Mentor and coach more junior engineers, including thorough pull request reviews and receptiveness to critical feedback.

Job description

As a Senior Data Platform Engineer II, you will architect and manage the high-performance, distributed data environments that power our healthcare analytics. You will move beyond traditional maintenance to ensure our Databricks Lakehouse and Snowflake environments scale indefinitely. You will be responsible for the health, optimization, and security of our data platforms, making complex data accessible and expressive for web applications and AI.

Primary Duties:
  • Develop and implement scalable and performant solutions.
  • Partner, as a peer, with Engineering Managers, Product Managers, and stakeholders throughout Aledade to develop and execute technical roadmaps using Agile processes.
  • Mentor and coach more junior engineers including thorough pull request reviews for other developers and be receptive to critical feedback on your own work.

  • Minimum Qualifications:
  • BS/BTech (or higher) in Computer Science, Engineering or a related field or equivalent experience.
  • 6+ years experience as an engineer building and optimizing highly scalable distributed data systems (e.g., Databricks, Spark, or Snowflake).
  • 3+ years of experience working with SQL and data modeling on large multi-table data sets.
  • 3+ years of experience acting as a trusted technical decision-maker in a team setting, solving for short-term and long-term business value.
  • 3+ years of experience coaching other engineers.

  • Preferred KSA’s:

  • Platform & Infrastructure (The "Databricks/Cloud" Core)
  • Databricks & Lakehouse Architecture: Deep expertise in managing Databricks workspaces, including Unity Catalog for data governance, lineage, and fine-grained access control.
  • Infrastructure as Code (IaC): Advanced proficiency with Terraform (or similar) to automate the provisioning and scaling of Databricks clusters, cloud resources (AWS preferred), and networking.
  • Snowflake Proficiency (Nice-to-Have): Experience managing Snowflake environments, specifically focusing on warehouse cost optimization, security integration, and secure data sharing.
  • Modern Database Internals: In-depth knowledge of distributed systems, including partitioning, liquid clustering/Z-Ordering, sharding, and high-availability strategies for petabyte-scale data.
  • Performance, Reliability & DevOps
  • Observability & Optimization: Proven track record in performance monitoring and query tuning for distributed workloads to ensure system reliability and cost-efficiency.
  • Data Engineering Lifecycle: Experience designing and optimizing high-throughput ETL/ELT pipelines and ingestion systems (batch and streaming) using Spark.
  • Deployment & Orchestration: Experience building robust CI/CD pipelines for data infrastructure and deploying services using containerization (Docker, Kubernetes).

  • Security, Compliance & Domain Knowledge
  • Sensitive Data Handling: Expertise in building systems that handle protected information, with specific experience in HIPAA and SOX compliance frameworks.
  • Healthcare Data Expertise: Experience navigating health-tech data complexities, such as Electronic Health Records (EHR), clinical data formats (HL7/FHIR), and claims data.

  • Physical Requirements:
  • Sitting for prolonged periods of time. Extensive use of computers and keyboard. Occasional walking and lifting may be required.
  • Data Engineer Related jobs

    Other jobs at Aledade, Inc.

    We help you get seen. Not ignored.

    We help you get seen faster — by the right people.

    🚀

    Auto-Apply

    We apply for you — automatically and instantly.

    Save time, skip forms, and stay on top of every opportunity. Because you can't get seen if you're not in the race.

    AI Match Feedback

    Know your real match before you apply.

    Get a detailed AI assessment of your profile against each job posting. Because getting seen starts with passing the filters.

    Upgrade to Premium. Apply smarter and get noticed.

    Upgrade to Premium

    Join thousands of professionals who got noticed and hired faster.