DataBricks Architect [EGN] at Crodu

Remote: 
Full Remote
Contract: 
Work from: 

Offer summary

Qualifications:

8+ years of experience in data engineering or related roles, Strong hands-on experience with Databricks and Apache Spark, Excellent knowledge of Python and experience with cloud migrations, Strong communication and teamwork skills, with a proactive mindset..

Key responsibilities:

  • Ensuring secure data storage and processing DICOM data
  • Validating data, building processing pipelines, and creating cohorts
  • Planning and executing database migrations
  • Collaborating closely with a cross-functional team including data engineers and data scientists.

CRODU logo
CRODU Information Technology & Services Scaleup https://crodu.com/
51 - 200 Employees
See all jobs

Job description

🌴 Work setup: long-term, full-time, 100% remote 👈
Start: ASAP 👈

Hey there! 👋
We’re looking for Databricks Architects for our US-based client. The work focuses on areas like migration, data ingestion, and optimization of Databricks-based solutions. The client has an ongoing need for specialists - their projects tend to be short-term (with a high chance of extension), and since the demand is steady, they’re usually able to offer new assignments after each one wraps up.

Right now, they’re looking for experts to join an AI/ML healthcare project. The platform is analyzing both text data and images generated by medical devices (X-rays, MRI, etc.). The gathered data will be migrated to a Databricks-powered cloud database. The platform will manage the full data lifecycle and include built-in features for compliance, auditing, cohort creation, and model reuse. The ultimate goal is to address issues in current data systems (e.g., scattered sources, manual processes, insufficient security).
We're looking for engineers fluent in Python, with strong cloud experience and solid knowledge of Databricks and Apache Spark. The projects are mostly for US-based companies - in most cases, working hours require only a slight overlap (like 10:00–18:00 CET), but we’re flexible on the schedule.

Scope:

📍 Ensuring secure data storage
📍 Processing and indexing DICOM data
📍 Validating data, building processing pipelines, creating and sharing cohorts
📍 Planning and executing database migrations
📍 Collaborating closely with a cross-functional team (data engineers, data scientists, clinical informaticians, support teams, etc.)

What do we need?
⚡️ 8+ years of experience in data engineering or related roles
⚡️ Strong hands-on experience with Databricks and Apache Spark
⚡️ Excellent knowledge of Python
⚡️ Experience with cloud migrations
⚡️ Hands-on experience with AWS (especially Amazon S3)
⚡️ Previous work on AI/ML projects
⚡️ Strong communication & teamwork skills
⚡️ A proactive mindset and independence
⚡️ English good enough for smooth daily communication
Nice to have:

⚡️ Experience with Azure (Data Factory, Synapse, Logic Apps, Data Lake)
⚡️ Skills in designing/optimizing data pipelines using DBT, SSIS, TimeXtender or similar (ETL/ELT tools)
⚡️ Experience with big data or NoSQL platforms (e.g. Redshift, Hadoop, EMR, Google Data)

What do we offer and how do we work?
🎯 Open communication throughout the recruitment process and beyond – we care about transparency
🎯 We keep our hiring process simple and candidate-friendly
🎯 We work remote-first – travel is kept to a minimum
🎯 We offer private medical care (Medicover) and Multisport card for contractors

How to apply? 👈
Just send us your application via the form!

Required profile

Experience

Industry :
Information Technology & Services
Spoken language(s):
English
Check out the description to know which languages are mandatory.

Other Skills

  • Open Mindset
  • Teamwork
  • Communication

Related jobs