Logo for Featherless AI

Machine Learning Engineer — Distillation

Roles & Responsibilities

  • Strong background in machine learning or deep learning
  • Hands-on experience with model distillation (LLMs or other neural networks)
  • Solid understanding of training dynamics, loss functions, and optimization
  • Experience with PyTorch (or JAX) and modern ML tooling

Requirements:

  • Design and implement knowledge distillation pipelines (teacher–student, self-distillation, multi-teacher, etc.)
  • Distill large foundation models into smaller, faster, and cheaper models for inference
  • Run and analyze large-scale training experiments to evaluate quality, latency, and cost tradeoffs
  • Collaborate with research to translate new distillation ideas into production-ready code

Job description

About the Role

We’re looking for a Machine Learning Engineer focused on model distillation to help us build smaller, faster, and more efficient models without sacrificing quality. You’ll work at the intersection of research and production—taking cutting-edge techniques and turning them into systems that scale.

This is a hands-on role with real ownership: you’ll design distillation pipelines, run large-scale experiments, and ship models used in production.

What You’ll Do

  • Design and implement knowledge distillation pipelines (teacher–student, self-distillation, multi-teacher, etc.)

  • Distill large foundation models into smaller, faster, and cheaper models for inference

  • Run and analyze large-scale training experiments to evaluate quality, latency, and cost tradeoffs

  • Collaborate with research to translate new distillation ideas into production-ready code

  • Optimize training and inference performance (memory, throughput, latency)

  • Contribute to internal tooling, evaluation frameworks, and experiment tracking

  • (Optional) Contribute back to open-source models, tooling, or research

What We’re Looking For

  • Strong background in machine learning or deep learning

  • Hands-on experience with model distillation (LLMs or other neural networks)

  • Solid understanding of training dynamics, loss functions, and optimization

  • Experience with PyTorch (or JAX) and modern ML tooling

  • Comfort running experiments on multi-GPU or distributed setups

  • Ability to reason about model quality vs. performance tradeoffs

  • Pragmatic mindset: you care about shipping, not just papers

Nice to Have

  • Experience distilling LLMs or large sequence models

  • Experience with inference optimization (quantization, pruning, kernels, etc.)

  • Familiarity with evaluation for language models

  • Open-source contributions or research publications

  • Experience in early-stage or fast-moving startups

Why Join

  • Work on core model quality and cost efficiency—not side projects

  • High ownership and direct impact on product and roadmap

  • Small, senior team with strong research + engineering culture

  • Competitive compensation + meaningful equity

  • Remote-friendly, async-first environment

Machine Learning Engineer Related jobs

Other jobs at Featherless AI

We help you get seen. Not ignored.

We help you get seen faster — by the right people.

🚀

Auto-Apply

We apply for you — automatically and instantly.

Save time, skip forms, and stay on top of every opportunity. Because you can't get seen if you're not in the race.

AI Match Feedback

Know your real match before you apply.

Get a detailed AI assessment of your profile against each job posting. Because getting seen starts with passing the filters.

Upgrade to Premium. Apply smarter and get noticed.

Upgrade to Premium

Join thousands of professionals who got noticed and hired faster.