Help Luma build some of the biggest & fastest AI supercomputing clusters in the world! As a HighPerformance Computing engineer, you’ll work at the intersection of hardware and software, designing systems that deliver the maximum possible performance for running largescale AI models. We work at the very cutting edge of speed and scale, combining the traditions of HighPerformance Computing (HPC) in a modern cloud environment.
For this role, it’s important you understand how to combine CPU’s, GPU’s, and network devices into systems that are then deployed at a large scale to peak efficiency. You understand the lowest levels of the software platforms that sit on top of this hardware, including how to best optimize the Linux kernel and userspace code. You are capable of writing code to automate the monitoring and healing of these systems, commanding a large number of servers with few people.
Responsibilities
In this role, you will work closely with and directly accelerate machine learning researchers, but dont need to be a machine learning expert yourself.
We value people who can quickly obtain a deep technical understanding of new domains and enjoy being selfdirected and identifying the most important problems to solve.
You’ll be managing training HPC clusters at Luma from provisioning to performance tuning.
Areas of work will include observability, distributed job tracing, GPU diagnostics, software environment management and additional tooling plus work on the actual code to enable necessary features.
We believe that increasing compute is a huge lever to AI progress. You will have a direct impact on our ability to grow to an unprecedented scale and likewise produce unprecedented results.
Experience
8+ years experience as infrastructure engineer or Devops in large and complex distributed systems.
Deep understanding of networking, bonus points for experience in HPC networking.
Experience developing highquality software in a generalpurpose programming language, preferably including Python.
Excellent problemsolving skills and attention to detail.
Experience with GPUs in large scale clusters is strongly preferred.
Strong knowledge of observability and monitoring in distributed systems.
Tenacious at troubleshooting hardware and network topology failures in distributed systems
Independently driven and able to own problems and build solutions from endtoend.
Experience with large scale data center operations, proficiency in cloud orchestration and system tools.
Your application is reviewed by real people.
Liberty University
Tether.io
Experian
Wilson Sonsini Goodrich & Rosati
Penn Interactive