**Who this is for** This role is for experienced Machine Learning Engineers who are passionate about optimizing the inference performance of large-scale generat
Work type: remote
Location: United States
Type: Full-time
**Who this is for** This role is for experienced Machine Learning Engineers who are passionate about optimizing the inference performance of large-scale generative AI models for real-world production environments. **Key highlights** You will work with cutting-edge high-performance computing resources to prototype and deploy multi-modal generative pipelines while collaborating with top-tier research teams at Stability AI. **You might be a good fit if you...** - Have 7+ years of experience in productionizing machine learning inference pipelines. - Possess expert knowledge in Python and high-performance frameworks like Triton or TensorRT. - Maintain a deep understanding of diffusion model architectures and Nvidia GPU profiling tools. - Are comfortable working with cloud orchestration (Kubernetes) and containerization technologies.
Generative AI Inference Engineer
About the role:
We are seeking passionate Machine Learning Engineers to join our Inference team, focusing on the creative applications of generative AI models. The ideal candidate will have substantial experience developing and running inference for multi-modal models. A deep understanding of diffusion model architectures and familiarity with workflow tools like ComfyUI are a big plus. You will be expected to leverage and push the boundaries of state-of-the-art inference optimization techniques for multi-modal generative models. This role offers the opportunity to work alongside top researchers and engineers, utilizing cutting-edge high-performance computing resources to make a significant impact in the rapidly evolving field of generative AI.
Responsibilities:
We are an equal opportunity employer and do not discriminate on the basis of race, religion, national origin, gender, sexual orientation, age, veteran status, disability or other legally protected statuses.