Founding Engineer, ML Inference
2 weeks ago
We're looking for a Founding Engineer, ML Inference with deep expertise in high-performance ML engineering. This is a highly technical, high-impact role focused on squeezing every drop of performance from real-time generative media models.
The information below covers the role requirements, expected candidate experience, and accompanying qualifications.
You'll work across the model-serving stack, designing novel inference frameworks, optimizing inference performance, and shaping Reactor's competitive edge in ultra-low-latency, high-throughput environments. We want to establish new inference frameworks in this domain and you will be able to own this part of our stack.
What You'll Do • Drive our frontier position on real-time model performance for diffusion models
• Design and implement a high-performance in-house inference runtime
• Implement optimizations using torch.compile, custom CUDA kernels, and specialized inference frameworks
• Optimize neural network models for inference through quantization, pruning, and architectural modifications while maintaining accuracy
• Profile and benchmark model performance to identify computational bottlenecks
• Collaborate directly with model partner teams to directly integrate their models into our platform
Required Skills • Strong foundation in systems programming, with a track record of identifying and resolving bottlenecks
• Deep expertise in the ML infrastructure stack: ◦ PyTorch, TensorRT, TransformerEngine, Nsight, ONNX Runtime
◦ Model compilation, quantization (INT8/FP16), and advanced serving architectures
• Working knowledge of GPU hardware (NVIDIA) and the ability to dive deep into the stack as needed
• Strong understanding of transformer architectures and modern ML model optimization techniques
Logistics We are based in-person in San Francisco. We believe the best ideas and work come from being together.
• Competitive San Francisco salary and meaningful early equity.
• We sponsor visas. We are committed to working through the process together for the right candidates. If you're currently outside the US, we're also committed to helping you relocate to the US throughout this process. xrczosw
• We offer generous health, dental, and vision coverage, and relocation support as needed.
If this sounds like you, we'd love to hear from you.
-
Founding Engineer, ML Inference
2 weeks ago
San Francisco, United States Reactor Full timeWe're looking for a Founding Engineer, ML Inference with deep expertise in high-performance ML engineering. This is a highly technical, high-impact role focused on squeezing every drop of performance from real-time generative media models.You'll work across the model-serving stack, designing novel inference frameworks, optimizing inference performance, and...
-
San Francisco, CA, United States Reactor Full timeA pioneering technology firm in San Francisco is seeking a Founding Engineer for ML Inference. Please ensure you read the below overview and requirements for this employment opportunity completely. This highly technical role focuses on optimizing real-time generative media models. You'll design novel inference frameworks and work to maintain...
-
San Francisco, United States Reactor Full timeA pioneering technology firm in San Francisco is seeking a Founding Engineer for ML Inference. This highly technical role focuses on optimizing real-time generative media models. You'll design novel inference frameworks and work to maintain competitive advantage in ultra-low-latency environments. The ideal candidate has a strong foundation in systems...
-
ML Ops Engineer — Agentic AI Lab
3 weeks ago
San Francisco, CA, United States Fabrion Full timeAbout the Role ML Ops Engineer — Agentic AI Lab (Founding Team) — Location: San Francisco Bay Area — Type: Full-Time — Compensation: Competitive salary + meaningful equity (founding tier) Backed by 8VC, we're building a world-class team to tackle one of the industry’s most critical infrastructure problems. Our AI Lab is pioneering the future of...
-
Sr. Software Engineer, ML Edge Inference
6 hours ago
San Francisco, United States Serve Robotics Full timeSr. Software Engineer, ML Edge Inference Engineer Join to apply for the Sr. Software Engineer, ML Edge Inference Engineer role at Serve Robotics. Base pay range $190,000.00/yr - $240,000.00/yr At Serve Robotics, we’re reimagining how things move in cities. Our personable sidewalk robot is our vision for the future. It’s designed to take deliveries away...
-
ML Engineer
6 hours ago
San Jose, United States Photalabs Full timeAt Phota Labs, we’re building visual GenAI that helps people capture, express, and relive their memories — in ways that feel effortless, personal, and emotionally resonant. Our core technology enables personalized image generation that faithfully reflects who you are and the moments you experienced. Our first goal is to bring visual GenAI into everyday...
-
ML/AI Founding Engineer
3 weeks ago
San Francisco, United States Navi AI Full timeML/AI Founding Engineer Navi AI As an AI/ML Engineer you’ll design and deploy machine learning systems that power everything from student training flights to commercial airline operations to high-performance fighter jet sorties. At Navi, artificial intelligence isn’t an experiment—it’s mission‑critical. We’re building the AI co‑pilot that will...
-
ML Platform Engineer
4 weeks ago
San Francisco, CA, United States Apple Inc. Full timeA leading technology company in San Francisco is seeking a dedicated ML Engineer to enhance model training and inference workloads in the cloud. The information below covers the role requirements, expected candidate experience, and accompanying qualifications. This role provides a unique opportunity to collaborate with senior ML engineers and improve the...
-
Founding ML Researcher
2 hours ago
San Francisco, California, United States Unsiloed AI Full timeWe are hiring a Founding ML Researcher in San Francisco.We are building a small, talent-dense team. This role will define the engineering archetype at Unsiloed AI and set the ceiling for the team. We strongly believe technical DNA compounds (or degrades) with every hire and hence the first few matter disproportionately. You will be expected to operate...
-
Staff ML Engineer, Inference Platform
2 weeks ago
Sunnyvale, CA, United States General Motors Full timeHybrid This role is categorized as hybrid. This means the successful candidate is expected to report to the Sunnyvale Tecnical Center, CA at least three times per week, at minimum or other frequency dictated by the business. This job is eligible for relocation assistance. Making sure you fit the guidelines as an applicant for this role is essential, please...