Exceptional generalist engineers for AI inference engine development, optimizing CUDA kernels and designing distributed systems. Fully remote opportunity with a focus on autonomy.

Inferact is a startup founded by creators and core maintainers of vLLM, the most popular open-source LLM inference engine.
Our mission is to grow vLLM as the world's AI inference engine and accelerate AI progress by making inference cheaper and faster.
Browse and apply for open jobs at Inferact.
Exceptional generalist engineers for AI inference engine development, optimizing CUDA kernels and designing distributed systems. Fully remote opportunity with a focus on autonomy.