The future of AI is open-source. Let's build together.
Artificial Intelligence • Cloud Computing • LLM • Open Source • Decentralized Computing
June 7
🏢 In-office - San Francisco
The future of AI is open-source. Let's build together.
Artificial Intelligence • Cloud Computing • LLM • Open Source • Decentralized Computing
• Seeking a Machine Learning Engineer to join our Inference Engine team, focusing on optimizing and enhancing the performance of AI inference systems • Involves working with state-of-the-art large language models and ensuring they run efficiently and effectively at scale • Opportunity to collaborate closely with AI researchers and engineers to create cutting-edge AI solutions
• 3+ years of experience writing high-performance, well-tested, production-quality code • Proficiency with Python and PyTorch • Demonstrated experience in building high performance libraries and tooling • Excellent understanding of low-level operating systems concepts including multi-threading, memory management, networking, storage, performance, and scale • Preferred: Knowledge of existing AI inference systems such as TGI, vLLM, TensorRT-LLM, Optimum • Preferred: Knowledge of AI inference techniques such as speculative decoding • Preferred: Knowledge of CUDA/Triton programming • Nice to have: Knowledge of Rust, Cython and compilers
• competitive compensation • startup equity • health insurance • other competitive benefits
Apply Now