The future of AI is open-source. Let's build together.
Artificial Intelligence • Cloud Computing • LLM • Open Source • Decentralized Computing
June 26
🏢 In-office - San Francisco
The future of AI is open-source. Let's build together.
Artificial Intelligence • Cloud Computing • LLM • Open Source • Decentralized Computing
• Develop and optimize large language model (LLM) inference frameworks • Optimize GPU and CUDA performance for high-performance LLM inference engine • Design and implement distributed inference infrastructure for LLM • Build monitoring and management tools for online inference servers reliability and scalability
• Bachelor’s degree in Computer Science, Computer Engineering, or a relevant technical field, or equivalent practical experience • Proficient in C/C++, Python, or Rust with a strong understanding of algorithms and data structures • At least 3 years of experience working in ML infrastructure • Solid understanding of deep learning training frameworks such as PyTorch and TensorFlow
• Competitive compensation • Startup equity • Health insurance • Remote work flexibility
Apply Now