Single-source cross-platform GPU LLM inference with Slang and Rust
- Track: AI Plumbers
- Room: UD2.120 (Chavanne)
- Day: Saturday
- Start: 13:05
- End: 13:25
- Video only: ud2120
- Chat: Join the conversation!
Leveraging Rust and Khronos' emerging Slang initiative, we introduce our efforts toward a cross-platform GPU LLM inference ecosystem. With a single-source approach we aim to minimize backend-specific code and foster community participation by writing inference kernels once and run them everywhere.
Speakers
| Crozet Sébastien |