This talk explores building a complete self-hosted LLM stack in Rust: Paddler, a distributed load balancer for serving LLMs at scale, and Poet, a static site generator that consumes those LLMs for AI-powered content features.

This talk explores building a complete self-hosted LLM stack in Rust: Paddler, a distributed load balancer for serving LLMs at scale, and Poet, a static site generator that consumes those LLMs for AI-powered content features.
We'll dive into the hard problems: async request routing across dynamic agent fleets, integrating with llama.cpp's C++ codebase, managing KV cache in custom slots, and implementing zero-to-N autoscaling with request buffering. You'll see how Rust's ownership model prevented entire classes of bugs in distributed state management, and walk away with concrete patterns for building and consuming LLM infrastructure in production.
This talk explores what it means to write scientific software that lives up to the standards we expect of science itself.
I'll share a few tricks to help you write cleaner, more powerful declarative macros. You'll also get a sneak peek at the nightly features to see what's coming next macro_rules! world.
In this introductory talk, we will explore what it means to "Ratatuify" the Rust package manager, Cargo.
In this talk, we’ll re-create the core ideas of Karpathy’s micrograd, but entirely in Rust.
In my session, I will present the https://hotpath.rs crate and explain how it compares to other profiling tools available.
This talk puts popular Rust rewrites to the test. We'll examine how these tools stack up against their battle-tested predecessors, looking at real-world performance, compilation times, binary sizes, feature completeness, and ecosystem maturity.