Loops is a Tinker-compatible training SDK for post-training large models at long sequence lengths. It lets you deploy dedicated training and sampling servers for any supported base model, then run your existing Tinker scripts with minimal changes. Baseten’s Loops SDK also includes primitives for async RL, so you can train across long-horizon workloads without pipeline bubbles that increase wall-clock time and execution variance. You can then deploy sampler checkpoints directly to the Baseten Inference Stack.Documentation Index
Fetch the complete documentation index at: https://docs.baseten.co/llms.txt
Use this file to discover all available pages before exploring further.
How Loops works
Loops provides API-driven training infrastructure by deploying training servers that execute traditional forward and backward passes plus optimizer steps. It decouples RL inference into sampling servers, making RL inference scalable for compute-intensive workloads. Trainer and sampler stay synchronized via weight transfers that you can await synchronously or asynchronously, so you can stay on-policy or run bounded off-policy algorithms. In Loops, you own your checkpoints. You can download them as presigned URLs or deploy them onto Baseten’s Inference Stack via UI, CLI, or API. If you’re not sure Loops is the right path for your team, the Training overview compares Loops withtruss train (the bring-your-own-container alternative) side by side.