Baseten home page
Search...
⌘K
Get started
Overview
Quick start
Concepts
Why Baseten
How Baseten works
Development
Concepts
Model APIs
Developing a model
Developing a Chain
Deployment
Concepts
Deployments
Environments
Resources
Autoscaling
Inference
Concepts
Call your model
Streaming
Async inference
Structured LLM output
Output formats
Integrations
Training
Overview
Getting started
Concepts
Management
Deploying checkpoints
Observability
Metrics
Status and health
Security
Exporting metrics
Tracing
Billing and usage
Troubleshooting
Deployments
Inference
Support
Return to Baseten
Baseten home page
Search...
⌘K
Ask AI
Support
Return to Baseten
Return to Baseten
Search...
Navigation
Get started
Quick start
Documentation
Examples
Reference
Status
Documentation
Examples
Reference
Status
Get started
Quick start
1
What modality are you working with?
Choose from common modalities like LLMs, transcription, and image generation to get started quickly.
LLMs
Build and deploy large language models
Transcription
Transcribe audio and video
Image generation
Rapidly generate images
Text to speech
Build humanlike experiences
Compound AI
Build real-time AI-native applications
Embeddings
Process millions of data points
Custom models
Deploy any model
2
Select a model or guide to get started...
Choose a use case or modality above first…
Was this page helpful?
Yes
No
Previous
Why Baseten
Baseten delivers fast, scalable AI/ML inference with enterprise-grade security and reliability—whether in our cloud or yours.
Next
Assistant
Responses are generated using AI and may contain mistakes.