Changelog
See our latest feature releases, product improvements and bug fixes
Introducing two new products: Model APIs and Training
Today we're introducing two new products: Baseten Model APIs and Training. Model APIs are frontier models running on the Baseten Inference Stack, purpose-built for production. Today we’re launching...
May 19, 2025Introducing our new brand
We're thrilled to introduce our new brand! We believe inference is the foundation of all AI going forward. That's what our new look is all about: Baseten is the building blocks of AI. Baseten is...
Apr 15, 2025Early Access: Announcing B200s on Baseten
We're thrilled to announce early access to NVIDIA B200 GPUs on Baseten! From benchmarks on models like DeepSeek R1, Llama 4, and Qwen, we’re already seeing 5x higher throughput, over 2x better cost...
Apr 14, 2025Flexible instance types per model deployment
Model deployments now support changing instance types, enabling you to experiment with different hardware configurations and use specific hardware for staging, development, and production...
Apr 10, 2025Stream Baseten logs from the terminal
For users who love working in the terminal, we're excited to announce truss push --tail , which streams Baseten logs directly to your command line. You no longer need to switch context between your...
Apr 7, 2025Docs refresh
We’ve overhauled the Baseten docs to make them more readable, structured, and easier to navigate for both new and returning users. Some highlights: New homepage to help new users get started All-new...
Mar 21, 2025Baseten is now fully OpenAI compatible
The OpenAI SDK has become a standard for interacting with AI models, making it extremely important in the inference space. We’re happy to announce official OpenAI-compatible APIs for both chat...
Feb 10, 2025Baseten Chains is now GA: Deploy ultra-low-latency compound AI systems
Now with improved performance, robustness, and an even more delightful DevEx since our beta launch, we’re thrilled to announce the general availability of Baseten Chains for production compound AI!...
Jan 30, 2025Health checks are now customizable
We run health checks on your deployments to ensure they’re able to run inference. Now, you can customize these checks to monitor anything , from tracking 500 errors to detecting CUDA issues and more....
Jan 21, 2025GPU metrics now available on MIG instance types
We've expanded our metrics support to include GPU memory usage and utilization for MIG (Multi-Instance GPU) instance types. These metrics were previously unavailable for MIG configurations. This...