New metrics dashboard customization
We’ve revamped our metrics dashboard to make monitoring and debugging easier! Here’s what’s new: Unified view : All metrics are now displayed on a single page—no more clicking between tabs. This...
See our latest feature releases, product improvements and bug fixes
Dec 20, 2024
We’ve revamped our metrics dashboard to make monitoring and debugging easier! Here’s what’s new: Unified view : All metrics are now displayed on a single page—no more clicking between tabs. This...
Dec 19, 2024
Our new Speculative Decoding integration lets you leverage speculative decoding as part of our streamlined TensorRT-LLM Engine Builder flow. Just modify the new speculator configuration in the Engine...
Dec 13, 2024
We’ve added several new endpoints to our REST API, giving you even more control over your deployments, environments, and resources. Here’s what’s new: Deletion Endpoints Delete a model:...
Dec 13, 2024
Our async inference service now supports delivering async predict results to your webhook endpoints over HTTP/2. This means faster, more efficient connections for your webhook integrations. Don’t...
Dec 6, 2024
Get more visibility into activity across your workspace, models, and Chains with the new Activity Feed ! Click the Activity tab to view a detailed list of changes, including who made them and when....
Dec 6, 2024
[No action needed] As of truss version 0.9.55, the flag --trusted in truss push is no longer needed to use secrets in your deployed models. Secrets specified in your config.yaml will automatically be...
Dec 5, 2024
Our new Custom Servers feature lets you deploy production-ready model servers directly from Docker images using just a YAML file. We built Truss to simplify the process of deploying AI models in...
Nov 22, 2024
Debugging just got a little easier! Now, when filtering logs, you can view surrounding events by simply clicking on an event's timestamp. The logs will expand to show events immediately before and...
Oct 31, 2024
As part of ongoing improvements to Baseten’s infrastructure platform, we’re working on giving you more flexibility in how resources are provisioned for each model deployment. In the interim, we’re...
Oct 15, 2024
We're excited to introduce canary deployments on Baseten, designed to phase in new deployments with minimal impact on production latency and uptime. When enabled for a model, Baseten gradually shifts...