Baseten Deprecates DeepSeek v3 0324 and GLM 4.6 Model APIs on May 1
Baseten has announced the deprecation of two model APIs while simultaneously launching discounted cache token pricing for all remaining model endpoints.
Baseten has announced the deprecation of two model APIs while simultaneously launching discounted cache token pricing for all remaining model endpoints.
Baseten will retire the DeepSeek v3 0324 and GLM 4.6 Model APIs at 5pm PT on May 1st, 2026. The deprecation was announced via the company's changelog on April 17th, giving developers roughly two weeks' notice before the cutoff. Any applications still routing requests through these endpoints will need to migrate before the deadline to avoid service disruption. On the pricing side, Baseten introduced cache token pricing for Model APIs effective April 16th, 2026. Cached input tokens now attract a discounted billing rate across all models, with the sole exception of GPT-OSS. The discounted rate applies automatically to the relevant portion of each request, requiring no manual configuration from users. A third changelog entry, published April 6th, added a quality-of-life improvement to Baseten's logs viewer. Users can now copy logs directly to their clipboard or download them as CSV or JSON files via a new export menu positioned next to the search box.