AI: A single prompt alone consume almost all of the 128k tokens of gpt5/sonnet4 #37274
UdittLamba
started this conversation in
LLMs and Zed Agent
Replies: 0 comments
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Uh oh!
There was an error while loading. Please reload this page.
-
Summary
Previously, a single 128k context lasted for multiple prompts, but now it is a single response for a single prompt that consumes most of the tokens. Continuing to a new thread also doesn't;t help, as the "new from summary" is bugged out as well
Description
Steps to trigger the problem:
Expected Behavior: single context should last for multiple complex prompts
Actual Behavior: it barely lasts a single prompt after the last update
Model Provider Details
Other Details (MCPs, other settings, etc): none
Zed Version and System Specs
Zed: v0.201.8 (Zed)
OS: Linux Wayland manjaro unknown
Memory: 31.1 GiB
Architecture: x86_64
GPU: AMD Radeon RX 7900 XTX (RADV NAVI31) || radv || Mesa 25.2.1-arch1.4
Beta Was this translation helpful? Give feedback.
All reactions