GPT-5.4 & fast mode live #552
Replies: 3 comments 1 reply
-
|
Working great! I have a couple issues I want to bring up, I'll make issues or PRs tomorrow. TUI selection for compaction. I believe there are some config values you can put in for this, but 5.4 does really well even with larger context use. I was testing close to the 1m limit. There is an issue where large console output from a process Every Code is watching that causes the session to basically become unresponsive. Just a few minutes ago I had a session I really wanted to recover so we found the problem. Anyway, 100% on gpt-5.4 and your choice to hide the other models. I'll write up something tomorrow for these, unless you fix them before I have time, as you usually do. GUI/remote view is coming along. I've had to trash and start over a few times. I did have it fully working at one point with a decent UI, and I really like it. It's so nice to be able to step away and answer quick messages or questions or just check on progress. |
Beta Was this translation helpful? Give feedback.
-
|
@cbusillo Yeah good ideas! I'm adding in the 1M compaction toggle. I'm also adding in an "1M Auto" model (probably the new default) which will automatically pick the best time to compact between 100k and 1M. Basically and the start of each turn, it'll decide based on the turn (new user input & prior convo) if the turn will use most of the prior context and/or if it'll run out of context before completing. Also keeping in mind the 2x cost of > 290k. So this will allow sessions to go above the 290k when needed, but avoid it most of them time. I've had quite a few sessions keep looping on compaction on difficult tasks and have the ability to automatically into 1M when useful. |
Beta Was this translation helpful? Give feedback.
-
|
Are you aware of the leaking into assistant messages? I assume you want those cleaned up somehow? Or do you not use memory. I love new features sometimes only because they are new features and tend to turn everything on. I did turn off fast mode. I already burn through two Pro accounts a week in 4-6 days. |
Beta Was this translation helpful? Give feedback.
Uh oh!
There was an error while loading. Please reload this page.
-
I've cleaned up the /model list as well to just show GPT-5.4, GPT-5.3-Codex and GPT-5.3-Codex-Spark.
You can still launch Code with the model param (e.g. -m gpt-5.2) to access earlier models.
Open to feedback if you think this is too early!
Beta Was this translation helpful? Give feedback.
All reactions