Tag: token limit
-
Simon Willison’s Weblog: GPT-5 pro
Source URL: https://simonwillison.net/2025/Oct/6/gpt-5-pro/ Source: Simon Willison’s Weblog Title: GPT-5 pro Feedly Summary: GPT-5 pro Here’s OpenAI’s model documentation for their GPT-5 pro model, released to their API today at their DevDay event. It has similar base characteristics to GPT-5: both share a September 30, 2024 knowledge cutoff and 400,000 context limit. GPT-5 pro has maximum…
-
Simon Willison’s Weblog: Introducing gpt-realtime
Source URL: https://simonwillison.net/2025/Sep/1/introducing-gpt-realtime/#atom-everything Source: Simon Willison’s Weblog Title: Introducing gpt-realtime Feedly Summary: Introducing gpt-realtime Released a few days ago (August 28th), gpt-realtime is OpenAI’s new “most advanced speech-to-speech model". It looks like this is a replacement for the older gpt-4o-realtime-preview model that was released last October. This is a slightly confusing release. The previous realtime…
-
Simon Willison’s Weblog: How to Fix Your Context
Source URL: https://simonwillison.net/2025/Jun/29/how-to-fix-your-context/#atom-everything Source: Simon Willison’s Weblog Title: How to Fix Your Context Feedly Summary: How to Fix Your Context Drew Breunig has been publishing some very detailed notes on context engineering recently. In How Long Contexts Fail he described four common patterns for context rot, which he summarizes like so: Context Poisoning: When a…
-
Simon Willison’s Weblog: Quoting Workaccount2 on Hacker News
Source URL: https://simonwillison.net/2025/Jun/18/context-rot/#atom-everything Source: Simon Willison’s Weblog Title: Quoting Workaccount2 on Hacker News Feedly Summary: They poison their own context. Maybe you can call it context rot, where as context grows and especially if it grows with lots of distractions and dead ends, the output quality falls off rapidly. Even with good context the rot…