Tag: token limit

  • Simon Willison’s Weblog: GPT-5 pro

    Source URL: https://simonwillison.net/2025/Oct/6/gpt-5-pro/ Source: Simon Willison’s Weblog Title: GPT-5 pro Feedly Summary: GPT-5 pro Here’s OpenAI’s model documentation for their GPT-5 pro model, released to their API today at their DevDay event. It has similar base characteristics to GPT-5: both share a September 30, 2024 knowledge cutoff and 400,000 context limit. GPT-5 pro has maximum…

  • Simon Willison’s Weblog: Introducing gpt-realtime

    Source URL: https://simonwillison.net/2025/Sep/1/introducing-gpt-realtime/#atom-everything Source: Simon Willison’s Weblog Title: Introducing gpt-realtime Feedly Summary: Introducing gpt-realtime Released a few days ago (August 28th), gpt-realtime is OpenAI’s new “most advanced speech-to-speech model". It looks like this is a replacement for the older gpt-4o-realtime-preview model that was released last October. This is a slightly confusing release. The previous realtime…

  • Simon Willison’s Weblog: How to Fix Your Context

    Source URL: https://simonwillison.net/2025/Jun/29/how-to-fix-your-context/#atom-everything Source: Simon Willison’s Weblog Title: How to Fix Your Context Feedly Summary: How to Fix Your Context Drew Breunig has been publishing some very detailed notes on context engineering recently. In How Long Contexts Fail he described four common patterns for context rot, which he summarizes like so: Context Poisoning: When a…