Tag: llm-pricing

  • Simon Willison’s Weblog: GPT-5 pro

    Source URL: https://simonwillison.net/2025/Oct/6/gpt-5-pro/ Source: Simon Willison’s Weblog Title: GPT-5 pro Feedly Summary: GPT-5 pro Here’s OpenAI’s model documentation for their GPT-5 pro model, released to their API today at their DevDay event. It has similar base characteristics to GPT-5: both share a September 30, 2024 knowledge cutoff and 400,000 context limit. GPT-5 pro has maximum…

  • Simon Willison’s Weblog: Claude Sonnet 4.5 is probably the "best coding model in the world" (at least for now)

    Source URL: https://simonwillison.net/2025/Sep/29/claude-sonnet-4-5/ Source: Simon Willison’s Weblog Title: Claude Sonnet 4.5 is probably the "best coding model in the world" (at least for now) Feedly Summary: Anthropic released Claude Sonnet 4.5 today, with a very bold set of claims: Claude Sonnet 4.5 is the best coding model in the world. It’s the strongest model for…

  • Simon Willison’s Weblog: Grok 4 Fast

    Source URL: https://simonwillison.net/2025/Sep/20/grok-4-fast/ Source: Simon Willison’s Weblog Title: Grok 4 Fast Feedly Summary: Grok 4 Fast New hosted reasoning model from xAI that’s designed to be fast and extremely competitive on price. It has a 2 million token context window and “was trained end-to-end with tool-use reinforcement learning". It’s priced at $0.20/million input tokens and…

  • Simon Willison’s Weblog: gpt-5 and gpt-5-mini rate limit updates

    Source URL: https://simonwillison.net/2025/Sep/12/gpt-5-rate-limits/#atom-everything Source: Simon Willison’s Weblog Title: gpt-5 and gpt-5-mini rate limit updates Feedly Summary: gpt-5 and gpt-5-mini rate limit updates OpenAI have increased the rate limits for their two main GPT-5 models. These look significant: gpt-5 Tier 1: 30K → 500K TPM (1.5M batch) Tier 2: 450K → 1M (3M batch) Tier 3:…

  • Simon Willison’s Weblog: Load Llama-3.2 WebGPU in your browser from a local folder

    Source URL: https://simonwillison.net/2025/Sep/8/webgpu-local-folder/#atom-everything Source: Simon Willison’s Weblog Title: Load Llama-3.2 WebGPU in your browser from a local folder Feedly Summary: Load Llama-3.2 WebGPU in your browser from a local folder Inspired by a comment on Hacker News I decided to see if it was possible to modify the transformers.js-examples/tree/main/llama-3.2-webgpu Llama 3.2 chat demo (online here,…

  • Simon Willison’s Weblog: Introducing gpt-realtime

    Source URL: https://simonwillison.net/2025/Sep/1/introducing-gpt-realtime/#atom-everything Source: Simon Willison’s Weblog Title: Introducing gpt-realtime Feedly Summary: Introducing gpt-realtime Released a few days ago (August 28th), gpt-realtime is OpenAI’s new “most advanced speech-to-speech model". It looks like this is a replacement for the older gpt-4o-realtime-preview model that was released last October. This is a slightly confusing release. The previous realtime…

  • Simon Willison’s Weblog: Claude Sonnet 4 now supports 1M tokens of context

    Source URL: https://simonwillison.net/2025/Aug/12/claude-sonnet-4-1m/ Source: Simon Willison’s Weblog Title: Claude Sonnet 4 now supports 1M tokens of context Feedly Summary: Claude Sonnet 4 now supports 1M tokens of context Gemini and OpenAI both have million token models, so it’s good to see Anthropic catching up. This is 5x the previous 200,000 context length limit of the…

  • Simon Willison’s Weblog: Claude Opus 4.1

    Source URL: https://simonwillison.net/2025/Aug/5/claude-opus-41/ Source: Simon Willison’s Weblog Title: Claude Opus 4.1 Feedly Summary: Claude Opus 4.1 Surprise new model from Anthropic today – Claude Opus 4.1, which they describe as “a drop-in replacement for Opus 4". My favorite thing about this model is the version number – treating this as a .1 version increment looks…

  • Simon Willison’s Weblog: Faster inference

    Source URL: https://simonwillison.net/2025/Aug/1/faster-inference/ Source: Simon Willison’s Weblog Title: Faster inference Feedly Summary: Two interesting examples of inference speed as a flagship feature of LLM services today. First, Cerebras announced two new monthly plans for their extremely high speed hosted model service: Cerebras Code Pro ($50/month, 1,000 messages a day) and Cerebras Code Max ($200/month, 5,000/day).…

  • Simon Willison’s Weblog: Quoting Anthropic

    Source URL: https://simonwillison.net/2025/Jul/28/anthropic/ Source: Simon Willison’s Weblog Title: Quoting Anthropic Feedly Summary: We’re rolling out new weekly rate limits for Claude Pro and Max in late August. We estimate they’ll apply to less than 5% of subscribers based on current usage. […] Some of the biggest Claude Code fans are running it continuously in the…