Tag: llm-pricing

  • Simon Willison’s Weblog: GPT-4.1: Three new million token input models from OpenAI, including their cheapest model yet

    Source URL: https://simonwillison.net/2025/Apr/14/gpt-4-1/ Source: Simon Willison’s Weblog Title: GPT-4.1: Three new million token input models from OpenAI, including their cheapest model yet Feedly Summary: OpenAI introduced three new models this morning: GPT-4.1, GPT-4.1 mini and GPT-4.1 nano. These are API-only models right now, not available through the ChatGPT interface (though you can try them out…

  • Simon Willison’s Weblog: LLM pricing calculator (updated)

    Source URL: https://simonwillison.net/2025/Apr/10/llm-pricing-calculator/#atom-everything Source: Simon Willison’s Weblog Title: LLM pricing calculator (updated) Feedly Summary: LLM pricing calculator (updated) I updated my LLM pricing calculator this morning (Claude transcript) to show the prices of various hosted models in a sorted table, defaulting to lowest price first. Amazon Nova and Google Gemini continue to dominate the lower…

  • Simon Willison’s Weblog: Gemini 2.5 Pro Preview pricing

    Source URL: https://simonwillison.net/2025/Apr/4/gemini-25-pro-pricing/ Source: Simon Willison’s Weblog Title: Gemini 2.5 Pro Preview pricing Feedly Summary: Gemini 2.5 Pro Preview pricing Google’s Gemini 2.5 Pro is currently the top model on LM Arena and, from my own testing, a superb model for OCR, audio transcription and long-context coding. You can now pay for it! The new…

  • Simon Willison’s Weblog: OpenAI platform: o1-pro

    Source URL: https://simonwillison.net/2025/Mar/19/o1-pro/ Source: Simon Willison’s Weblog Title: OpenAI platform: o1-pro Feedly Summary: OpenAI platform: o1-pro OpenAI have a new most-expensive model: o1-pro can now be accessed through their API at a hefty $150/million tokens for input and $600/million tokens for output. That’s 10x the price of their o1 and o1-preview models and a full…

  • Simon Willison’s Weblog: Gemini 2.0 Flash and Flash-Lite

    Source URL: https://simonwillison.net/2025/Feb/25/gemini-20-flash-and-flash-lite/ Source: Simon Willison’s Weblog Title: Gemini 2.0 Flash and Flash-Lite Feedly Summary: Gemini 2.0 Flash and Flash-Lite Gemini 2.0 Flash-Lite is now generally available – previously it was available just as a preview – and has announced pricing. The model is $0.075/million input tokens and $0.030/million output – the same price as…

  • Simon Willison’s Weblog: Quoting Sam Altman

    Source URL: https://simonwillison.net/2025/Feb/9/sam-altman/#atom-everything Source: Simon Willison’s Weblog Title: Quoting Sam Altman Feedly Summary: The cost to use a given level of AI falls about 10x every 12 months, and lower prices lead to much more use. You can see this in the token cost from GPT-4 in early 2023 to GPT-4o in mid-2024, where the…

  • Simon Willison’s Weblog: Gemini 2.0 is now available to everyone

    Source URL: https://simonwillison.net/2025/Feb/5/gemini-2/ Source: Simon Willison’s Weblog Title: Gemini 2.0 is now available to everyone Feedly Summary: Gemini 2.0 is now available to everyone Big new Gemini 2.0 releases today: Gemini 2.0 Pro (Experimental) is Google’s “best model yet for coding performance and complex prompts" – currently available as a free preview. Gemini 2.0 Flash…

  • Simon Willison’s Weblog: DeepSeek_V3.pdf

    Source URL: https://simonwillison.net/2024/Dec/26/deepseek-v3/#atom-everything Source: Simon Willison’s Weblog Title: DeepSeek_V3.pdf Feedly Summary: DeepSeek_V3.pdf The DeepSeek v3 paper (and model card) are out, after yesterday’s mysterious release of the undocumented model weights. Plenty of interesting details in here. The model pre-trained on 14.8 trillion “high-quality and diverse tokens" (not otherwise documented). Following this, we conduct post-training, including…

  • Simon Willison’s Weblog: Claude 3.5 Haiku price drops by 20%

    Source URL: https://simonwillison.net/2024/Dec/5/claude-35-haiku-price-drops-by-20/#atom-everything Source: Simon Willison’s Weblog Title: Claude 3.5 Haiku price drops by 20% Feedly Summary: Claude 3.5 Haiku price drops by 20% Buried in this otherwise quite dry post about Anthropic’s ongoing partnership with AWS: To make this model even more accessible for a wide range of use cases, we’re lowering the price…

  • Simon Willison’s Weblog: First impressions of the new Amazon Nova LLMs (via a new llm-bedrock plugin)

    Source URL: https://simonwillison.net/2024/Dec/4/amazon-nova/ Source: Simon Willison’s Weblog Title: First impressions of the new Amazon Nova LLMs (via a new llm-bedrock plugin) Feedly Summary: Amazon released three new Large Language Models yesterday at their AWS re:Invent conference. The new model family is called Amazon Nova and comes in three sizes: Micro, Lite and Pro. I built…