Tag: .NET

  • Simon Willison’s Weblog: Models can prompt now

    Source URL: https://simonwillison.net/2025/Sep/14/models-can-prompt/#atom-everything Source: Simon Willison’s Weblog Title: Models can prompt now Feedly Summary: Here’s an interesting example of models incrementally improving over time: I am finding that today’s leading models are competent at writing prompts for themselves and each other. A year ago I was quite skeptical of the pattern where models are used…

  • Simon Willison’s Weblog: gpt-5 and gpt-5-mini rate limit updates

    Source URL: https://simonwillison.net/2025/Sep/12/gpt-5-rate-limits/#atom-everything Source: Simon Willison’s Weblog Title: gpt-5 and gpt-5-mini rate limit updates Feedly Summary: gpt-5 and gpt-5-mini rate limit updates OpenAI have increased the rate limits for their two main GPT-5 models. These look significant: gpt-5 Tier 1: 30K → 500K TPM (1.5M batch) Tier 2: 450K → 1M (3M batch) Tier 3:…

  • Simon Willison’s Weblog: Comparing the memory implementations of Claude and ChatGPT

    Source URL: https://simonwillison.net/2025/Sep/12/claude-memory/#atom-everything Source: Simon Willison’s Weblog Title: Comparing the memory implementations of Claude and ChatGPT Feedly Summary: Claude Memory: A Different Philosophy Shlok Khemani has been doing excellent work reverse-engineering LLM systems and documenting his discoveries. Last week he wrote about ChatGPT memory. This week it’s Claude. Claude’s memory system has two fundamental characteristics.…

  • Simon Willison’s Weblog: Defeating Nondeterminism in LLM Inference

    Source URL: https://simonwillison.net/2025/Sep/11/defeating-nondeterminism/#atom-everything Source: Simon Willison’s Weblog Title: Defeating Nondeterminism in LLM Inference Feedly Summary: Defeating Nondeterminism in LLM Inference A very common question I see about LLMs concerns why they can’t be made to deliver the same response to the same prompt by setting a fixed random number seed. Like many others I had…

  • Simon Willison’s Weblog: Claude API: Web fetch tool

    Source URL: https://simonwillison.net/2025/Sep/10/claude-web-fetch-tool/#atom-everything Source: Simon Willison’s Weblog Title: Claude API: Web fetch tool Feedly Summary: Claude API: Web fetch tool New in the Claude API: if you pass the web-fetch-2025-09-10 beta header you can add {“type": "web_fetch_20250910", "name": "web_fetch", "max_uses": 5} to your "tools" list and Claude will gain the ability to fetch content from…

  • Simon Willison’s Weblog: Recreating the Apollo AI adoption rate chart with GPT-5, Python and Pyodide

    Source URL: https://simonwillison.net/2025/Sep/9/apollo-ai-adoption/#atom-everything Source: Simon Willison’s Weblog Title: Recreating the Apollo AI adoption rate chart with GPT-5, Python and Pyodide Feedly Summary: Apollo Global Management’s “Chief Economist" Dr. Torsten Sløk released this interesting chart which appears to show a slowdown in AI adoption rates among large (>250 empoloyees) companies: Here’s the full description that accompanied…

  • Simon Willison’s Weblog: Anthropic status: Model output quality

    Source URL: https://simonwillison.net/2025/Sep/9/anthropic-model-output-quality/ Source: Simon Willison’s Weblog Title: Anthropic status: Model output quality Feedly Summary: Anthropic status: Model output quality Anthropic previously reported model serving bugs that affected Claude Opus 4 and 4.1 for 56.5 hours. They’ve now fixed additional bugs affecting “a small percentage" of Sonnet 4 requests for almost a month, plus a…

  • Simon Willison’s Weblog: Load Llama-3.2 WebGPU in your browser from a local folder

    Source URL: https://simonwillison.net/2025/Sep/8/webgpu-local-folder/#atom-everything Source: Simon Willison’s Weblog Title: Load Llama-3.2 WebGPU in your browser from a local folder Feedly Summary: Load Llama-3.2 WebGPU in your browser from a local folder Inspired by a comment on Hacker News I decided to see if it was possible to modify the transformers.js-examples/tree/main/llama-3.2-webgpu Llama 3.2 chat demo (online here,…