Tag: llms

  • Simon Willison’s Weblog: llm-prices.com

    Source URL: https://simonwillison.net/2025/May/7/llm-prices/#atom-everything Source: Simon Willison’s Weblog Title: llm-prices.com Feedly Summary: llm-prices.com I’ve been maintaining a simple LLM pricing calculator since October last year. I finally decided to split it out to its own domain name (previously it was hosted at tools.simonwillison.net/llm-prices), running on Cloudflare Pages. The site runs out of my simonw/llm-prices GitHub repository.…

  • Simon Willison’s Weblog: Saying "hi" to Microsoft’s Phi-4-reasoning

    Source URL: https://simonwillison.net/2025/May/6/phi-4-reasoning/#atom-everything Source: Simon Willison’s Weblog Title: Saying "hi" to Microsoft’s Phi-4-reasoning Feedly Summary: Microsoft released a new sub-family of models a few days ago: Phi-4 reasoning. They introduced them in this blog post celebrating a year since the release of Phi-3: Today, we are excited to introduce Phi-4-reasoning, Phi-4-reasoning-plus, and Phi-4-mini-reasoning – marking…

  • Simon Willison’s Weblog: Gemini 2.5 Pro Preview: even better coding performance

    Source URL: https://simonwillison.net/2025/May/6/gemini-25-pro-preview/#atom-everything Source: Simon Willison’s Weblog Title: Gemini 2.5 Pro Preview: even better coding performance Feedly Summary: Gemini 2.5 Pro Preview: even better coding performance New Gemini 2.5 Pro “Google I/O edition" model, released a few weeks ahead of that annual developer conference. They claim even better frontend coding performance, highlighting their #1 ranking…

  • Simon Willison’s Weblog: Quoting Daniel Stenberg

    Source URL: https://simonwillison.net/2025/May/6/daniel-stenberg/#atom-everything Source: Simon Willison’s Weblog Title: Quoting Daniel Stenberg Feedly Summary: That’s it. I’ve had it. I’m putting my foot down on this craziness. 1. Every reporter submitting security reports on #Hackerone for #curl now needs to answer this question: “Did you use an AI to find the problem or generate this submission?"…

  • Simon Willison’s Weblog: What people get wrong about the leading Chinese open models: Adoption and censorship

    Source URL: https://simonwillison.net/2025/May/6/what-people-get-wrong-about-the-leading-chinese-models/#atom-everything Source: Simon Willison’s Weblog Title: What people get wrong about the leading Chinese open models: Adoption and censorship Feedly Summary: What people get wrong about the leading Chinese open models: Adoption and censorship While I’ve been enjoying trying out Alibaba’s Qwen 3 a lot recently, Nathan Lambert focuses on the elephant in…

  • Simon Willison’s Weblog: Quoting Max Woolf

    Source URL: https://simonwillison.net/2025/May/5/max-woolf/#atom-everything Source: Simon Willison’s Weblog Title: Quoting Max Woolf Feedly Summary: Two things can be true simultaneously: (a) LLM provider cost economics are too negative to return positive ROI to investors, and (b) LLMs are useful for solving problems that are meaningful and high impact, albeit not to the AGI hype that would…

  • Simon Willison’s Weblog: Feed a video to a vision LLM as a sequence of JPEG frames on the CLI (also LLM 0.25)

    Source URL: https://simonwillison.net/2025/May/5/llm-video-frames/#atom-everything Source: Simon Willison’s Weblog Title: Feed a video to a vision LLM as a sequence of JPEG frames on the CLI (also LLM 0.25) Feedly Summary: The new llm-video-frames plugin can turn a video file into a sequence of JPEG frames and feed them directly into a long context vision LLM such…

  • Simon Willison’s Weblog: Quoting Arvind Narayanan

    Source URL: https://simonwillison.net/2025/May/5/arvind-narayanan/#atom-everything Source: Simon Willison’s Weblog Title: Quoting Arvind Narayanan Feedly Summary: [On using generative AI for work despite the risk of errors:] AI is helpful despite being error-prone if it is faster to verify the output than it is to do the work yourself. For example, if you’re using it to find a…

  • Simon Willison’s Weblog: Dummy’s Guide to Modern LLM Sampling

    Source URL: https://simonwillison.net/2025/May/4/llm-sampling/#atom-everything Source: Simon Willison’s Weblog Title: Dummy’s Guide to Modern LLM Sampling Feedly Summary: Dummy’s Guide to Modern LLM Sampling This is an extremely useful, detailed set of explanations by @AlpinDale covering the various different sampling strategies used by modern LLMs. LLMs return a set of next-token probabilities for every token in their…

  • Simon Willison’s Weblog: Qwen3-8B

    Source URL: https://simonwillison.net/2025/May/2/qwen3-8b/#atom-everything Source: Simon Willison’s Weblog Title: Qwen3-8B Feedly Summary: Having tried a few of the Qwen 3 models now my favorite is a bit of a surprise to me: I’m really enjoying Qwen3-8B. I’ve been running prompts through the MLX 4bit quantized version, mlx-community/Qwen3-8B-4bit. I’m using llm-mlx like this: llm install llm-mlx llm…