Tag: Tags:

  • Simon Willison’s Weblog: llm-fragments-go

    Source URL: https://simonwillison.net/2025/Apr/10/llm-fragments-go/#atom-everything Source: Simon Willison’s Weblog Title: llm-fragments-go Feedly Summary: llm-fragments-go Filippo Valsorda released the first plugin by someone other than me that uses LLM’s new register_fragment_loaders() plugin hook I announced the other day. Install with llm install llm-fragments-go and then: You can feed the docs of a Go package into LLM using the…

  • Simon Willison’s Weblog: An LLM Query Understanding Service

    Source URL: https://simonwillison.net/2025/Apr/9/an-llm-query-understanding-service/#atom-everything Source: Simon Willison’s Weblog Title: An LLM Query Understanding Service Feedly Summary: An LLM Query Understanding Service Doug Turnbull recently wrote about how all search is structured now: Many times, even a small open source LLM will be able to turn a search query into reasonable structure at relatively low cost. In…

  • Simon Willison’s Weblog: Mistral Small 3.1 on Ollama

    Source URL: https://simonwillison.net/2025/Apr/8/mistral-small-31-on-ollama/#atom-everything Source: Simon Willison’s Weblog Title: Mistral Small 3.1 on Ollama Feedly Summary: Mistral Small 3.1 on Ollama Mistral Small 3.1 (previously) is now available through Ollama, providing an easy way to run this multi-modal (vision) model on a Mac (and other platforms, though I haven’t tried them myself yet). I had to…

  • Simon Willison’s Weblog: Quoting lmarena.ai

    Source URL: https://simonwillison.net/2025/Apr/8/lmaren/#atom-everything Source: Simon Willison’s Weblog Title: Quoting lmarena.ai Feedly Summary: We’ve seen questions from the community about the latest release of Llama-4 on Arena. To ensure full transparency, we’re releasing 2,000+ head-to-head battle results for public review. […] In addition, we’re also adding the HF version of Llama-4-Maverick to Arena, with leaderboard results…

  • Simon Willison’s Weblog: Quoting Andriy Burkov

    Source URL: https://simonwillison.net/2025/Apr/6/andriy-burkov/#atom-everything Source: Simon Willison’s Weblog Title: Quoting Andriy Burkov Feedly Summary: […] The disappointing releases of both GPT-4.5 and Llama 4 have shown that if you don’t train a model to reason with reinforcement learning, increasing its size no longer provides benefits. Reinforcement learning is limited only to domains where a reward can…

  • Simon Willison’s Weblog: Note on 5th April 2025

    Source URL: https://simonwillison.net/2025/Apr/5/llama-4-notes/#atom-everything Source: Simon Willison’s Weblog Title: Note on 5th April 2025 Feedly Summary: Dropping a model release as significant as Llama 4 on a weekend is plain unfair! So far the best place to learn about the new model family is this post on the Meta AI blog. You can try them out…

  • Simon Willison’s Weblog: Quoting Ahmed Al-Dahle

    Source URL: https://simonwillison.net/2025/Apr/5/llama-4/#atom-everything Source: Simon Willison’s Weblog Title: Quoting Ahmed Al-Dahle Feedly Summary: The Llama series have been re-designed to use state of the art mixture-of-experts (MoE) architecture and natively trained with multimodality. We’re dropping Llama 4 Scout & Llama 4 Maverick, and previewing Llama 4 Behemoth. 📌 Llama 4 Scout is highest performing small…

  • Simon Willison’s Weblog: Quoting Sam Altman

    Source URL: https://simonwillison.net/2025/Apr/4/sam-altman/#atom-everything Source: Simon Willison’s Weblog Title: Quoting Sam Altman Feedly Summary: change of plans: we are going to release o3 and o4-mini after all, probably in a couple of weeks, and then do GPT-5 in a few months — Sam Altman Tags: sam-altman, generative-ai, openai, ai, llms AI Summary and Description: Yes Summary:…