Tag: Tags:
-
Simon Willison’s Weblog: llm-fragments-go
Source URL: https://simonwillison.net/2025/Apr/10/llm-fragments-go/#atom-everything Source: Simon Willison’s Weblog Title: llm-fragments-go Feedly Summary: llm-fragments-go Filippo Valsorda released the first plugin by someone other than me that uses LLM’s new register_fragment_loaders() plugin hook I announced the other day. Install with llm install llm-fragments-go and then: You can feed the docs of a Go package into LLM using the…
-
Simon Willison’s Weblog: An LLM Query Understanding Service
Source URL: https://simonwillison.net/2025/Apr/9/an-llm-query-understanding-service/#atom-everything Source: Simon Willison’s Weblog Title: An LLM Query Understanding Service Feedly Summary: An LLM Query Understanding Service Doug Turnbull recently wrote about how all search is structured now: Many times, even a small open source LLM will be able to turn a search query into reasonable structure at relatively low cost. In…
-
Simon Willison’s Weblog: Model Context Protocol has prompt injection security problems
Source URL: https://simonwillison.net/2025/Apr/9/mcp-prompt-injection/#atom-everything Source: Simon Willison’s Weblog Title: Model Context Protocol has prompt injection security problems Feedly Summary: As more people start hacking around with implementations of MCP (the Model Context Protocol, a new standard for making tools available to LLM-powered systems) the security implications of tools built on that protocol are starting to come…
-
Simon Willison’s Weblog: Mistral Small 3.1 on Ollama
Source URL: https://simonwillison.net/2025/Apr/8/mistral-small-31-on-ollama/#atom-everything Source: Simon Willison’s Weblog Title: Mistral Small 3.1 on Ollama Feedly Summary: Mistral Small 3.1 on Ollama Mistral Small 3.1 (previously) is now available through Ollama, providing an easy way to run this multi-modal (vision) model on a Mac (and other platforms, though I haven’t tried them myself yet). I had to…
-
Simon Willison’s Weblog: Quoting lmarena.ai
Source URL: https://simonwillison.net/2025/Apr/8/lmaren/#atom-everything Source: Simon Willison’s Weblog Title: Quoting lmarena.ai Feedly Summary: We’ve seen questions from the community about the latest release of Llama-4 on Arena. To ensure full transparency, we’re releasing 2,000+ head-to-head battle results for public review. […] In addition, we’re also adding the HF version of Llama-4-Maverick to Arena, with leaderboard results…
-
Simon Willison’s Weblog: Long context support in LLM 0.24 using fragments and template plugins
Source URL: https://simonwillison.net/2025/Apr/7/long-context-llm/#atom-everything Source: Simon Willison’s Weblog Title: Long context support in LLM 0.24 using fragments and template plugins Feedly Summary: LLM 0.24 is now available with new features to help take advantage of the increasingly long input context supported by modern LLMs. (LLM is my command-line tool and Python library for interacting with LLMs,…
-
Simon Willison’s Weblog: Quoting Andriy Burkov
Source URL: https://simonwillison.net/2025/Apr/6/andriy-burkov/#atom-everything Source: Simon Willison’s Weblog Title: Quoting Andriy Burkov Feedly Summary: […] The disappointing releases of both GPT-4.5 and Llama 4 have shown that if you don’t train a model to reason with reinforcement learning, increasing its size no longer provides benefits. Reinforcement learning is limited only to domains where a reward can…
-
Simon Willison’s Weblog: Note on 5th April 2025
Source URL: https://simonwillison.net/2025/Apr/5/llama-4-notes/#atom-everything Source: Simon Willison’s Weblog Title: Note on 5th April 2025 Feedly Summary: Dropping a model release as significant as Llama 4 on a weekend is plain unfair! So far the best place to learn about the new model family is this post on the Meta AI blog. You can try them out…
-
Simon Willison’s Weblog: Quoting Ahmed Al-Dahle
Source URL: https://simonwillison.net/2025/Apr/5/llama-4/#atom-everything Source: Simon Willison’s Weblog Title: Quoting Ahmed Al-Dahle Feedly Summary: The Llama series have been re-designed to use state of the art mixture-of-experts (MoE) architecture and natively trained with multimodality. We’re dropping Llama 4 Scout & Llama 4 Maverick, and previewing Llama 4 Behemoth. 📌 Llama 4 Scout is highest performing small…
-
Simon Willison’s Weblog: Quoting Sam Altman
Source URL: https://simonwillison.net/2025/Apr/4/sam-altman/#atom-everything Source: Simon Willison’s Weblog Title: Quoting Sam Altman Feedly Summary: change of plans: we are going to release o3 and o4-mini after all, probably in a couple of weeks, and then do GPT-5 in a few months — Sam Altman Tags: sam-altman, generative-ai, openai, ai, llms AI Summary and Description: Yes Summary:…