Tag: cpp

  • Simon Willison’s Weblog: Introducing Gemma 3n: The developer guide

    Source URL: https://simonwillison.net/2025/Jun/26/gemma-3n/ Source: Simon Willison’s Weblog Title: Introducing Gemma 3n: The developer guide Feedly Summary: Introducing Gemma 3n: The developer guide Extremely consequential new open weights model release from Google today: Multimodal by design: Gemma 3n natively supports image, audio, video, and text inputs and text outputs. Optimized for on-device: Engineered with a focus…

  • Docker: Behind the scenes: How we designed Docker Model Runner and what’s next

    Source URL: https://www.docker.com/blog/behind-the-scenes-how-we-designed-docker-model-runner-and-whats-next/ Source: Docker Title: Behind the scenes: How we designed Docker Model Runner and what’s next Feedly Summary: The last few years have made it clear that AI models will continue to be a fundamental component of many applications. The catch is that they’re also a fundamentally different type of component, with complex…

  • Simon Willison’s Weblog: llm-llama-server 0.2

    Source URL: https://simonwillison.net/2025/May/28/llama-server-tools/ Source: Simon Willison’s Weblog Title: llm-llama-server 0.2 Feedly Summary: llm-llama-server 0.2 Here’s a second option for using LLM’s new tool support against local models (the first was via llm-ollama). It turns out the llama.cpp ecosystem has pretty robust OpenAI-compatible tool support already, so my llm-llama-server plugin only needed a quick upgrade to…

  • Simon Willison’s Weblog: Large Language Models can run tools in your terminal with LLM 0.26

    Source URL: https://simonwillison.net/2025/May/27/llm-tools/ Source: Simon Willison’s Weblog Title: Large Language Models can run tools in your terminal with LLM 0.26 Feedly Summary: LLM 0.26 is out with the biggest new feature since I started the project: support for tools. You can now use the LLM CLI tool – and Python library – to grant LLMs…

  • Simon Willison’s Weblog: Trying out llama.cpp’s new vision support

    Source URL: https://simonwillison.net/2025/May/10/llama-cpp-vision/#atom-everything Source: Simon Willison’s Weblog Title: Trying out llama.cpp’s new vision support Feedly Summary: This llama.cpp server vision support via libmtmd pull request – via Hacker News – was merged earlier today. The PR finally adds full support for vision models to the excellent llama.cpp project. It’s documented on this page, but the…