Tag: Gemma

  • Simon Willison’s Weblog: Vision Language Models (Better, Faster, Stronger)

    Source URL: https://simonwillison.net/2025/May/13/vision-language-models/#atom-everything Source: Simon Willison’s Weblog Title: Vision Language Models (Better, Faster, Stronger) Feedly Summary: Vision Language Models (Better, Faster, Stronger) Extremely useful review of the last year in vision and multi-modal LLMs. So much has happened! I’m particularly excited about the range of small open weight vision models that are now available. Models…

  • Simon Willison’s Weblog: Trying out llama.cpp’s new vision support

    Source URL: https://simonwillison.net/2025/May/10/llama-cpp-vision/#atom-everything Source: Simon Willison’s Weblog Title: Trying out llama.cpp’s new vision support Feedly Summary: This llama.cpp server vision support via libmtmd pull request – via Hacker News – was merged earlier today. The PR finally adds full support for vision models to the excellent llama.cpp project. It’s documented on this page, but the…

  • Simon Willison’s Weblog: What people get wrong about the leading Chinese open models: Adoption and censorship

    Source URL: https://simonwillison.net/2025/May/6/what-people-get-wrong-about-the-leading-chinese-models/#atom-everything Source: Simon Willison’s Weblog Title: What people get wrong about the leading Chinese open models: Adoption and censorship Feedly Summary: What people get wrong about the leading Chinese open models: Adoption and censorship While I’ve been enjoying trying out Alibaba’s Qwen 3 a lot recently, Nathan Lambert focuses on the elephant in…

  • Cloud Blog: Create chatbots that speak different languages with Gemini, Gemma, Translation LLM, and Model Context Protocol

    Source URL: https://cloud.google.com/blog/products/ai-machine-learning/build-multilingual-chatbots-with-gemini-gemma-and-mcp/ Source: Cloud Blog Title: Create chatbots that speak different languages with Gemini, Gemma, Translation LLM, and Model Context Protocol Feedly Summary: Your customers might not all speak the same language. If you operate internationally or serve a diverse customer base, you need your chatbot to meet them where they are – whether…

  • Simon Willison’s Weblog: Note on 20th April 2025

    Source URL: https://simonwillison.net/2025/Apr/20/janky-license/#atom-everything Source: Simon Willison’s Weblog Title: Note on 20th April 2025 Feedly Summary: Now that Llama has very real competition in open weight models (Gemma 3, latest Mistrals, DeepSeek, Qwen) I think their janky license is becoming much more of a liability for them. It’s just limiting enough that it could be the…

  • Simon Willison’s Weblog: llm-fragments-github 0.2

    Source URL: https://simonwillison.net/2025/Apr/20/llm-fragments-github/#atom-everything Source: Simon Willison’s Weblog Title: llm-fragments-github 0.2 Feedly Summary: llm-fragments-github 0.2 I upgraded my llm-fragments-github plugin to add a new fragment type called issue. It lets you pull the entire content of a GitHub issue thread into your prompt as a concatenated Markdown file. (If you haven’t seen fragments before I introduced…

  • Simon Willison’s Weblog: Gemma 3 QAT Models

    Source URL: https://simonwillison.net/2025/Apr/19/gemma-3-qat-models/ Source: Simon Willison’s Weblog Title: Gemma 3 QAT Models Feedly Summary: Gemma 3 QAT Models Interesting release from Google, as a follow-up to Gemma 3 from last month: To make Gemma 3 even more accessible, we are announcing new versions optimized with Quantization-Aware Training (QAT) that dramatically reduces memory requirements while maintaining…

  • Slashdot: Microsoft Researchers Develop Hyper-Efficient AI Model That Can Run On CPUs

    Source URL: https://slashdot.org/story/25/04/17/2224205/microsoft-researchers-develop-hyper-efficient-ai-model-that-can-run-on-cpus?utm_source=rss1.0mainlinkanon&utm_medium=feed Source: Slashdot Title: Microsoft Researchers Develop Hyper-Efficient AI Model That Can Run On CPUs Feedly Summary: AI Summary and Description: Yes Summary: Microsoft has launched BitNet b1.58 2B4T, a highly efficient 1-bit AI model featuring 2 billion parameters, optimized for CPU use and accessible under an MIT license. It surpasses competitors in…