Tag: ollama

  • Simon Willison’s Weblog: An LLM Query Understanding Service

    Source URL: https://simonwillison.net/2025/Apr/9/an-llm-query-understanding-service/#atom-everything Source: Simon Willison’s Weblog Title: An LLM Query Understanding Service Feedly Summary: An LLM Query Understanding Service Doug Turnbull recently wrote about how all search is structured now: Many times, even a small open source LLM will be able to turn a search query into reasonable structure at relatively low cost. In…

  • Cloud Blog: Introducing Firebase Studio and agentic developer tools to build with Gemini

    Source URL: https://cloud.google.com/blog/products/application-development/firebase-studio-lets-you-build-full-stack-ai-apps-with-gemini/ Source: Cloud Blog Title: Introducing Firebase Studio and agentic developer tools to build with Gemini Feedly Summary: Millions of developers use Firebase to engage their users, powering over 70 billion instances of apps every day, everywhere — from mobile devices and web browsers, to embedded platforms and agentic experiences. But full-stack development…

  • Simon Willison’s Weblog: Mistral Small 3.1 on Ollama

    Source URL: https://simonwillison.net/2025/Apr/8/mistral-small-31-on-ollama/#atom-everything Source: Simon Willison’s Weblog Title: Mistral Small 3.1 on Ollama Feedly Summary: Mistral Small 3.1 on Ollama Mistral Small 3.1 (previously) is now available through Ollama, providing an easy way to run this multi-modal (vision) model on a Mac (and other platforms, though I haven’t tried them myself yet). I had to…

  • Simon Willison’s Weblog: Function calling with Gemma

    Source URL: https://simonwillison.net/2025/Mar/26/function-calling-with-gemma/#atom-everything Source: Simon Willison’s Weblog Title: Function calling with Gemma Feedly Summary: Function calling with Gemma Google’s Gemma 3 model (the 27B variant is particularly capable, I’ve been trying it out via Ollama) supports function calling exclusively through prompt engineering. The official documentation describes two recommended prompts – both of them suggest that…

  • Simon Willison’s Weblog: Mistral Small 3.1

    Source URL: https://simonwillison.net/2025/Mar/17/mistral-small-31/#atom-everything Source: Simon Willison’s Weblog Title: Mistral Small 3.1 Feedly Summary: Mistral Small 3.1 Mistral Small 3 came out in January and was a notable, genuinely excellent local model that used an Apache 2.0 license. Mistral Small 3.1 offers a significant improvement: it’s multi-modal (images) and has an increased 128,000 token context length,…

  • Simon Willison’s Weblog: Notes on Google’s Gemma 3

    Source URL: https://simonwillison.net/2025/Mar/12/gemma-3/ Source: Simon Willison’s Weblog Title: Notes on Google’s Gemma 3 Feedly Summary: Google’s Gemma team released an impressive new model today (under their not-open-source Gemma license). Gemma 3 comes in four sizes – 1B, 4B, 12B, and 27B – and while 1B is text-only the larger three models are all multi-modal for…

  • Simon Willison’s Weblog: Notes on Google’s Gemma 3

    Source URL: https://simonwillison.net/2025/Mar/12/notes-on-googles-gemma-3/ Source: Simon Willison’s Weblog Title: Notes on Google’s Gemma 3 Feedly Summary: Google’s Gemma team released an impressive new model today (under their not-open-source Gemma license). Gemma 3 comes in four sizes – 1B, 4B, 12B, and 27B – and while 1B is text-only the larger three models are all multi-modal for…

  • Hacker News: Gemma3 – The current strongest model that fits on a single GPU

    Source URL: https://ollama.com/library/gemma3 Source: Hacker News Title: Gemma3 – The current strongest model that fits on a single GPU Feedly Summary: Comments AI Summary and Description: Yes **Summary:** The text discusses the features and capabilities of the Gemma 3 models developed by Google, which are built on Gemini technology and designed for multimodal tasks. Their…

  • Cloud Blog: How to deploy serverless AI with Gemma 3 on Cloud Run

    Source URL: https://cloud.google.com/blog/products/ai-machine-learning/serverless-ai-with-gemma-3-on-cloud-run/ Source: Cloud Blog Title: How to deploy serverless AI with Gemma 3 on Cloud Run Feedly Summary: Today, we introduced Gemma 3, a family of lightweight, open models built with the cutting-edge technology behind Gemini 2.0. The Gemma 3 family of models have been designed for speed and portability, empowering developers to…

  • Hacker News: A Practical Guide to Running Local LLMs

    Source URL: https://spin.atomicobject.com/running-local-llms/ Source: Hacker News Title: A Practical Guide to Running Local LLMs Feedly Summary: Comments AI Summary and Description: Yes Summary: The text discusses the intricacies of running local large language models (LLMs), emphasizing their applications in privacy-critical situations and the potential benefits of various tools like Ollama and Llama.cpp. It provides insights…