Tag: ollama

  • Simon Willison’s Weblog: Notes on Google’s Gemma 3

    Source URL: https://simonwillison.net/2025/Mar/12/notes-on-googles-gemma-3/ Source: Simon Willison’s Weblog Title: Notes on Google’s Gemma 3 Feedly Summary: Google’s Gemma team released an impressive new model today (under their not-open-source Gemma license). Gemma 3 comes in four sizes – 1B, 4B, 12B, and 27B – and while 1B is text-only the larger three models are all multi-modal for…

  • Hacker News: Gemma3 – The current strongest model that fits on a single GPU

    Source URL: https://ollama.com/library/gemma3 Source: Hacker News Title: Gemma3 – The current strongest model that fits on a single GPU Feedly Summary: Comments AI Summary and Description: Yes **Summary:** The text discusses the features and capabilities of the Gemma 3 models developed by Google, which are built on Gemini technology and designed for multimodal tasks. Their…

  • Cloud Blog: How to deploy serverless AI with Gemma 3 on Cloud Run

    Source URL: https://cloud.google.com/blog/products/ai-machine-learning/serverless-ai-with-gemma-3-on-cloud-run/ Source: Cloud Blog Title: How to deploy serverless AI with Gemma 3 on Cloud Run Feedly Summary: Today, we introduced Gemma 3, a family of lightweight, open models built with the cutting-edge technology behind Gemini 2.0. The Gemma 3 family of models have been designed for speed and portability, empowering developers to…

  • Hacker News: A Practical Guide to Running Local LLMs

    Source URL: https://spin.atomicobject.com/running-local-llms/ Source: Hacker News Title: A Practical Guide to Running Local LLMs Feedly Summary: Comments AI Summary and Description: Yes Summary: The text discusses the intricacies of running local large language models (LLMs), emphasizing their applications in privacy-critical situations and the potential benefits of various tools like Ollama and Llama.cpp. It provides insights…

  • Hacker News: Local Deep Research – ArXiv, wiki and other searches included

    Source URL: https://github.com/LearningCircuit/local-deep-research Source: Hacker News Title: Local Deep Research – ArXiv, wiki and other searches included Feedly Summary: Comments AI Summary and Description: Yes Summary: This text outlines a sophisticated AI-powered research assistant designed for deep analysis through local and cloud-based LLM integrations, promoting privacy and comprehensive research capabilities. The focus on privacy, advanced…

  • Simon Willison’s Weblog: What’s new in the world of LLMs, for NICAR 2025

    Source URL: https://simonwillison.net/2025/Mar/8/nicar-llms/ Source: Simon Willison’s Weblog Title: What’s new in the world of LLMs, for NICAR 2025 Feedly Summary: I presented two sessions at the NICAR 2025 data journalism conference this year. The first was this one based on my review of LLMs in 2024, extended by several months to cover everything that’s happened…

  • Hacker News: Letta: Letta is a framework for creating LLM services with memory

    Source URL: https://github.com/letta-ai/letta Source: Hacker News Title: Letta: Letta is a framework for creating LLM services with memory Feedly Summary: Comments AI Summary and Description: Yes Summary: The text outlines the installation and usage of the Letta platform, a tool for managing and deploying large language model (LLM) agents. It highlights how to set up…

  • Simon Willison’s Weblog: QwQ-32B: Embracing the Power of Reinforcement Learning

    Source URL: https://simonwillison.net/2025/Mar/5/qwq-32b/#atom-everything Source: Simon Willison’s Weblog Title: QwQ-32B: Embracing the Power of Reinforcement Learning Feedly Summary: QwQ-32B: Embracing the Power of Reinforcement Learning New Apache 2 licensed reasoning model from Qwen: We are excited to introduce QwQ-32B, a model with 32 billion parameters that achieves performance comparable to DeepSeek-R1, which boasts 671 billion parameters…

  • Simon Willison’s Weblog: llm-mistral 0.11

    Source URL: https://simonwillison.net/2025/Mar/4/llm-mistral-011/#atom-everything Source: Simon Willison’s Weblog Title: llm-mistral 0.11 Feedly Summary: llm-mistral 0.11 I added schema support to this plugin which adds support for the Mistral API to LLM. Release notes: Support for LLM schemas. #19 -o prefix ‘{‘ option for forcing a response prefix. #18 Schemas now work with OpenAI, Anthropic, Gemini and…