Tag: mlx

  • Hacker News: Mlx-community/OLMo-2-0325-32B-Instruct-4bit

    Source URL: https://simonwillison.net/2025/Mar/16/olmo2/ Source: Hacker News Title: Mlx-community/OLMo-2-0325-32B-Instruct-4bit Feedly Summary: Comments AI Summary and Description: Yes Summary: The text discusses the OLMo 2 model, which claims to be a superior, fully open alternative to GPT-3.5 Turbo and GPT-4o mini. It provides installation instructions for running this model on a Mac, highlighting its ease of access…

  • Simon Willison’s Weblog: Mistral Small 3.1

    Source URL: https://simonwillison.net/2025/Mar/17/mistral-small-31/#atom-everything Source: Simon Willison’s Weblog Title: Mistral Small 3.1 Feedly Summary: Mistral Small 3.1 Mistral Small 3 came out in January and was a notable, genuinely excellent local model that used an Apache 2.0 license. Mistral Small 3.1 offers a significant improvement: it’s multi-modal (images) and has an increased 128,000 token context length,…

  • Simon Willison’s Weblog: mlx-community/OLMo-2-0325-32B-Instruct-4bit

    Source URL: https://simonwillison.net/2025/Mar/16/olmo2/#atom-everything Source: Simon Willison’s Weblog Title: mlx-community/OLMo-2-0325-32B-Instruct-4bit Feedly Summary: mlx-community/OLMo-2-0325-32B-Instruct-4bit OLMo 2 32B claims to be “the first fully-open model (all data, code, weights, and details are freely available) to outperform GPT3.5-Turbo and GPT-4o mini". Thanks to the MLX project here’s a recipe that worked for me to run it on my Mac,…

  • Simon Willison’s Weblog: Notes on Google’s Gemma 3

    Source URL: https://simonwillison.net/2025/Mar/12/gemma-3/ Source: Simon Willison’s Weblog Title: Notes on Google’s Gemma 3 Feedly Summary: Google’s Gemma team released an impressive new model today (under their not-open-source Gemma license). Gemma 3 comes in four sizes – 1B, 4B, 12B, and 27B – and while 1B is text-only the larger three models are all multi-modal for…

  • Simon Willison’s Weblog: Notes on Google’s Gemma 3

    Source URL: https://simonwillison.net/2025/Mar/12/notes-on-googles-gemma-3/ Source: Simon Willison’s Weblog Title: Notes on Google’s Gemma 3 Feedly Summary: Google’s Gemma team released an impressive new model today (under their not-open-source Gemma license). Gemma 3 comes in four sizes – 1B, 4B, 12B, and 27B – and while 1B is text-only the larger three models are all multi-modal for…

  • Simon Willison’s Weblog: QwQ-32B: Embracing the Power of Reinforcement Learning

    Source URL: https://simonwillison.net/2025/Mar/5/qwq-32b/#atom-everything Source: Simon Willison’s Weblog Title: QwQ-32B: Embracing the Power of Reinforcement Learning Feedly Summary: QwQ-32B: Embracing the Power of Reinforcement Learning New Apache 2 licensed reasoning model from Qwen: We are excited to introduce QwQ-32B, a model with 32 billion parameters that achieves performance comparable to DeepSeek-R1, which boasts 671 billion parameters…

  • Simon Willison’s Weblog: Run LLMs on macOS using llm-mlx and Apple’s MLX framework

    Source URL: https://simonwillison.net/2025/Feb/15/llm-mlx/#atom-everything Source: Simon Willison’s Weblog Title: Run LLMs on macOS using llm-mlx and Apple’s MLX framework Feedly Summary: llm-mlx is a brand new plugin for my LLM Python Library and CLI utility which builds on top of Apple’s excellent MLX array framework library and mlx-lm package. If you’re a terminal user or Python…

  • Hacker News: Show HN: VimLM – A Local, Offline Coding Assistant for Vim

    Source URL: https://github.com/JosefAlbers/VimLM Source: Hacker News Title: Show HN: VimLM – A Local, Offline Coding Assistant for Vim Feedly Summary: Comments AI Summary and Description: Yes Summary: VimLM is an innovative coding assistant that integrates a local Large Language Model (LLM) into the Vim environment, allowing users to enhance their coding workflow with AI-driven code…