Tag: mlx

  • Simon Willison’s Weblog: deepseek-ai/DeepSeek-V3-0324

    Source URL: https://simonwillison.net/2025/Mar/24/deepseek/ Source: Simon Willison’s Weblog Title: deepseek-ai/DeepSeek-V3-0324 Feedly Summary: deepseek-ai/DeepSeek-V3-0324 Chinese AI lab DeepSeek just released the latest version of their enormous DeepSeek v3 model, baking the release date into the name DeepSeek-V3-0324. The license is MIT, the README is empty and the release adds up a to a total of 641 GB…

  • Simon Willison’s Weblog: Qwen2.5-VL-32B: Smarter and Lighter

    Source URL: https://simonwillison.net/2025/Mar/24/qwen25-vl-32b/#atom-everything Source: Simon Willison’s Weblog Title: Qwen2.5-VL-32B: Smarter and Lighter Feedly Summary: Qwen2.5-VL-32B: Smarter and Lighter The second big open weight LLM release from China today – the first being DeepSeek v3-0324. Qwen’s previous vision model was Qwen2.5 VL, released in January in 3B, 7B and 72B sizes. Today’s release is a 32B…

  • Simon Willison’s Weblog: deepseek-ai/DeepSeek-V3-0324

    Source URL: https://simonwillison.net/2025/Mar/24/deepseek/ Source: Simon Willison’s Weblog Title: deepseek-ai/DeepSeek-V3-0324 Feedly Summary: deepseek-ai/DeepSeek-V3-0324 Chinese AI lab DeepSeek just released the latest version of their enormous DeepSeek v3 model, baking the release date into the name DeepSeek-V3-0324. The license is MIT, the README is empty and the release adds up a to a total of 641 GB…

  • Hacker News: Mlx-community/OLMo-2-0325-32B-Instruct-4bit

    Source URL: https://simonwillison.net/2025/Mar/16/olmo2/ Source: Hacker News Title: Mlx-community/OLMo-2-0325-32B-Instruct-4bit Feedly Summary: Comments AI Summary and Description: Yes Summary: The text discusses the OLMo 2 model, which claims to be a superior, fully open alternative to GPT-3.5 Turbo and GPT-4o mini. It provides installation instructions for running this model on a Mac, highlighting its ease of access…

  • Simon Willison’s Weblog: Mistral Small 3.1

    Source URL: https://simonwillison.net/2025/Mar/17/mistral-small-31/#atom-everything Source: Simon Willison’s Weblog Title: Mistral Small 3.1 Feedly Summary: Mistral Small 3.1 Mistral Small 3 came out in January and was a notable, genuinely excellent local model that used an Apache 2.0 license. Mistral Small 3.1 offers a significant improvement: it’s multi-modal (images) and has an increased 128,000 token context length,…

  • Simon Willison’s Weblog: mlx-community/OLMo-2-0325-32B-Instruct-4bit

    Source URL: https://simonwillison.net/2025/Mar/16/olmo2/#atom-everything Source: Simon Willison’s Weblog Title: mlx-community/OLMo-2-0325-32B-Instruct-4bit Feedly Summary: mlx-community/OLMo-2-0325-32B-Instruct-4bit OLMo 2 32B claims to be “the first fully-open model (all data, code, weights, and details are freely available) to outperform GPT3.5-Turbo and GPT-4o mini". Thanks to the MLX project here’s a recipe that worked for me to run it on my Mac,…

  • Simon Willison’s Weblog: Notes on Google’s Gemma 3

    Source URL: https://simonwillison.net/2025/Mar/12/gemma-3/ Source: Simon Willison’s Weblog Title: Notes on Google’s Gemma 3 Feedly Summary: Google’s Gemma team released an impressive new model today (under their not-open-source Gemma license). Gemma 3 comes in four sizes – 1B, 4B, 12B, and 27B – and while 1B is text-only the larger three models are all multi-modal for…

  • Simon Willison’s Weblog: Notes on Google’s Gemma 3

    Source URL: https://simonwillison.net/2025/Mar/12/notes-on-googles-gemma-3/ Source: Simon Willison’s Weblog Title: Notes on Google’s Gemma 3 Feedly Summary: Google’s Gemma team released an impressive new model today (under their not-open-source Gemma license). Gemma 3 comes in four sizes – 1B, 4B, 12B, and 27B – and while 1B is text-only the larger three models are all multi-modal for…

  • Simon Willison’s Weblog: QwQ-32B: Embracing the Power of Reinforcement Learning

    Source URL: https://simonwillison.net/2025/Mar/5/qwq-32b/#atom-everything Source: Simon Willison’s Weblog Title: QwQ-32B: Embracing the Power of Reinforcement Learning Feedly Summary: QwQ-32B: Embracing the Power of Reinforcement Learning New Apache 2 licensed reasoning model from Qwen: We are excited to introduce QwQ-32B, a model with 32 billion parameters that achieves performance comparable to DeepSeek-R1, which boasts 671 billion parameters…

  • Simon Willison’s Weblog: Run LLMs on macOS using llm-mlx and Apple’s MLX framework

    Source URL: https://simonwillison.net/2025/Feb/15/llm-mlx/#atom-everything Source: Simon Willison’s Weblog Title: Run LLMs on macOS using llm-mlx and Apple’s MLX framework Feedly Summary: llm-mlx is a brand new plugin for my LLM Python Library and CLI utility which builds on top of Apple’s excellent MLX array framework library and mlx-lm package. If you’re a terminal user or Python…