Tag: mlx

  • Simon Willison’s Weblog: Run Your Own AI

    Source URL: https://simonwillison.net/2025/Jun/3/run-your-own-ai/ Source: Simon Willison’s Weblog Title: Run Your Own AI Feedly Summary: Run Your Own AI Anthony Lewis published this neat, concise tutorial on using my LLM tool to run local models on your own machine, using llm-mlx. An under-appreciated way to contribute to open source projects is to publish unofficial guides like…

  • Simon Willison’s Weblog: Saying "hi" to Microsoft’s Phi-4-reasoning

    Source URL: https://simonwillison.net/2025/May/6/phi-4-reasoning/#atom-everything Source: Simon Willison’s Weblog Title: Saying "hi" to Microsoft’s Phi-4-reasoning Feedly Summary: Microsoft released a new sub-family of models a few days ago: Phi-4 reasoning. They introduced them in this blog post celebrating a year since the release of Phi-3: Today, we are excited to introduce Phi-4-reasoning, Phi-4-reasoning-plus, and Phi-4-mini-reasoning – marking…

  • Simon Willison’s Weblog: Qwen3-8B

    Source URL: https://simonwillison.net/2025/May/2/qwen3-8b/#atom-everything Source: Simon Willison’s Weblog Title: Qwen3-8B Feedly Summary: Having tried a few of the Qwen 3 models now my favorite is a bit of a surprise to me: I’m really enjoying Qwen3-8B. I’ve been running prompts through the MLX 4bit quantized version, mlx-community/Qwen3-8B-4bit. I’m using llm-mlx like this: llm install llm-mlx llm…

  • Simon Willison’s Weblog: llm-fragments-github 0.2

    Source URL: https://simonwillison.net/2025/Apr/20/llm-fragments-github/#atom-everything Source: Simon Willison’s Weblog Title: llm-fragments-github 0.2 Feedly Summary: llm-fragments-github 0.2 I upgraded my llm-fragments-github plugin to add a new fragment type called issue. It lets you pull the entire content of a GitHub issue thread into your prompt as a concatenated Markdown file. (If you haven’t seen fragments before I introduced…

  • Simon Willison’s Weblog: Gemma 3 QAT Models

    Source URL: https://simonwillison.net/2025/Apr/19/gemma-3-qat-models/ Source: Simon Willison’s Weblog Title: Gemma 3 QAT Models Feedly Summary: Gemma 3 QAT Models Interesting release from Google, as a follow-up to Gemma 3 from last month: To make Gemma 3 even more accessible, we are announcing new versions optimized with Quantization-Aware Training (QAT) that dramatically reduces memory requirements while maintaining…

  • Simon Willison’s Weblog: deepseek-ai/DeepSeek-V3-0324

    Source URL: https://simonwillison.net/2025/Mar/24/deepseek/ Source: Simon Willison’s Weblog Title: deepseek-ai/DeepSeek-V3-0324 Feedly Summary: deepseek-ai/DeepSeek-V3-0324 Chinese AI lab DeepSeek just released the latest version of their enormous DeepSeek v3 model, baking the release date into the name DeepSeek-V3-0324. The license is MIT, the README is empty and the release adds up a to a total of 641 GB…

  • Simon Willison’s Weblog: Qwen2.5-VL-32B: Smarter and Lighter

    Source URL: https://simonwillison.net/2025/Mar/24/qwen25-vl-32b/#atom-everything Source: Simon Willison’s Weblog Title: Qwen2.5-VL-32B: Smarter and Lighter Feedly Summary: Qwen2.5-VL-32B: Smarter and Lighter The second big open weight LLM release from China today – the first being DeepSeek v3-0324. Qwen’s previous vision model was Qwen2.5 VL, released in January in 3B, 7B and 72B sizes. Today’s release is a 32B…

  • Simon Willison’s Weblog: deepseek-ai/DeepSeek-V3-0324

    Source URL: https://simonwillison.net/2025/Mar/24/deepseek/ Source: Simon Willison’s Weblog Title: deepseek-ai/DeepSeek-V3-0324 Feedly Summary: deepseek-ai/DeepSeek-V3-0324 Chinese AI lab DeepSeek just released the latest version of their enormous DeepSeek v3 model, baking the release date into the name DeepSeek-V3-0324. The license is MIT, the README is empty and the release adds up a to a total of 641 GB…