Tag: mlx

  • Simon Willison’s Weblog: Qwen2.5 VL! Qwen2.5 VL! Qwen2.5 VL!

    Source URL: https://simonwillison.net/2025/Jan/27/qwen25-vl-qwen25-vl-qwen25-vl/ Source: Simon Willison’s Weblog Title: Qwen2.5 VL! Qwen2.5 VL! Qwen2.5 VL! Feedly Summary: Qwen2.5 VL! Qwen2.5 VL! Qwen2.5 VL! Hot on the heels of yesterday’s Qwen2.5-1M, here’s Qwen2.5 VL (with an excitable announcement title) – the latest in Qwen’s series of vision LLMs. They’re releasing multiple versions: base models and instruction tuned…

  • Simon Willison’s Weblog: The impact of competition and DeepSeek on Nvidia

    Source URL: https://simonwillison.net/2025/Jan/27/deepseek-nvidia/ Source: Simon Willison’s Weblog Title: The impact of competition and DeepSeek on Nvidia Feedly Summary: The impact of competition and DeepSeek on Nvidia Long, excellent piece by Jeffrey Emanuel capturing the current state of the AI/LLM industry. The original title is “The Short Case for Nvidia Stock" – I’m using the Hacker…

  • Simon Willison’s Weblog: r1.py script to run R1 with a min-thinking-tokens parameter

    Source URL: https://simonwillison.net/2025/Jan/22/r1py/ Source: Simon Willison’s Weblog Title: r1.py script to run R1 with a min-thinking-tokens parameter Feedly Summary: r1.py script to run R1 with a min-thinking-tokens parameter Fantastically creative hack by Theia Vogel. The DeepSeek R1 family of models output their chain of thought inside a …</think> block. Theia found that you can intercept…

  • Simon Willison’s Weblog: Run DeepSeek R1 or V3 with MLX Distributed

    Source URL: https://simonwillison.net/2025/Jan/22/mlx-distributed/ Source: Simon Willison’s Weblog Title: Run DeepSeek R1 or V3 with MLX Distributed Feedly Summary: Run DeepSeek R1 or V3 with MLX Distributed Handy detailed instructions from Awni Hannun on running the enormous DeepSeek R1 or v3 models on a cluster of Macs using the distributed communication feature of Apple’s MLX library.…

  • Simon Willison’s Weblog: microsoft/phi-4

    Source URL: https://simonwillison.net/2025/Jan/8/phi-4/ Source: Simon Willison’s Weblog Title: microsoft/phi-4 Feedly Summary: microsoft/phi-4 Here’s the official release of Microsoft’s Phi-4 LLM, now officially under an MIT license. A few weeks ago I covered the earlier unofficial versions, where I talked about how the model used synthetic training data in some really interesting ways. It benchmarks favorably…

  • Simon Willison’s Weblog: Weeknotes: Starting 2025 a little slow

    Source URL: https://simonwillison.net/2025/Jan/4/weeknotes/#atom-everything Source: Simon Willison’s Weblog Title: Weeknotes: Starting 2025 a little slow Feedly Summary: I published my review of 2024 in LLMs and then got into a fight with most of the internet over the phone microphone targeted ads conspiracy theory. In my last weeknotes I talked about how December in LLMs has…

  • Simon Willison’s Weblog: Things we learned out about LLMs in 2024

    Source URL: https://simonwillison.net/2024/Dec/31/llms-in-2024/#atom-everything Source: Simon Willison’s Weblog Title: Things we learned out about LLMs in 2024 Feedly Summary: A lot has happened in the world of Large Language Models over the course of 2024. Here’s a review of things we figured out about the field in the past twelve months, plus my attempt at identifying…

  • Simon Willison’s Weblog: Trying out QvQ – Qwen’s new visual reasoning model

    Source URL: https://simonwillison.net/2024/Dec/24/qvq/#atom-everything Source: Simon Willison’s Weblog Title: Trying out QvQ – Qwen’s new visual reasoning model Feedly Summary: I thought we were done for major model releases in 2024, but apparently not: Alibaba’s Qwen team just dropped the Apache2 2 licensed QvQ-72B-Preview, “an experimental research model focusing on enhancing visual reasoning capabilities". Their blog…