Tag: mlx

  • Simon Willison’s Weblog: Introducing EmbeddingGemma

    Source URL: https://simonwillison.net/2025/Sep/4/embedding-gemma/#atom-everything Source: Simon Willison’s Weblog Title: Introducing EmbeddingGemma Feedly Summary: Introducing EmbeddingGemma Brand new open weights (under the slightly janky Gemma license) 308M parameter embedding model from Google: Based on the Gemma 3 architecture, EmbeddingGemma is trained on 100+ languages and is small enough to run on less than 200MB of RAM with…

  • Simon Willison’s Weblog: Open weight LLMs exhibit inconsistent performance across providers

    Source URL: https://simonwillison.net/2025/Aug/15/inconsistent-performance/ Source: Simon Willison’s Weblog Title: Open weight LLMs exhibit inconsistent performance across providers Feedly Summary: Artificial Analysis published a new benchmark the other day, this time focusing on how an individual model – OpenAI’s gpt-oss-120b – performs across different hosted providers. The results showed some surprising differences. Here’s the one with the…

  • Simon Willison’s Weblog: Qwen3-4B Instruct and Thinking

    Source URL: https://simonwillison.net/2025/Aug/6/qwen3-4b-instruct-and-thinking/ Source: Simon Willison’s Weblog Title: Qwen3-4B Instruct and Thinking Feedly Summary: Qwen3-4B Instruct and Thinking Yet another interesting model from Qwen—these are tiny compared to their other recent releases (just 4B parameters, 7.5GB on Hugging Face and even smaller when quantized) but with a 262,144 context length, which Qwen suggest is essential…

  • Simon Willison’s Weblog: XBai o4

    Source URL: https://simonwillison.net/2025/Aug/3/xbai-o4/#atom-everything Source: Simon Willison’s Weblog Title: XBai o4 Feedly Summary: XBai o4 Yet another open source (Apache 2.0) LLM from a Chinese AI lab. This model card claims: XBai o4 excels in complex reasoning capabilities and has now completely surpassed OpenAI-o3-mini in Medium mode. This a 32.8 billion parameter model released by MetaStone…

  • Simon Willison’s Weblog: Trying out Qwen3 Coder Flash using LM Studio and Open WebUI and LLM

    Source URL: https://simonwillison.net/2025/Jul/31/qwen3-coder-flash/ Source: Simon Willison’s Weblog Title: Trying out Qwen3 Coder Flash using LM Studio and Open WebUI and LLM Feedly Summary: Qwen just released their sixth model(!) for this July called Qwen3-Coder-30B-A3B-Instruct – listed as Qwen3-Coder-Flash in their chat.qwen.ai interface. It’s 30.5B total parameters with 3.3B active at any one time. This means…

  • Simon Willison’s Weblog: Qwen/Qwen3-30B-A3B-Instruct-2507

    Source URL: https://simonwillison.net/2025/Jul/29/qwen3-30b-a3b-instruct-2507/ Source: Simon Willison’s Weblog Title: Qwen/Qwen3-30B-A3B-Instruct-2507 Feedly Summary: Qwen/Qwen3-30B-A3B-Instruct-2507 New model update from Qwen, improving on their previous Qwen3-30B-A3B release from late April. In their tweet they said: Smarter, faster, and local deployment-friendly. ✨ Key Enhancements: ✅ Enhanced reasoning, coding, and math skills ✅ Broader multilingual knowledge ✅ Improved long-context understanding (up…

  • Simon Willison’s Weblog: My 2.5 year old laptop can write Space Invaders in JavaScript now

    Source URL: https://simonwillison.net/2025/Jul/29/space-invaders/ Source: Simon Willison’s Weblog Title: My 2.5 year old laptop can write Space Invaders in JavaScript now Feedly Summary: I wrote about the new GLM-4.5 model family yesterday – new open weight (MIT licensed) models from Z.ai in China which their benchmarks claim score highly in coding even against models such as…

  • Simon Willison’s Weblog: Introducing Gemma 3n: The developer guide

    Source URL: https://simonwillison.net/2025/Jun/26/gemma-3n/ Source: Simon Willison’s Weblog Title: Introducing Gemma 3n: The developer guide Feedly Summary: Introducing Gemma 3n: The developer guide Extremely consequential new open weights model release from Google today: Multimodal by design: Gemma 3n natively supports image, audio, video, and text inputs and text outputs. Optimized for on-device: Engineered with a focus…

  • Simon Willison’s Weblog: Comma v0.1 1T and 2T – 7B LLMs trained on openly licensed text

    Source URL: https://simonwillison.net/2025/Jun/7/comma/#atom-everything Source: Simon Willison’s Weblog Title: Comma v0.1 1T and 2T – 7B LLMs trained on openly licensed text Feedly Summary: It’s been a long time coming, but we finally have some promising LLMs to try out which are trained entirely on openly licensed text! EleutherAI released the Pile four and a half…