Tag: pelican

  • Simon Willison’s Weblog: OpenAI’s new open weight (Apache 2) models are really good

    Source URL: https://simonwillison.net/2025/Aug/5/gpt-oss/ Source: Simon Willison’s Weblog Title: OpenAI’s new open weight (Apache 2) models are really good Feedly Summary: The long promised OpenAI open weight models are here, and they are very impressive. They’re available under proper open source licenses – Apache 2.0 – and come in two sizes, 120B and 20B. OpenAI’s own…

  • Simon Willison’s Weblog: Claude Opus 4.1

    Source URL: https://simonwillison.net/2025/Aug/5/claude-opus-41/ Source: Simon Willison’s Weblog Title: Claude Opus 4.1 Feedly Summary: Claude Opus 4.1 Surprise new model from Anthropic today – Claude Opus 4.1, which they describe as “a drop-in replacement for Opus 4". My favorite thing about this model is the version number – treating this as a .1 version increment looks…

  • Simon Willison’s Weblog: XBai o4

    Source URL: https://simonwillison.net/2025/Aug/3/xbai-o4/#atom-everything Source: Simon Willison’s Weblog Title: XBai o4 Feedly Summary: XBai o4 Yet another open source (Apache 2.0) LLM from a Chinese AI lab. This model card claims: XBai o4 excels in complex reasoning capabilities and has now completely surpassed OpenAI-o3-mini in Medium mode. This a 32.8 billion parameter model released by MetaStone…

  • Simon Willison’s Weblog: Trying out Qwen3 Coder Flash using LM Studio and Open WebUI and LLM

    Source URL: https://simonwillison.net/2025/Jul/31/qwen3-coder-flash/ Source: Simon Willison’s Weblog Title: Trying out Qwen3 Coder Flash using LM Studio and Open WebUI and LLM Feedly Summary: Qwen just released their sixth model(!) for this July called Qwen3-Coder-30B-A3B-Instruct – listed as Qwen3-Coder-Flash in their chat.qwen.ai interface. It’s 30.5B total parameters with 3.3B active at any one time. This means…

  • Simon Willison’s Weblog: Qwen3-30B-A3B-Thinking-2507

    Source URL: https://simonwillison.net/2025/Jul/30/qwen3-30b-a3b-thinking-2507/ Source: Simon Willison’s Weblog Title: Qwen3-30B-A3B-Thinking-2507 Feedly Summary: Qwen3-30B-A3B-Thinking-2507 Yesterday was Qwen3-30B-A3B-Instruct-2507. Qwen are clearly committed to their new split between reasoning and non-reasoning models (a reversal from Qwen 3 in April), because today they released the new reasoning partner to yesterday’s model: Qwen3-30B-A3B-Thinking-2507. I’m surprised at how poorly this reasoning mode…

  • Simon Willison’s Weblog: Qwen/Qwen3-30B-A3B-Instruct-2507

    Source URL: https://simonwillison.net/2025/Jul/29/qwen3-30b-a3b-instruct-2507/ Source: Simon Willison’s Weblog Title: Qwen/Qwen3-30B-A3B-Instruct-2507 Feedly Summary: Qwen/Qwen3-30B-A3B-Instruct-2507 New model update from Qwen, improving on their previous Qwen3-30B-A3B release from late April. In their tweet they said: Smarter, faster, and local deployment-friendly. ✨ Key Enhancements: ✅ Enhanced reasoning, coding, and math skills ✅ Broader multilingual knowledge ✅ Improved long-context understanding (up…

  • Simon Willison’s Weblog: My 2.5 year old laptop can write Space Invaders in JavaScript now

    Source URL: https://simonwillison.net/2025/Jul/29/space-invaders/ Source: Simon Willison’s Weblog Title: My 2.5 year old laptop can write Space Invaders in JavaScript now Feedly Summary: I wrote about the new GLM-4.5 model family yesterday – new open weight (MIT licensed) models from Z.ai in China which their benchmarks claim score highly in coding even against models such as…

  • Simon Willison’s Weblog: Qwen3-235B-A22B-Thinking-2507

    Source URL: https://simonwillison.net/2025/Jul/25/qwen3-235b-a22b-thinking-2507/#atom-everything Source: Simon Willison’s Weblog Title: Qwen3-235B-A22B-Thinking-2507 Feedly Summary: Qwen3-235B-A22B-Thinking-2507 The third Qwen model release week, following Qwen3-235B-A22B-Instruct-2507 on Monday 21st and Qwen3-Coder-480B-A35B-Instruct on Tuesday 22nd. Those two were both non-reasoning models – a change from the previous models in the Qwen 3 family which combined reasoning and non-reasoning in the same model,…