Tag: Qwen
-
Simon Willison’s Weblog: AbsenceBench: Language Models Can’t Tell What’s Missing
Source URL: https://simonwillison.net/2025/Jun/20/absencebench/#atom-everything Source: Simon Willison’s Weblog Title: AbsenceBench: Language Models Can’t Tell What’s Missing Feedly Summary: AbsenceBench: Language Models Can’t Tell What’s Missing Here’s another interesting result to file under the “jagged frontier" of LLMs, where their strengths and weaknesses are often unintuitive. Long context models have been getting increasingly good at passing "Needle…
-
Simon Willison’s Weblog: Qwen3 Embedding
Source URL: https://simonwillison.net/2025/Jun/8/qwen3-embedding/#atom-everything Source: Simon Willison’s Weblog Title: Qwen3 Embedding Feedly Summary: Qwen3 Embedding New family of embedding models from Qwen, in three sizes: 0.6B, 4B, 8B – and two categories: Text Embedding and Text Reranking. The full collection can be browsed on Hugging Face. The smallest available model is the 0.6B Q8 one, which…
-
Cloud Blog: Building a Production Multimodal Fine-Tuning Pipeline
Source URL: https://cloud.google.com/blog/topics/developers-practitioners/building-a-production-multimodal-fine-tuning-pipeline/ Source: Cloud Blog Title: Building a Production Multimodal Fine-Tuning Pipeline Feedly Summary: Looking to fine-tune multimodal AI models for your specific domain but facing infrastructure and implementation challenges? This guide demonstrates how to overcome the multimodal implementation gap using Google Cloud and Axolotl, with a complete hands-on example fine-tuning Gemma 3 on…
-
Simon Willison’s Weblog: Shisa V2 405B: Japan’s Highest Performing LLM
Source URL: https://simonwillison.net/2025/Jun/3/shisa-v2/ Source: Simon Willison’s Weblog Title: Shisa V2 405B: Japan’s Highest Performing LLM Feedly Summary: Shisa V2 405B: Japan’s Highest Performing LLM Leonard Lin and Adam Lensenmayer have been working on Shisa for a while. They describe their latest release as “Japan’s Highest Performing LLM". Shisa V2 405B is the highest-performing LLM ever…
-
Simon Willison’s Weblog: deepseek-ai/DeepSeek-R1-0528
Source URL: https://simonwillison.net/2025/May/31/deepseek-aideepseek-r1-0528/ Source: Simon Willison’s Weblog Title: deepseek-ai/DeepSeek-R1-0528 Feedly Summary: deepseek-ai/DeepSeek-R1-0528 Sadly the trend for terrible naming of models has infested the Chinese AI labs as well. DeepSeek-R1-0528 is a brand new and much improved open weights reasoning model from DeepSeek, a major step up from the DeepSeek R1 they released back in January.…
-
Simon Willison’s Weblog: Large Language Models can run tools in your terminal with LLM 0.26
Source URL: https://simonwillison.net/2025/May/27/llm-tools/ Source: Simon Willison’s Weblog Title: Large Language Models can run tools in your terminal with LLM 0.26 Feedly Summary: LLM 0.26 is out with the biggest new feature since I started the project: support for tools. You can now use the LLM CLI tool – and Python library – to grant LLMs…
-
Simon Willison’s Weblog: Devstral
Source URL: https://simonwillison.net/2025/May/21/devstral/#atom-everything Source: Simon Willison’s Weblog Title: Devstral Feedly Summary: Devstral New Apache 2.0 licensed LLM release from Mistral, this time specifically trained for code. Devstral achieves a score of 46.8% on SWE-Bench Verified, outperforming prior open-source SoTA models by more than 6% points. When evaluated under the same test scaffold (OpenHands, provided by…
-
Simon Willison’s Weblog: qwen2.5vl in Ollama
Source URL: https://simonwillison.net/2025/May/18/qwen25vl-in-ollama/#atom-everything Source: Simon Willison’s Weblog Title: qwen2.5vl in Ollama Feedly Summary: qwen2.5vl in Ollama Ollama announced a complete overhaul of their vision support the other day. Here’s the first new model they’ve shipped since then – a packaged version of Qwen 2.5 VL which was first released on January 26th 2025. Here are…
-
Simon Willison’s Weblog: Building software on top of Large Language Models
Source URL: https://simonwillison.net/2025/May/15/building-on-llms/#atom-everything Source: Simon Willison’s Weblog Title: Building software on top of Large Language Models Feedly Summary: I presented a three hour workshop at PyCon US yesterday titled Building software on top of Large Language Models. The goal of the workshop was to give participants everything they needed to get started writing code that…
-
Simon Willison’s Weblog: Vision Language Models (Better, Faster, Stronger)
Source URL: https://simonwillison.net/2025/May/13/vision-language-models/#atom-everything Source: Simon Willison’s Weblog Title: Vision Language Models (Better, Faster, Stronger) Feedly Summary: Vision Language Models (Better, Faster, Stronger) Extremely useful review of the last year in vision and multi-modal LLMs. So much has happened! I’m particularly excited about the range of small open weight vision models that are now available. Models…