Tag: context length

  • Simon Willison’s Weblog: Claude Sonnet 4 now supports 1M tokens of context

    Source URL: https://simonwillison.net/2025/Aug/12/claude-sonnet-4-1m/ Source: Simon Willison’s Weblog Title: Claude Sonnet 4 now supports 1M tokens of context Feedly Summary: Claude Sonnet 4 now supports 1M tokens of context Gemini and OpenAI both have million token models, so it’s good to see Anthropic catching up. This is 5x the previous 200,000 context length limit of the…

  • Simon Willison’s Weblog: Qwen3-4B Instruct and Thinking

    Source URL: https://simonwillison.net/2025/Aug/6/qwen3-4b-instruct-and-thinking/ Source: Simon Willison’s Weblog Title: Qwen3-4B Instruct and Thinking Feedly Summary: Qwen3-4B Instruct and Thinking Yet another interesting model from Qwen—these are tiny compared to their other recent releases (just 4B parameters, 7.5GB on Hugging Face and even smaller when quantized) but with a 262,144 context length, which Qwen suggest is essential…

  • Simon Willison’s Weblog: Qwen3-235B-A22B-Thinking-2507

    Source URL: https://simonwillison.net/2025/Jul/25/qwen3-235b-a22b-thinking-2507/#atom-everything Source: Simon Willison’s Weblog Title: Qwen3-235B-A22B-Thinking-2507 Feedly Summary: Qwen3-235B-A22B-Thinking-2507 The third Qwen model release week, following Qwen3-235B-A22B-Instruct-2507 on Monday 21st and Qwen3-Coder-480B-A35B-Instruct on Tuesday 22nd. Those two were both non-reasoning models – a change from the previous models in the Qwen 3 family which combined reasoning and non-reasoning in the same model,…

  • Simon Willison’s Weblog: TimeScope: How Long Can Your Video Large Multimodal Model Go?

    Source URL: https://simonwillison.net/2025/Jul/23/timescope/#atom-everything Source: Simon Willison’s Weblog Title: TimeScope: How Long Can Your Video Large Multimodal Model Go? Feedly Summary: TimeScope: How Long Can Your Video Large Multimodal Model Go? New open source benchmark for evaluating vision LLMs on how well they handle long videos: TimeScope probes the limits of long-video capabilities by inserting several…

  • Simon Willison’s Weblog: Qwen3-Coder: Agentic Coding in the World

    Source URL: https://simonwillison.net/2025/Jul/22/qwen3-coder/ Source: Simon Willison’s Weblog Title: Qwen3-Coder: Agentic Coding in the World Feedly Summary: Qwen3-Coder: Agentic Coding in the World It turns out that as I was typing up my notes on Qwen3-235B-A22B-Instruct-2507 the Qwen team were unleashing something much bigger: Today, we’re announcing Qwen3-Coder, our most agentic code model to date. Qwen3-Coder…

  • Cloud Blog: How to use Gemini 2.5 to fine-tune video outputs on Vertex AI

    Source URL: https://cloud.google.com/blog/products/ai-machine-learning/how-to-fine-tune-video-outputs-using-vertex-ai/ Source: Cloud Blog Title: How to use Gemini 2.5 to fine-tune video outputs on Vertex AI Feedly Summary: Recently, we announced Gemini 2.5 is generally available on Vertex AI. As part of this update, tuning capabilities have extended beyond text outputs – now, you can tune image, audio, and video outputs on…