Tag: context handling

  • Docker: IBM Granite 4.0 Models Now Available on Docker Hub

    Source URL: https://www.docker.com/blog/ibm-granite-4-0-models-now-available-on-docker-hub/ Source: Docker Title: IBM Granite 4.0 Models Now Available on Docker Hub Feedly Summary: Developers can now discover and run IBM’s latest open-source Granite 4.0 language models from the Docker Hub model catalog, and start building in minutes with Docker Model Runner. Granite 4.0 pairs strong, enterprise-ready performance with a lightweight footprint,…

  • AWS News Blog: Qwen models are now available in Amazon Bedrock

    Source URL: https://aws.amazon.com/blogs/aws/qwen-models-are-now-available-in-amazon-bedrock/ Source: AWS News Blog Title: Qwen models are now available in Amazon Bedrock Feedly Summary: Amazon Bedrock has expanded its model offerings with the addition of Qwen 3 foundation models enabling users to access and deploy them in a fully managed, serverless environment. These models feature both mixture-of-experts (MoE) and dense architectures…

  • Simon Willison’s Weblog: Claude Sonnet 4 now supports 1M tokens of context

    Source URL: https://simonwillison.net/2025/Aug/12/claude-sonnet-4-1m/ Source: Simon Willison’s Weblog Title: Claude Sonnet 4 now supports 1M tokens of context Feedly Summary: Claude Sonnet 4 now supports 1M tokens of context Gemini and OpenAI both have million token models, so it’s good to see Anthropic catching up. This is 5x the previous 200,000 context length limit of the…

  • Simon Willison’s Weblog: Anthropic: How we built our multi-agent research system

    Source URL: https://simonwillison.net/2025/Jun/14/multi-agent-research-system/#atom-everything Source: Simon Willison’s Weblog Title: Anthropic: How we built our multi-agent research system Feedly Summary: Anthropic: How we built our multi-agent research system OK, I’m sold on multi-agent LLM systems now. I’ve been pretty skeptical of these until recently: why make your life more complicated by running multiple different prompts in parallel…

  • Simon Willison’s Weblog: o3-pro

    Source URL: https://simonwillison.net/2025/Jun/10/o3-pro/ Source: Simon Willison’s Weblog Title: o3-pro Feedly Summary: o3-pro OpenAI released o3-pro today, which they describe as a “version of o3 with more compute for better responses". It’s only available via the newer Responses API. I’ve added it to my llm-openai-plugin plugin which uses that new API, so you can try it…

  • Simon Willison’s Weblog: GPT-4.1: Three new million token input models from OpenAI, including their cheapest model yet

    Source URL: https://simonwillison.net/2025/Apr/14/gpt-4-1/ Source: Simon Willison’s Weblog Title: GPT-4.1: Three new million token input models from OpenAI, including their cheapest model yet Feedly Summary: OpenAI introduced three new models this morning: GPT-4.1, GPT-4.1 mini and GPT-4.1 nano. These are API-only models right now, not available through the ChatGPT interface (though you can try them out…

  • Slashdot: In ‘Milestone’ for Open Source, Meta Releases New Benchmark-Beating Llama 4 Models

    Source URL: https://news.slashdot.org/story/25/04/06/182233/in-milestone-for-open-source-meta-releases-new-benchmark-beating-llama-4-models?utm_source=rss1.0mainlinkanon&utm_medium=feed Source: Slashdot Title: In ‘Milestone’ for Open Source, Meta Releases New Benchmark-Beating Llama 4 Models Feedly Summary: AI Summary and Description: Yes Summary: Mark Zuckerberg recently announced the launch of four new Llama Large Language Models (LLMs) that reinforce Meta’s commitment to open source AI. These models, particularly Llama 4 Scout and…