Tag: 0 Flash
-
Simon Willison’s Weblog: gpt-5 and gpt-5-mini rate limit updates
Source URL: https://simonwillison.net/2025/Sep/12/gpt-5-rate-limits/#atom-everything Source: Simon Willison’s Weblog Title: gpt-5 and gpt-5-mini rate limit updates Feedly Summary: gpt-5 and gpt-5-mini rate limit updates OpenAI have increased the rate limits for their two main GPT-5 models. These look significant: gpt-5 Tier 1: 30K → 500K TPM (1.5M batch) Tier 2: 450K → 1M (3M batch) Tier 3:…
-
Cloud Blog: Gemini momentum continues with launch of 2.5 Flash-Lite and general availability of 2.5 Flash and Pro on Vertex AI
Source URL: https://cloud.google.com/blog/products/ai-machine-learning/gemini-2-5-flash-lite-flash-pro-ga-vertex-ai/ Source: Cloud Blog Title: Gemini momentum continues with launch of 2.5 Flash-Lite and general availability of 2.5 Flash and Pro on Vertex AI Feedly Summary: The momentum of the Gemini 2.5 era continues to build. Following our recent announcements, we’re empowering enterprise builders and developers with even greater access to the intelligence,…
-
Cloud Blog: Multimodal agents tutorial: How to use Gemini, Langchain, and LangGraph to build agents for object detection
Source URL: https://cloud.google.com/blog/products/ai-machine-learning/build-multimodal-agents-using-gemini-langchain-and-langgraph/ Source: Cloud Blog Title: Multimodal agents tutorial: How to use Gemini, Langchain, and LangGraph to build agents for object detection Feedly Summary: Here’s a common scenario when building AI agents that might feel confusing: How can you use the latest Gemini models and an open-source framework like LangChain and LangGraph to create…
-
Simon Willison’s Weblog: How often do LLMs snitch? Recreating Theo’s SnitchBench with LLM
Source URL: https://simonwillison.net/2025/May/31/snitchbench-with-llm/#atom-everything Source: Simon Willison’s Weblog Title: How often do LLMs snitch? Recreating Theo’s SnitchBench with LLM Feedly Summary: A fun new benchmark just dropped! Inspired by the Claude 4 system card – which showed that Claude 4 might just rat you out to the authorities if you told it to “take initiative" in…
-
Simon Willison’s Weblog: Create and edit images with Gemini 2.0 in preview
Source URL: https://simonwillison.net/2025/May/7/gemini-images-preview/#atom-everything Source: Simon Willison’s Weblog Title: Create and edit images with Gemini 2.0 in preview Feedly Summary: Create and edit images with Gemini 2.0 in preview Gemini 2.0 Flash has had image generation capabilities for a while now, and they’re now available via the paid Gemini API – at 3.9 cents per generated…
-
Cloud Blog: Build live voice-driven agentic applications with Vertex AI Gemini Live API
Source URL: https://cloud.google.com/blog/products/ai-machine-learning/build-voice-driven-applications-with-live-api/ Source: Cloud Blog Title: Build live voice-driven agentic applications with Vertex AI Gemini Live API Feedly Summary: Across industries, enterprises need efficient and proactive solutions. Imagine frontline professionals using voice commands and visual input to diagnose issues, access vital information, and initiate processes in real-time. The Gemini 2.0 Flash Live API empowers…
-
Simon Willison’s Weblog: Start building with Gemini 2.5 Flash
Source URL: https://simonwillison.net/2025/Apr/17/start-building-with-gemini-25-flash/ Source: Simon Willison’s Weblog Title: Start building with Gemini 2.5 Flash Feedly Summary: Start building with Gemini 2.5 Flash Google Gemini’s latest model is Gemini 2.5 Flash, available in (paid) preview as gemini-2.5-flash-preview-04-17. Building upon the popular foundation of 2.0 Flash, this new version delivers a major upgrade in reasoning capabilities, while…
-
Simon Willison’s Weblog: GPT-4.1: Three new million token input models from OpenAI, including their cheapest model yet
Source URL: https://simonwillison.net/2025/Apr/14/gpt-4-1/ Source: Simon Willison’s Weblog Title: GPT-4.1: Three new million token input models from OpenAI, including their cheapest model yet Feedly Summary: OpenAI introduced three new models this morning: GPT-4.1, GPT-4.1 mini and GPT-4.1 nano. These are API-only models right now, not available through the ChatGPT interface (though you can try them out…
-
Slashdot: Meta Got Caught Gaming AI Benchmarks
Source URL: https://tech.slashdot.org/story/25/04/08/133257/meta-got-caught-gaming-ai-benchmarks?utm_source=rss1.0mainlinkanon&utm_medium=feed Source: Slashdot Title: Meta Got Caught Gaming AI Benchmarks Feedly Summary: AI Summary and Description: Yes Summary: Meta’s release of the Llama 4 models, Scout and Maverick, has stirred the competitive landscape of AI. Maverick’s claims of superiority over established models like GPT-4o and Gemini 2.0 Flash raise questions about evaluation fairness,…
-
Simon Willison’s Weblog: Long context support in LLM 0.24 using fragments and template plugins
Source URL: https://simonwillison.net/2025/Apr/7/long-context-llm/#atom-everything Source: Simon Willison’s Weblog Title: Long context support in LLM 0.24 using fragments and template plugins Feedly Summary: LLM 0.24 is now available with new features to help take advantage of the increasingly long input context supported by modern LLMs. (LLM is my command-line tool and Python library for interacting with LLMs,…