Tag: vision-llms
-
Simon Willison’s Weblog: Trying out QvQ – Qwen’s new visual reasoning model
Source URL: https://simonwillison.net/2024/Dec/24/qvq/#atom-everything Source: Simon Willison’s Weblog Title: Trying out QvQ – Qwen’s new visual reasoning model Feedly Summary: I thought we were done for major model releases in 2024, but apparently not: Alibaba’s Qwen team just dropped the Apache2 2 licensed QvQ-72B-Preview, “an experimental research model focusing on enhancing visual reasoning capabilities". Their blog…
-
Simon Willison’s Weblog: Gemini 2.0 Flash: An outstanding multi-modal LLM with a sci-fi streaming mode
Source URL: https://simonwillison.net/2024/Dec/11/gemini-2/#atom-everything Source: Simon Willison’s Weblog Title: Gemini 2.0 Flash: An outstanding multi-modal LLM with a sci-fi streaming mode Feedly Summary: Huge announcment from Google this morning: Introducing Gemini 2.0: our new AI model for the agentic era. There’s a ton of stuff in there (including updates on Project Astra and the new Project…
-
Simon Willison’s Weblog: SmolVLM – small yet mighty Vision Language Model
Source URL: https://simonwillison.net/2024/Nov/28/smolvlm/#atom-everything Source: Simon Willison’s Weblog Title: SmolVLM – small yet mighty Vision Language Model Feedly Summary: SmolVLM – small yet mighty Vision Language Model I’ve been having fun playing with this new vision model from the Hugging Face team behind SmolLM. They describe it as: […] a 2B VLM, SOTA for its memory…
-
Simon Willison’s Weblog: Say hello to gemini-exp-1121
Source URL: https://simonwillison.net/2024/Nov/22/gemini-exp-1121/#atom-everything Source: Simon Willison’s Weblog Title: Say hello to gemini-exp-1121 Feedly Summary: Say hello to gemini-exp-1121 Google Gemini’s Logan Kilpatrick on Twitter: Say hello to gemini-exp-1121! Our latest experimental gemini model, with: significant gains on coding performance stronger reasoning capabilities improved visual understanding Available on Google AI Studio and the Gemini API right…
-
Simon Willison’s Weblog: Pixtral Large
Source URL: https://simonwillison.net/2024/Nov/18/pixtral-large/ Source: Simon Willison’s Weblog Title: Pixtral Large Feedly Summary: Pixtral Large New today from Mistral: Today we announce Pixtral Large, a 124B open-weights multimodal model built on top of Mistral Large 2. Pixtral Large is the second model in our multimodal family and demonstrates frontier-level image understanding. The weights are out on…
-
Simon Willison’s Weblog: Claude API: PDF support (beta)
Source URL: https://simonwillison.net/2024/Nov/1/claude-api-pdf-support-beta/#atom-everything Source: Simon Willison’s Weblog Title: Claude API: PDF support (beta) Feedly Summary: Claude API: PDF support (beta) Claude 3.5 Sonnet now accepts PDFs as attachments: The new Claude 3.5 Sonnet (claude-3-5-sonnet-20241022) model now supports PDF input and understands both text and visual content within documents. I just released llm-claude-3 0.7 with support…
-
Simon Willison’s Weblog: You can now run prompts against images, audio and video in your terminal using LLM
Source URL: https://simonwillison.net/2024/Oct/29/llm-multi-modal/#atom-everything Source: Simon Willison’s Weblog Title: You can now run prompts against images, audio and video in your terminal using LLM Feedly Summary: I released LLM 0.17 last night, the latest version of my combined CLI tool and Python library for interacting with hundreds of different Large Language Models such as GPT-4o, Llama,…
-
Simon Willison’s Weblog: LLM Pictionary
Source URL: https://simonwillison.net/2024/Oct/26/llm-pictionary/ Source: Simon Willison’s Weblog Title: LLM Pictionary Feedly Summary: LLM Pictionary Inspired by my SVG pelicans on a bicycle, Paul Calcraft built this brilliant system where different vision LLMs can play Pictionary with each other, taking it in turns to progressively draw SVGs while the other models see if they can guess…