Tag: vision-llms
-
Simon Willison’s Weblog: GPT-4.1: Three new million token input models from OpenAI, including their cheapest model yet
Source URL: https://simonwillison.net/2025/Apr/14/gpt-4-1/ Source: Simon Willison’s Weblog Title: GPT-4.1: Three new million token input models from OpenAI, including their cheapest model yet Feedly Summary: OpenAI introduced three new models this morning: GPT-4.1, GPT-4.1 mini and GPT-4.1 nano. These are API-only models right now, not available through the ChatGPT interface (though you can try them out…
-
Simon Willison’s Weblog: Mistral Small 3.1 on Ollama
Source URL: https://simonwillison.net/2025/Apr/8/mistral-small-31-on-ollama/#atom-everything Source: Simon Willison’s Weblog Title: Mistral Small 3.1 on Ollama Feedly Summary: Mistral Small 3.1 on Ollama Mistral Small 3.1 (previously) is now available through Ollama, providing an easy way to run this multi-modal (vision) model on a Mac (and other platforms, though I haven’t tried them myself yet). I had to…
-
Simon Willison’s Weblog: Putting Gemini 2.5 Pro through its paces
Source URL: https://simonwillison.net/2025/Mar/25/gemini/ Source: Simon Willison’s Weblog Title: Putting Gemini 2.5 Pro through its paces Feedly Summary: There’s a new release from Google Gemini this morning: the first in the Gemini 2.5 series. Google call it “a thinking model, designed to tackle increasingly complex problems". It’s already sat at the top of the LM Arena…
-
Simon Willison’s Weblog: Mistral Small 3.1
Source URL: https://simonwillison.net/2025/Mar/17/mistral-small-31/#atom-everything Source: Simon Willison’s Weblog Title: Mistral Small 3.1 Feedly Summary: Mistral Small 3.1 Mistral Small 3 came out in January and was a notable, genuinely excellent local model that used an Apache 2.0 license. Mistral Small 3.1 offers a significant improvement: it’s multi-modal (images) and has an increased 128,000 token context length,…
-
Simon Willison’s Weblog: Notes on Google’s Gemma 3
Source URL: https://simonwillison.net/2025/Mar/12/gemma-3/ Source: Simon Willison’s Weblog Title: Notes on Google’s Gemma 3 Feedly Summary: Google’s Gemma team released an impressive new model today (under their not-open-source Gemma license). Gemma 3 comes in four sizes – 1B, 4B, 12B, and 27B – and while 1B is text-only the larger three models are all multi-modal for…
-
Simon Willison’s Weblog: What’s new in the world of LLMs, for NICAR 2025
Source URL: https://simonwillison.net/2025/Mar/8/nicar-llms/ Source: Simon Willison’s Weblog Title: What’s new in the world of LLMs, for NICAR 2025 Feedly Summary: I presented two sessions at the NICAR 2025 data journalism conference this year. The first was this one based on my review of LLMs in 2024, extended by several months to cover everything that’s happened…
-
Simon Willison’s Weblog: Mistral OCR
Source URL: https://simonwillison.net/2025/Mar/7/mistral-ocr/#atom-everything Source: Simon Willison’s Weblog Title: Mistral OCR Feedly Summary: Mistral OCR New closed-source specialist OCR model by Mistral – you can feed it images or a PDF and it produces Markdown with optional embedded images. It’s available via their API, or it’s “available to self-host on a selective basis" for people with…
-
Simon Willison’s Weblog: llm-ollama 0.9.0
Source URL: https://simonwillison.net/2025/Mar/4/llm-ollama-090/ Source: Simon Willison’s Weblog Title: llm-ollama 0.9.0 Feedly Summary: llm-ollama 0.9.0 This release of the llm-ollama plugin adds support for schemas, thanks to a PR by Adam Compton. Ollama provides very robust support for this pattern thanks to their structured outputs feature, which works across all of the models that they support…
-
Simon Willison’s Weblog: olmOCR
Source URL: https://simonwillison.net/2025/Feb/26/olmocr/#atom-everything Source: Simon Willison’s Weblog Title: olmOCR Feedly Summary: olmOCR New from Ai2 – olmOCR is “an open-source tool designed for high-throughput conversion of PDFs and other documents into plain text while preserving natural reading order". At its core is allenai/olmOCR-7B-0225-preview, a Qwen2-VL-7B-Instruct variant trained on ~250,000 pages of diverse PDF content (both…
-
Simon Willison’s Weblog: Quoting Mark Zuckerberg
Source URL: https://simonwillison.net/2025/Jan/30/mark-zuckerberg/#atom-everything Source: Simon Willison’s Weblog Title: Quoting Mark Zuckerberg Feedly Summary: Llama 4 is making great progress in training. Llama 4 mini is done with pre-training and our reasoning models and larger model are looking good too. Our goal with Llama 3 was to make open source competitive with closed models, and our…