Tag: Tags:

  • Simon Willison’s Weblog: Apple Is Delaying the ‘More Personalized Siri’ Apple Intelligence Features

    Source URL: https://simonwillison.net/2025/Mar/8/delaying-personalized-siri/#atom-everything Source: Simon Willison’s Weblog Title: Apple Is Delaying the ‘More Personalized Siri’ Apple Intelligence Features Feedly Summary: Apple Is Delaying the ‘More Personalized Siri’ Apple Intelligence Features Apple told John Gruber (and other Apple press) this about the new “personalized" Siri: It’s going to take us longer than we thought to deliver…

  • Simon Willison’s Weblog: State-of-the-art text embedding via the Gemini API

    Source URL: https://simonwillison.net/2025/Mar/7/gemini-embeddings/#atom-everything Source: Simon Willison’s Weblog Title: State-of-the-art text embedding via the Gemini API Feedly Summary: State-of-the-art text embedding via the Gemini API Gemini just released their new text embedding model, with the snappy name gemini-embedding-exp-03-07. It supports 8,000 input tokens – up from 3,000 – and outputs vectors that are a lot larger…

  • Simon Willison’s Weblog: Mistral OCR

    Source URL: https://simonwillison.net/2025/Mar/7/mistral-ocr/#atom-everything Source: Simon Willison’s Weblog Title: Mistral OCR Feedly Summary: Mistral OCR New closed-source specialist OCR model by Mistral – you can feed it images or a PDF and it produces Markdown with optional embedded images. It’s available via their API, or it’s “available to self-host on a selective basis" for people with…

  • Simon Willison’s Weblog: Demo of ChatGPT Code Interpreter running in o3-mini-high

    Source URL: https://simonwillison.net/2025/Mar/5/code-interpreter/ Source: Simon Willison’s Weblog Title: Demo of ChatGPT Code Interpreter running in o3-mini-high Feedly Summary: Demo of ChatGPT Code Interpreter running in o3-mini-high OpenAI made GPT-4.5 available to Plus ($20/month) users today. I was a little disappointed with GPT-4.5 when I tried it through the API, but having access in the ChatGPT…

  • Simon Willison’s Weblog: Career Update: Google DeepMind -> Anthropic

    Source URL: https://simonwillison.net/2025/Mar/5/google-deepmind-anthropic/ Source: Simon Willison’s Weblog Title: Career Update: Google DeepMind -> Anthropic Feedly Summary: Career Update: Google DeepMind -> Anthropic Nicholas Carlini (previously) on joining Anthropic, driven partly by his frustration at friction he encountered publishing his research at Google DeepMind after their merge with Google Brain. His area of expertise is adversarial…

  • Simon Willison’s Weblog: QwQ-32B: Embracing the Power of Reinforcement Learning

    Source URL: https://simonwillison.net/2025/Mar/5/qwq-32b/#atom-everything Source: Simon Willison’s Weblog Title: QwQ-32B: Embracing the Power of Reinforcement Learning Feedly Summary: QwQ-32B: Embracing the Power of Reinforcement Learning New Apache 2 licensed reasoning model from Qwen: We are excited to introduce QwQ-32B, a model with 32 billion parameters that achieves performance comparable to DeepSeek-R1, which boasts 671 billion parameters…

  • Simon Willison’s Weblog: A Practical Guide to Implementing DeepSearch / DeepResearch

    Source URL: https://simonwillison.net/2025/Mar/4/deepsearch-deepresearch/#atom-everything Source: Simon Willison’s Weblog Title: A Practical Guide to Implementing DeepSearch / DeepResearch Feedly Summary: A Practical Guide to Implementing DeepSearch / DeepResearch I really like the definitions Han Xiao from Jina AI proposes for the terms DeepSearch and DeepResearch in this piece: DeepSearch runs through an iterative loop of searching, reading,…

  • Simon Willison’s Weblog: llm-mistral 0.11

    Source URL: https://simonwillison.net/2025/Mar/4/llm-mistral-011/#atom-everything Source: Simon Willison’s Weblog Title: llm-mistral 0.11 Feedly Summary: llm-mistral 0.11 I added schema support to this plugin which adds support for the Mistral API to LLM. Release notes: Support for LLM schemas. #19 -o prefix ‘{‘ option for forcing a response prefix. #18 Schemas now work with OpenAI, Anthropic, Gemini and…

  • Simon Willison’s Weblog: Quoting Ethan Mollick

    Source URL: https://simonwillison.net/2025/Mar/2/ethan-mollick/ Source: Simon Willison’s Weblog Title: Quoting Ethan Mollick Feedly Summary: After publishing this piece, I was contacted by Anthropic who told me that Sonnet 3.7 would not be considered a 10^26 FLOP model and cost a few tens of millions of dollars to train, though future models will be much bigger. —…