Tag: assisted programming

  • Simon Willison’s Weblog: Quoting Steve Yegge

    Source URL: https://simonwillison.net/2025/Mar/9/steve-yegge/ Source: Simon Willison’s Weblog Title: Quoting Steve Yegge Feedly Summary: I’ve been using Claude Code for a couple of days, and it has been absolutely ruthless in chewing through legacy bugs in my gnarly old code base. It’s like a wood chipper fueled by dollars. It can power through shockingly impressive tasks,…

  • Simon Willison’s Weblog: Demo of ChatGPT Code Interpreter running in o3-mini-high

    Source URL: https://simonwillison.net/2025/Mar/5/code-interpreter/ Source: Simon Willison’s Weblog Title: Demo of ChatGPT Code Interpreter running in o3-mini-high Feedly Summary: Demo of ChatGPT Code Interpreter running in o3-mini-high OpenAI made GPT-4.5 available to Plus ($20/month) users today. I was a little disappointed with GPT-4.5 when I tried it through the API, but having access in the ChatGPT…

  • Simon Willison’s Weblog: Quoting Catherine Wu

    Source URL: https://simonwillison.net/2025/Feb/24/catherine-wu/ Source: Simon Willison’s Weblog Title: Quoting Catherine Wu Feedly Summary: We find that Claude is really good at test driven development, so we often ask Claude to write tests first and then ask Claude to iterate against the tests. — Catherine Wu, Anthropic Tags: anthropic, claude, ai-assisted-programming, generative-ai, ai, llms, testing, tdd…

  • Simon Willison’s Weblog: My LLM codegen workflow atm

    Source URL: https://simonwillison.net/2025/Feb/21/my-llm-codegen-workflow-atm/#atom-everything Source: Simon Willison’s Weblog Title: My LLM codegen workflow atm Feedly Summary: My LLM codegen workflow atm Harper Reed describes his workflow for writing code with the assistance of LLMs. This is clearly a very well-thought out process, which has evolved a lot already and continues to change. Harper starts greenfield projects…

  • Simon Willison’s Weblog: Quoting Paul Gauthier

    Source URL: https://simonwillison.net/2025/Jan/26/paul-gauthier/ Source: Simon Willison’s Weblog Title: Quoting Paul Gauthier Feedly Summary: In my experience with AI coding, very large context windows aren’t useful in practice. Every model seems to get confused when you feed them more than ~25-30k tokens. The models stop obeying their system prompts, can’t correctly find/transcribe pieces of code in…

  • Simon Willison’s Weblog: Codestral 25.01

    Source URL: https://simonwillison.net/2025/Jan/13/codestral-2501/ Source: Simon Willison’s Weblog Title: Codestral 25.01 Feedly Summary: Codestral 25.01 Brand new code-focused model from Mistral. Unlike the first Codestral this one isn’t (yet) available as open weights. The model has a 256k token context – a new record for Mistral. The new model scored an impressive joint first place with…

  • Simon Willison’s Weblog: Building Python tools with a one-shot prompt using uv run and Claude Projects

    Source URL: https://simonwillison.net/2024/Dec/19/one-shot-python-tools/#atom-everything Source: Simon Willison’s Weblog Title: Building Python tools with a one-shot prompt using uv run and Claude Projects Feedly Summary: I’ve written a lot about how I’ve been using Claude to build one-shot HTML+JavaScript applications via Claude Artifacts. I recently started using a similar pattern to create one-shot Python utilities, using a…

  • Simon Willison’s Weblog: A new free tier for GitHub Copilot in VS Code

    Source URL: https://simonwillison.net/2024/Dec/18/free-tier-for-github-copilot/#atom-everything Source: Simon Willison’s Weblog Title: A new free tier for GitHub Copilot in VS Code Feedly Summary: A new free tier for GitHub Copilot in VS Code It’s easy to forget that GitHub Copilot was the first widely deployed feature built on top of generative AI, with its initial preview launching all…

  • Simon Willison’s Weblog: Phi-4 Technical Report

    Source URL: https://simonwillison.net/2024/Dec/15/phi-4-technical-report/ Source: Simon Willison’s Weblog Title: Phi-4 Technical Report Feedly Summary: Phi-4 Technical Report Phi-4 is the latest LLM from Microsoft Research. It has 14B parameters and claims to be a big leap forward in the overall Phi series. From Introducing Phi-4: Microsoft’s Newest Small Language Model Specializing in Complex Reasoning: Phi-4 outperforms…