Tag: system prompt
-
Simon Willison’s Weblog: Quoting Paul Gauthier
Source URL: https://simonwillison.net/2025/Jan/26/paul-gauthier/ Source: Simon Willison’s Weblog Title: Quoting Paul Gauthier Feedly Summary: In my experience with AI coding, very large context windows aren’t useful in practice. Every model seems to get confused when you feed them more than ~25-30k tokens. The models stop obeying their system prompts, can’t correctly find/transcribe pieces of code in…
-
Simon Willison’s Weblog: ChatGPT Operator system prompt
Source URL: https://simonwillison.net/2025/Jan/26/chatgpt-operator-system-prompt/#atom-everything Source: Simon Willison’s Weblog Title: ChatGPT Operator system prompt Feedly Summary: ChatGPT Operator system prompt Johann Rehberger snagged a copy of the ChatGPT Operator system prompt. As usual, the system prompt doubles as better written documentation than any of the official sources. It asks users for confirmation a lot: ## Confirmations Ask…
-
Chip Huyen: Common pitfalls when building generative AI applications
Source URL: https://huyenchip.com//2025/01/16/ai-engineering-pitfalls.html Source: Chip Huyen Title: Common pitfalls when building generative AI applications Feedly Summary: As we’re still in the early days of building applications with foundation models, it’s normal to make mistakes. This is a quick note with examples of some of the most common pitfalls that I’ve seen, both from public case…
-
Embrace The Red: Microsoft 365 Copilot Generated Images Accessible Without Authentication — Fixed!
Source URL: https://embracethered.com/blog/posts/2025/m365-copilot-image-generation-without-authentication/ Source: Embrace The Red Title: Microsoft 365 Copilot Generated Images Accessible Without Authentication — Fixed! Feedly Summary: I regularly look at how the system prompts of chatbots change over time. Updates frequently highlight new features being added, design changes that occur and potential areas that might benefit from more security scrutiny. A…
-
Simon Willison’s Weblog: Security ProbLLMs in xAI’s Grok: A Deep Dive
Source URL: https://simonwillison.net/2024/Dec/16/security-probllms-in-xais-grok/#atom-everything Source: Simon Willison’s Weblog Title: Security ProbLLMs in xAI’s Grok: A Deep Dive Feedly Summary: Security ProbLLMs in xAI’s Grok: A Deep Dive Adding xAI to the growing list of AI labs that shipped feature vulnerable to data exfiltration prompt injection attacks, but with the unfortunate addendum that they don’t seem to…
-
Simon Willison’s Weblog: PydanticAI
Source URL: https://simonwillison.net/2024/Dec/2/pydanticai/#atom-everything Source: Simon Willison’s Weblog Title: PydanticAI Feedly Summary: PydanticAI New project from Pydantic, which they describe as an “Agent Framework / shim to use Pydantic with LLMs". I asked which agent definition they are using and it’s the "system prompt with bundled tools" one. To their credit, they explain that in their…
-
Simon Willison’s Weblog: Leaked system prompts from Vercel v0
Source URL: https://simonwillison.net/2024/Nov/25/leaked-system-prompts-from-vercel-v0/#atom-everything Source: Simon Willison’s Weblog Title: Leaked system prompts from Vercel v0 Feedly Summary: Leaked system prompts from Vercel v0 v0 is Vercel’s entry in the increasingly crowded LLM-assisted development market – chat with a bot and have that bot build a full application for you. They’ve been iterating on it since launching…