Tag: prompt injection attack

  • The Register: Google to Iran: Yes, we see you using Gemini for phishing and scripting. We’re onto you

    Source URL: https://www.theregister.com/2025/01/31/state_spies_google_gemini/ Source: The Register Title: Google to Iran: Yes, we see you using Gemini for phishing and scripting. We’re onto you Feedly Summary: And you, China, Russia, North Korea … Guardrails block malware generation Google says it’s spotted Chinese, Russian, Iranian, and North Korean government agents using its Gemini AI for nefarious purposes,…

  • Simon Willison’s Weblog: How we estimate the risk from prompt injection attacks on AI systems

    Source URL: https://simonwillison.net/2025/Jan/29/prompt-injection-attacks-on-ai-systems/ Source: Simon Willison’s Weblog Title: How we estimate the risk from prompt injection attacks on AI systems Feedly Summary: How we estimate the risk from prompt injection attacks on AI systems The “Agentic AI Security Team" at Google DeepMind share some details on how they are researching indirect prompt injection attacks. They…

  • Google Online Security Blog: How we estimate the risk from prompt injection attacks on AI systems

    Source URL: https://security.googleblog.com/2025/01/how-we-estimate-risk-from-prompt.html Source: Google Online Security Blog Title: How we estimate the risk from prompt injection attacks on AI systems Feedly Summary: AI Summary and Description: Yes Summary: The text discusses emerging security challenges in modern AI systems, specifically focusing on a class of attacks called “indirect prompt injection.” It presents a comprehensive evaluation…

  • Simon Willison’s Weblog: ChatGPT Operator system prompt

    Source URL: https://simonwillison.net/2025/Jan/26/chatgpt-operator-system-prompt/#atom-everything Source: Simon Willison’s Weblog Title: ChatGPT Operator system prompt Feedly Summary: ChatGPT Operator system prompt Johann Rehberger snagged a copy of the ChatGPT Operator system prompt. As usual, the system prompt doubles as better written documentation than any of the official sources. It asks users for confirmation a lot: ## Confirmations Ask…

  • Simon Willison’s Weblog: Introducing Operator

    Source URL: https://simonwillison.net/2025/Jan/23/introducing-operator/ Source: Simon Willison’s Weblog Title: Introducing Operator Feedly Summary: Introducing Operator OpenAI released their “research preview" today of Operator, a cloud-based browser automation platform rolling out today to $200/month ChatGPT Pro subscribers. They’re calling this their first "agent". In the Operator announcement video Sam Altman defined that notoriously vague term like this:…

  • Simon Willison’s Weblog: Lessons From Red Teaming 100 Generative AI Products

    Source URL: https://simonwillison.net/2025/Jan/18/lessons-from-red-teaming/ Source: Simon Willison’s Weblog Title: Lessons From Red Teaming 100 Generative AI Products Feedly Summary: Lessons From Red Teaming 100 Generative AI Products New paper from Microsoft describing their top eight lessons learned red teaming (deliberately seeking security vulnerabilities in) 100 different generative AI models and products over the past few years.…

  • CSA: AI and Compliance for the Mid-Market

    Source URL: https://www.scrut.io/post/ai-and-compliance-for-the-mid-market Source: CSA Title: AI and Compliance for the Mid-Market Feedly Summary: AI Summary and Description: Yes **Summary:** The text emphasizes the urgent need for small and medium-sized businesses (SMBs) to adopt AI responsibly, given the potential cybersecurity vulnerabilities and evolving regulatory landscape associated with AI technologies. It outlines practical guidance and standards…

  • CSA: LLM Dragons: Why DSPM is the Key to AI Security

    Source URL: https://cloudsecurityalliance.org/articles/training-your-llm-dragons-why-dspm-is-the-key-to-ai-security Source: CSA Title: LLM Dragons: Why DSPM is the Key to AI Security Feedly Summary: AI Summary and Description: Yes Summary: The text emphasizes the security risks associated with AI implementations, particularly custom large language models (LLMs) and Microsoft Copilot. It outlines key threats such as data leakage and compliance failures and…

  • The Register: Just as your LLM once again goes off the rails, Cisco, Nvidia are at the door smiling

    Source URL: https://www.theregister.com/2025/01/17/nvidia_cisco_ai_guardrails_security/ Source: The Register Title: Just as your LLM once again goes off the rails, Cisco, Nvidia are at the door smiling Feedly Summary: Some of you have apparently already botched chatbots or allowed ‘shadow AI’ to creep in Cisco and Nvidia have both recognized that as useful as today’s AI may be,…