Tag: prompt injections

  • The Register: Microsoft dangles $10K for hackers to hijack LLM email service

    Source URL: https://www.theregister.com/2024/12/09/microsoft_llm_prompt_injection_challenge/ Source: The Register Title: Microsoft dangles $10K for hackers to hijack LLM email service Feedly Summary: Outsmart an AI, win a little Christmas cash Microsoft and friends have challenged AI hackers to break a simulated LLM-integrated email client with a prompt injection attack – and the winning teams will share a $10,000…

  • Embrace The Red: Terminal DiLLMa: LLM-powered Apps Can Hijack Your Terminal Via Prompt Injection

    Source URL: https://embracethered.com/blog/posts/2024/terminal-dillmas-prompt-injection-ansi-sequences/ Source: Embrace The Red Title: Terminal DiLLMa: LLM-powered Apps Can Hijack Your Terminal Via Prompt Injection Feedly Summary: Last week Leon Derczynski described how LLMs can output ANSI escape codes. These codes, also known as control characters, are interpreted by terminal emulators and modify behavior. This discovery resonates with areas I had…

  • Simon Willison’s Weblog: 0xfreysa/agent

    Source URL: https://simonwillison.net/2024/Nov/29/0xfreysaagent/#atom-everything Source: Simon Willison’s Weblog Title: 0xfreysa/agent Feedly Summary: 0xfreysa/agent Freysa describes itself as “the world’s first adversarial agent game". On 22nd November they released an LLM-driven application which people could pay to message (using Ethereum), with access to tools that could transfer a prize pool to the message sender, ending the game.…

  • Hacker News: The Beginner’s Guide to Visual Prompt Injections

    Source URL: https://www.lakera.ai/blog/visual-prompt-injections Source: Hacker News Title: The Beginner’s Guide to Visual Prompt Injections Feedly Summary: Comments AI Summary and Description: Yes Summary: The text discusses security vulnerabilities inherent in Large Language Models (LLMs), particularly focusing on visual prompt injections. As the reliance on models like GPT-4 increases for various tasks, concerns regarding the potential…

  • Schneier on Security: Prompt Injection Defenses Against LLM Cyberattacks

    Source URL: https://www.schneier.com/blog/archives/2024/11/prompt-injection-defenses-against-llm-cyberattacks.html Source: Schneier on Security Title: Prompt Injection Defenses Against LLM Cyberattacks Feedly Summary: Interesting research: “Hacking Back the AI-Hacker: Prompt Injection as a Defense Against LLM-driven Cyberattacks“: Large language models (LLMs) are increasingly being harnessed to automate cyberattacks, making sophisticated exploits more accessible and scalable. In response, we propose a new defense…

  • Wired: This Prompt Can Make an AI Chatbot Identify and Extract Personal Details From Your Chats

    Source URL: https://www.wired.com/story/ai-imprompter-malware-llm/ Source: Wired Title: This Prompt Can Make an AI Chatbot Identify and Extract Personal Details From Your Chats Feedly Summary: Security researchers created an algorithm that turns a malicious prompt into a set of hidden instructions that could send a user’s personal information to an attacker. AI Summary and Description: Yes Summary:…

  • Cloud Blog: Testing your LLMs differently: Security updates from our latest Cyber Snapshot Report

    Source URL: https://cloud.google.com/blog/products/identity-security/testing-your-llms-differently-security-updates-from-our-latest-cyber-snapshot-report/ Source: Cloud Blog Title: Testing your LLMs differently: Security updates from our latest Cyber Snapshot Report Feedly Summary: Web-based large-language models (LLM) are revolutionizing how we interact online. Instead of well-defined and structured queries, people can engage with applications and systems in a more natural and conversational manner — and the applications…