Tag: llms
-
Slashdot: AI-Generated Code Creates Major Security Risk Through ‘Package Hallucinations’
Source URL: https://developers.slashdot.org/story/25/04/29/1837239/ai-generated-code-creates-major-security-risk-through-package-hallucinations?utm_source=rss1.0mainlinkanon&utm_medium=feed Source: Slashdot Title: AI-Generated Code Creates Major Security Risk Through ‘Package Hallucinations’ Feedly Summary: AI Summary and Description: Yes Summary: The study highlights a critical vulnerability in AI-generated code, where a significant percentage of generated packages reference non-existent libraries, posing substantial risks for supply-chain attacks. This phenomenon is more prevalent in open…
-
Schneier on Security: Applying Security Engineering to Prompt Injection Security
Source URL: https://www.schneier.com/blog/archives/2025/04/applying-security-engineering-to-prompt-injection-security.html Source: Schneier on Security Title: Applying Security Engineering to Prompt Injection Security Feedly Summary: This seems like an important advance in LLM security against prompt injection: Google DeepMind has unveiled CaMeL (CApabilities for MachinE Learning), a new approach to stopping prompt-injection attacks that abandons the failed strategy of having AI models police…