Tag: risk factors

  • Schneier on Security: On Generative AI Security

    Source URL: https://www.schneier.com/blog/archives/2025/02/on-generative-ai-security.html Source: Schneier on Security Title: On Generative AI Security Feedly Summary: Microsoft’s AI Red Team just published “Lessons from Red Teaming 100 Generative AI Products.” Their blog post lists “three takeaways,” but the eight lessons in the report itself are more useful: Understand what the system can do and where it is…

  • Simon Willison’s Weblog: TextSynth Server

    Source URL: https://simonwillison.net/2024/Nov/21/textsynth-server/ Source: Simon Willison’s Weblog Title: TextSynth Server Feedly Summary: TextSynth Server I’d missed this: Fabrice Bellard (yes, that Fabrice Bellard) has a project called TextSynth Server which he describes like this: ts_server is a web server proposing a REST API to large language models. They can be used for example for text…

  • Cloud Blog: (In)tuned to Takeovers: Abusing Intune Permissions for Lateral Movement and Privilege Escalation in Entra ID Native Environments

    Source URL: https://cloud.google.com/blog/topics/threat-intelligence/abusing-intune-permissions-entra-id-environments/ Source: Cloud Blog Title: (In)tuned to Takeovers: Abusing Intune Permissions for Lateral Movement and Privilege Escalation in Entra ID Native Environments Feedly Summary: Written by: Thibault Van Geluwe de Berlaere, Karl Madden, Corné de Jong The Mandiant Red Team recently supported a client to visualize the possible impact of a compromise by…

  • METR Blog – METR: New Support Through The Audacious Project

    Source URL: https://metr.org/blog/2024-10-09-new-support-through-the-audacious-project/ Source: METR Blog – METR Title: New Support Through The Audacious Project Feedly Summary: AI Summary and Description: Yes Summary: The text discusses the Audacious Project’s funding initiative aimed at addressing global challenges through innovative solutions, particularly highlighting Project Canary’s focus on evaluating AI systems to ensure their safety and security. It…

  • Hacker News: Sabotage Evaluations for Frontier Models

    Source URL: https://www.anthropic.com/research/sabotage-evaluations Source: Hacker News Title: Sabotage Evaluations for Frontier Models Feedly Summary: Comments AI Summary and Description: Yes **Summary:** The text outlines a comprehensive series of evaluation techniques developed by the Anthropic Alignment Science team to assess potential sabotage capabilities in AI models. These evaluations are crucial for ensuring the safety and integrity…