• Автор темы News
  • Дата начала
  • " /> News - Researchers claim breakthrough in fight against AI’s frustrating security hole | SoftoolStore.de - Программное обеспечение, Avid Media Composer, Книги, Новости, Windows, Интернет-новости, Бесплатные прокси (HTTP, Socks 4, Socks 5)

    News Researchers claim breakthrough in fight against AI’s frustrating security hole

    News

    Команда форума
    Редактор
    Регистрация
    17 Февраль 2018
    Сообщения
    36 066
    Лучшие ответы
    0
    Баллы
    2 093
    Offline
    #1
    In the AI world, a vulnerability called "prompt injection" has haunted developers since chatbots went mainstream in 2022. Despite numerous attempts to solve this fundamental vulnerability—the digital equivalent of whispering secret instructions to override a system's intended behavior—no one has found a reliable solution. Until now, perhaps.

    Google DeepMind has unveiled CaMeL (CApabilities for MachinE Learning), a new approach to stopping prompt-injection attacks that abandons the failed strategy of having AI models police themselves. Instead, CaMeL treats language models as fundamentally untrusted components within a secure software framework, creating clear boundaries between user commands and potentially malicious content.

    Prompt injection has created a significant barrier to building trustworthy AI assistants, which may be why general-purpose big tech AI like Apple's Siri doesn't currently work like ChatGPT. As AI agents get integrated into email, calendar, banking, and document-editing processes, the consequences of prompt injection have shifted from hypothetical to existential. When agents can send emails, move money, or schedule appointments, a misinterpreted string isn't just an error—it's a dangerous exploit.

    Read full article

    Comments
     
    Сверху Снизу