Follow Slashdot stories on Twitter

 



Forgot your password?
typodupeerror
×
AI Security

Hackers Call Current AI Security Testing 'Bullshit' 50

Leading cybersecurity researchers at DEF CON, the world's largest hacker conference, have warned that current methods for securing AI systems are fundamentally flawed and require a complete rethink, according to the conference's inaugural "Hackers' Almanack" report [PDF].

The report, produced with the University of Chicago's Cyber Policy Initiative, challenges the effectiveness of "red teaming" -- where security experts probe AI systems for vulnerabilities -- saying this approach alone cannot adequately protect against emerging threats. "Public red teaming an AI model is not possible because documentation for what these models are supposed to even do is fragmented and the evaluations we include in the documentation are inadequate," said Sven Cattell, who leads DEF CON's AI Village.

Nearly 500 participants tested AI models at the conference, with even newcomers successfully finding vulnerabilities. The researchers called for adopting frameworks similar to the Common Vulnerabilities and Exposures (CVE) system used in traditional cybersecurity since 1999. This would create standardized ways to document and address AI vulnerabilities, rather than relying on occasional security audits.

Hackers Call Current AI Security Testing 'Bullshit'

Comments Filter:
  • For good or bad, these AIs don't not work like traditional software, you can't treat them the same. Yes it is a problem, but I don't see how CVE could work.

    • Agreed developing powerful AI is a fundamental change in software, like the phase change of going from liquid to steam when you boil water. There is no way conventional security practices by themselves will stop intelligent software from ultimately doing what it wants. Nor will they stop human attackers from compromising systems with huge poorly-understood attack surfaces (like from prompt injection).

      The deeper issue is that even when AI is "secure", it is being designed and wielded by powerful and wealth p

      • 1. Accept that AI systems will need to be policed based on their behavior the same way humans are, through the same sorts of things that Lawrence Lessig suggests in Code 2.0 shapes a lot of human behavior: rules, norms, prices, and architecture (both environmental and neural in this AI case). Regulating is a big issue that all of society will need to be involved in -- including the social sciences. (I read an essay about this recently, forget off-hand by whom.

        AI does not have agency and can't be "policed" "the same way humans are". Neither do I support the regulation of bags of weights. This will only aggregate power into the hands of corporations. AI is and will always be controlled via liability incurred by those with agency over it.

        2. Ensure that OpenAI (and any similar AI non-profit) stays true to its moral and legal roots as a non-profit with a mission of ensuring AI is open and accessible to all and is used for humane ends and devotes any financial assets it acquires to those ends. Ensure that there is no "self-dealing" involving key people at OpenAI. Related by me on the larger issue:

        What matters is the underlying technology be open not the service itself, financial BS or corporate mission statements.

        3. Recognize that any sufficiently advanced AI should have rights (a complex topic).

        Absolutely not, there is nothing more dangerous than this BS. Computers, algorithms, AIs are nothing more th

  • by david.emery ( 127135 ) on Tuesday February 11, 2025 @02:22PM (#65159741)

    As long as we accept systems with vulnerabilities that have to be discovered and patched, we'll be in this continuing doom loop. I've been critical of my university for its very successful (as measured by 'funding' and 'enrollment') computer security program, because it doesn't start with the fundamental premise that software should be constructed without vulnerabilities in the first place.

    But as any consultant will tell you, "If you can't solve the problem, there's lots of money involved in continuing to discuss it."

    • Is it even theoretically possible to construct software without vulnerabilities? Especially if you don't also construct the hardware it runs on? And control the environment it's run in?

      How do you differentiate software that truly has no vulnerabilities from software with vulnerabilities that simply haven't been discovered yet?
      • Well, it's kinda like aviation safety. What's the chance of a remaining bug? The commercial avionics standard is 10 ^ -9, and they have a lot of (expensive) techniques to achieve that.

        But we KNOW a lot of the vulnerabilities in coding, and in design. So starting with better programming languages, better coding techniques, code analysis (including proof-of-correctness, at least proof that certain kinds of bugs/problems don't exist, such as memory leaks or access uninitialized memory), and increased emphas

        • Doesn't aviation safety achieve those levels, at least in part, by controlling both the hardware the software is run on and the environment that it's run in, as much as possible? I imagine the chance of a remaining bugs goes up considerably if replacement parts are allowed to be sourced from eBay, or passengers with soldering irons gain access to the avionics bay.

          I'm certainly not advocating an ostrich approach here, but how do we apply that principle to software that can be downloaded and run on arbitr
          • Doesn't aviation safety achieve those levels, at least in part, by controlling both the hardware the software is run on and the environment that it's run in, as much as possible?
            From a system perspective, yes. But avionics software development uses a particular development methodology that is explicitly focused on preventing defects (where 'defect' is "doesn't implement precisely the specification".) In the most extreme cases, this has to look at all the branches of execution at the machine code/assembly

            • While I would indeed argue "You can't prove software to be 100% free of bugs", my follow-up would then be "therefore, prevention REQUIRES detection, rather than prevention REPLACING detection". One cannot focus on prevention without adequate detection capabilities....

              Prevention should always be the goal. But I suppose my point is: we should also acknowledge that goal might not be entirely achievable, and plan accordingly...
    • by dfghjk ( 711126 )

      "If you can't solve the problem, there's lots of money involved in continuing to discuss it."

      That's true even if you can solve the problem, or if you don't even know what the problem is, as is the case here.

  • This isn't really on-topic, but I do miss going to Def Con.

    Do people still play 'spot the fed'?

    Yes, I could still go. I just don't have much of a reason to go.

  • by DarkOx ( 621550 ) on Tuesday February 11, 2025 @02:39PM (#65159805) Journal

    CVEs or something similar would be fine. I mean why not; can't hurt to have a uniform reporting standard for know problems around specific models and host software, and integration software.

    but...

    If what we are talking about is LLMs, LLMs + RAG, and LLMs plus lets bolt it some of our APIs we already have and call it a customer service agent - well I don't think we really need anythign new.

    99% of the vulnerabilities fall into the same classes of issues you have LLM or no LLM, - CSRF, Authorization failures (object references etc), SSRF, content injection, service (sql and others) injection, etc. Just because an LLM or NLP thing-y transformed some imports instread of some JavaScript code before they got reflected out where and in fashion they can do something unintended does not fundamentally change anything.

    If you are sharing data not all user principles have access to in the LLMs context or in stuff it can access via rag without the current users session tokens/keys/whatever and hoping some system prompt will keep it from disclosing whatever well okay, you're an idiot. If you don't understand why in-band signalling and security don't mix there is no help for you.

    Where this gets a lot more interesting is if your model actually gets trained on the application users's data, ie new weights get created etc, not RAG. That opens up a whole lot new potential security considerations but really that is NOT 99% of the industry is doing, and where they are they are doing it with a high trust user pool, so not sure we are ready for a new discipline here in terms of need.

    Finally if you look at OWASP and NISTs efforts on this so far there is tone of stuff they are trying to classify as security issues that simply are not. Bias is not a security issue, most of the time. If you are trying to spot suspicious bulges to identify people carrying guns - ok it could be; but that is just your basic type-1, type-2 error problem again, if you are building something like that you know that is potential problem, you'd test for it specifically, not as part of security but as part of basic fitness testing. The rest of the time it is not the domain of security practitioners to decide of if the LLMs output might be 'offensive to the aboriginal population of ' that is broader organizational question and again belongs in QA land, not security land.

    I just don't dont see AU security testing as justifiably special unless you are actually ingesting raw data and training something.

    • You actually missed the concern.

      red-teaming isn't interested in any problems with how any particular inference engine may secure tool use by the ML, or whether or not rawdog makes sure the scripts it generates are actually safe to run.

      Their goal is to find ways to make the LLM circumvent fine-tuned safety training- i.e., refusing to do certain things.
      That is what they're wondering if a CVE is really sufficient for, and that kind of "vulnerability" isn't really contained in the list of "CSRF, Authorizat
  • "Program testing can be used to show the presence of bugs, but never to show their absence!" --Edsgar W. Dijkstra, 1970

  • I've been to enough Def Con's (8) to know. The claim that security is bullshit has been pretty constant. When I watched the EFF crack DES hashes in under a minute it started to sink in: yep, most security is bullshit. The most honest folks are safe-makers who don't claim to create uncrackable or unhackable systems. They just rate their safes in terms of how long it'll take to break in.

...though his invention worked superbly -- his theory was a crock of sewage from beginning to end. -- Vernor Vinge, "The Peace War"

Working...