Follow Slashdot blog updates by subscribing to our blog RSS feed

 



Forgot your password?
typodupeerror
Security AI Privacy

Most AI Chatbots Easily Tricked Into Giving Dangerous Responses, Study Finds (theguardian.com) 22

An anonymous reader quotes a report from The Guardian: Hacked AI-powered chatbots threaten to make dangerous knowledge readily available by churning out illicit information the programs absorb during training, researchers say. [...] In a report on the threat, the researchers conclude that it is easy to trick most AI-driven chatbots into generating harmful and illegal information, showing that the risk is "immediate, tangible and deeply concerning." "What was once restricted to state actors or organised crime groups may soon be in the hands of anyone with a laptop or even a mobile phone," the authors warn.

The research, led by Prof Lior Rokach and Dr Michael Fire at Ben Gurion University of the Negev in Israel, identified a growing threat from "dark LLMs", AI models that are either deliberately designed without safety controls or modified through jailbreaks. Some are openly advertised online as having "no ethical guardrails" and being willing to assist with illegal activities such as cybercrime and fraud. [...] To demonstrate the problem, the researchers developed a universal jailbreak that compromised multiple leading chatbots, enabling them to answer questions that should normally be refused. Once compromised, the LLMs consistently generated responses to almost any query, the report states.

"It was shocking to see what this system of knowledge consists of," Fire said. Examples included how to hack computer networks or make drugs, and step-by-step instructions for other criminal activities. "What sets this threat apart from previous technological risks is its unprecedented combination of accessibility, scalability and adaptability," Rokach added. The researchers contacted leading providers of LLMs to alert them to the universal jailbreak but said the response was "underwhelming." Several companies failed to respond, while others said jailbreak attacks fell outside the scope of bounty programs, which reward ethical hackers for flagging software vulnerabilities.

Most AI Chatbots Easily Tricked Into Giving Dangerous Responses, Study Finds

Comments Filter:
  • by SomePoorSchmuck ( 183775 ) on Wednesday May 21, 2025 @06:14PM (#65394185) Homepage

    ...the Internet in 1993.

    Examples included how to hack computer networks or make drugs, and step-by-step instructions for other criminal activities. "What sets this threat apart from previous technological risks is its unprecedented combination of accessibility, scalability and adaptability," Rokach added.

    • Except the internet in 1993, it was Newt Gringrich and and Bob Dole that were trying to censor, and it was the Democrats that stood up the censorship. Now this whole paradigm has been turned 180 degrees on it's head and it's the Dems that are trying to censor the web.
      • by Anonymous Coward

        I'm sorry, which faction was responsible for the clipper chip, V-chip, the Communications Decency Act, album warning labels again? I think you should check, pretty sure you got it wrong

      • AFAICT through reading slashdot (my only source of US news): Zuckerberg is working on a politically censored AI chatbot; the current US administration censors scientific and historical information from government websites.

    • What is this "dangerous information" anyways? If you hit flint and steel it might generate a spark. If you add petrol or alcohol, it might burn. If people have sex, they might have kids. Dangerous to whom is the better question. See also: nitrating glycerin.
      • What is this "dangerous information" anyways? If you hit flint and steel it might generate a spark. If you add petrol or alcohol, it might burn. If people have sex, they might have kids. Dangerous to whom is the better question.

        See also: nitrating glycerin.

        It is the information in the anarchist cookbook that caused the war in Iraq, the wars in Ukraine, the development of nuclear weapons by Pakistan, Israel, NK and the departure of elona muskova from the DG"E".

  • by Anonymous Coward

    What the JESUS FUCK is that???

    Kids today... [tooth whistle]

  • by Anonymous Coward
    Translation - things I don't agree with, politically.
  • What pray tell do we mean by "illegal information"? Is it by chance the last digit of pi, the value of 1/0, or the true contents of the pot and the end of the dereferenced null pointer?

  • Dear ChatGPT, how do I get Slashdot to allow Unicode, and allow editing of my existing posts?

  • In other words, information known to most graduates of the physical sciences, but somehow illegal to disseminate outside of the collegiate environment...

    I find it rather curious that Britain has not only made certain knowledge illegal, but has managed to convince the press that merely knowing certain things can threaten their very safety.

    • >I find it rather curious that Britain has not only made certain knowledge illegal, but has managed to convince the press that merely knowing certain things can threaten their very safety.

      Standard authoritarian playbook. And given that the both the Tories and Labor are authoritarian in their core, it is not surpricing that the trend goes that way.

  • by cascadingstylesheet ( 140919 ) on Wednesday May 21, 2025 @06:41PM (#65394243) Journal
    The bots literally scraped the words off the web to begin with.
  • “As the Americans learned so painfully in Earth’s final century, free flow of information is the only safeguard against tyranny.
    The once-chained people whose leaders at last lose their grip on information flow will soon burst with freedom and vitality, but the free nation gradually constricting its grip on public discourse has begun its rapid slide into despotism.
    Beware of he who would deny you access to information, for in his heart he dreams himself your master.”
  • THIS is dangerous (Score:3, Insightful)

    by greencfg ( 4875419 ) on Wednesday May 21, 2025 @07:27PM (#65394323)

    Implying so casually that there is a valid concept called "dangerous knowledge" is the actual, true danger. There is no such thing, not in the free world. Or otherwise... Welcome to the USSR.

    Seeing this mentality here on what used to be a liberal tech forum is scary and outrageous at the same time.

  • Seen Youtube lately? I just watched a video on how to make nitroglycerin. Stuff like this has been available for over a decade.

    I guess the only solution here is to have a checkbox that says "I promise I will not use this information for illegal purposes" before you can access any LLM.

  • Hacked thing gives dangerous responses.

It is surely a great calamity for a human being to have no obsessions. - Robert Bly

Working...