OpenAI To Limit New Model Release On Cybersecurity Fears (axios.com) 37
OpenAI is reportedly preparing a new cybersecurity product for a small group of partners, out of concern that a broader rollout could wreak havoc if it were released more widely. If that move sounds familiar, it's because Anthropic took a similar limited-release approach with its Mythos model and Project Glasswing initiative. Axios reports: OpenAI introduced its "Trusted Access for Cyber" pilot program in February after rolling out GPT-5.3-Codex, the company's most cyber-capable reasoning model. Organizations in the invite-only program are given access to "even more cyber capable or permissive models to accelerate legitimate defensive work," according to a blog post. At the time, OpenAI committed $10 million in API credits to participants. [...]
Restricting the rollout of a new frontier model makes "more sense" if companies are concerned about models' ability to write new exploits -- rather than about their ability to find bugs in the first place, Stanislav Fort, CEO of security firm Aisle, told Axios. Staggering the release of new AI models looks a lot like how cybersecurity vendors currently handle the disclosure of security flaws in software, Lee added. "It's the same debate we've had for decades around responsible vulnerability disclosure," Lee said.
Restricting the rollout of a new frontier model makes "more sense" if companies are concerned about models' ability to write new exploits -- rather than about their ability to find bugs in the first place, Stanislav Fort, CEO of security firm Aisle, told Axios. Staggering the release of new AI models looks a lot like how cybersecurity vendors currently handle the disclosure of security flaws in software, Lee added. "It's the same debate we've had for decades around responsible vulnerability disclosure," Lee said.
Hey what a coincidence... (Score:5, Insightful)
Anthropic announces that they have a super awesome AI product that's just too awesome for anyone for anyone to see.
And then immediately OpenAI has the exact same thing.
FOMO on "my technology is too scary to exist" is a fun twist.
I know, it's not the first time, someone even linked an article where OpenAI said the same sort of thing about GPT-2 back in 2019...
Re: (Score:2)
OpenAI: "This is just for a few select companies to have..."
Wait, doesn't that go against OpenAI's mission statement of OPEN ACCESS???
Re: Hey what a coincidence... (Score:3, Informative)
Re:Hey what a coincidence... (Score:4, Funny)
Wait, doesn't that go against OpenAI's mission statement of OPEN ACCESS???
I'm pretty sure the "open" part now refers to opening your wallet.
Re: (Score:2)
The difference is we can see what OpenAI puts out generally and see what Anthropic puts out and see pretty clearly which one is miles ahead of the other. I believe Anthropic, not so sure about OpenAI.
Re: (Score:3)
While Anthropic is generally more credible, they have indulged in performative bullshit for the sake of the hype train.
Frankly, if they didn't, they would have been screwed over no matter how well they actually made a product.
Not crazy about the "do things to open source projects, but obfuscate the fact that it's LLM originated" Anthropic thing either way.
Us too (Score:1)
So Anthropic demos Mythos and OpenAI has to say put out a press release.
Altman's house of cards is collapsing...
Re:Us too (Score:4, Insightful)
I do suspect that OpenAI will be the 'Netscape' of this bubble pop. Early mover that in many ways sparked something significant that got left behind by others that did it better.
I am so eager for a bubble pop to recalibrate expectations to properly leverage LLM as appropriate instead of the current madness. It will be an adjustment, but without the craze it won't be nearly so obnoxious.
Re: Us too (Score:2)
Who do you have in mind who did better than Netscape back then? The one thing IE did better was insert itself everywhere.
Re: (Score:2)
Re: (Score:2)
The problem being I haven't seen a good term that refers to the extended LLM scenario that is specific enough to exclude other things like machine vision.
Everyone is referring to the extended LLM scenario and despite things feeding improvements, it still cannot do everything that people promise/believe it can do. I have been inundated by project proposals that largely center around "screw everyone but my job, AI can replace everyone but me", and they are just full of bad ideas.
Basically, the good old "I ha
Re: (Score:3)
Re: (Score:2)
Re: (Score:2)
So Anthropic demos Mythos and OpenAI has to say put out a press release.
Altman's house of cards is collapsing...
Dude, the next model is gonna be so scary that they won't even let THEMSELVES use it.
Prolossus: The Beforbin Preject (Score:2)
Firesign Theatre Got It Right (Score:3)
This isn't a mirage (Score:3)
Re: (Score:2)
I mean in theory you can have the AI identify and fix the exploits. Yeah it's an arms race but as some point the defense will probably win.
Re: This isn't a mirage (Score:2)
Re: (Score:2)
Other than fixing your bugs.
Or don't computer (Score:2)
Just get everything off the internet. Shouldn't we be up to Internet 2027 already?
Re: This isn't a mirage (Score:2)
Re: (Score:2)
Fixing your bugs scales pretty well. What doesn't scale well is hoping nobody notices. Works great for the hockey pool you wrote in Python one weekend for your buddies. Not so well for stuff lots of people are going to use.
Current situation: you look for bugs, then you release your software and, if it's "scaled" a million other people look for bugs.
New situation: you run an AI bug scanner and fix your bugs, then you release your software a million other people run an AI bug scanner.
The only situation in whi
Re: (Score:2)
The same argument could be made around automated fuzzing. A new class of security misbehavior may be identified automatically, and it turns out you can use such tools to identify things to fix as well.
Of course, it could be a problem if it has a high false positive rate, where the attacker can hit false positives and barely be impacted but the false positives drive an impossible churn to keep up with on the defense side... Which frankly could be a thing based on my experience with LLM code review that can
Re: (Score:2)
IF models get that good at coding, that will probably mean the end of purpose built software anyway. At that point, everybody will be able to roll their own, tailored to the user's exact needs.
Re: This isn't a mirage (Score:2)
For user-facing stuff we're probably a few years away still.
Supposedly limited-release (Score:2)
Nation states will have sleeper agents who will grab a copy, send it home and go back to sleep waiting for the next big thing. How really dangerous it is - time will tell.
any guardrails on this model? (Score:2)
The real question is if this model doesn't have guardrails to "do no harm" or attempt to crack password or was tuned to specifically allow what they block the general public from.
Another bone headed move that we are starting to see from our AI overlords. Build powerful models that they gate keep and possibly lend to governments first.
I bet this will play out quite nicely as foreign countries and anyone with enough money to build a similar model, even a watered down model to fine the same security holes and
Re: (Score:1)
Walled gardens (Score:2)
The big frontier model providers are all moving towards an Enterprise-first approach to their products, gradually nerfing the 'value' lines and eliminating any subsidization paid plans may have previously (and currently) have. The free ride of investment subsidy couldn't have continued much longer and allow the companies to remain solvent.
This is just a part of that: super special tools and models will be kept amongst a small subset of enterprise companies and governments at a high premium, and they will gr
Heard this before (Score:2)
Are they playing the "AI is dangerous" card in hope of getting some cushy government regulated barriers to entry again?
Altman did it before...
Cybersecurity is a chimera .. (Score:2)
Chimera: a thing which is hoped for but is illusory or impossible to achieve.
Hit the wall hard (Score:2)
It is known as the digital apartheid (Score:1)