Follow Slashdot blog updates by subscribing to our blog RSS feed

 



Forgot your password?
typodupeerror
×
Encryption Data Storage Security

NIST Investigating Mass Flash Drive Vulnerability 71

Lucas123 writes with a followup to news we discussed earlier this week that the encryption on NIST-certified flash drives was cracked. "A number of leading manufacturers of encrypted flash drives have warned their customers of a security flaw uncovered by a German company. The devices in question use the AES 256-bit encryption algorithm and have been certified using the FIPS 140-2, but the flaw appears to circumvent the certification process by uncovering the password authentication code on host systems. The National Institute of Standards and Technology said it's investigating whether it needs to modify its standards to include password authentication software on host systems. Security specialist Bruce Schneier was blunt in his characterization of the flaw: 'It's a stupid crypto mistake and they screwed up and they should be rightfully embarrassed for making it.'"
This discussion has been archived. No new comments can be posted.

NIST Investigating Mass Flash Drive Vulnerability

Comments Filter:
  • by MichaelSmith ( 789609 ) on Saturday January 09, 2010 @06:13AM (#30706124) Homepage Journal

    Use PGP. Create a really long key, like 4096 bits.

    • by ls671 ( 1122017 ) *

      Bah, instead, I am still using an Enigma machine that my grandfather brought me back. He stole it from the ennemy while in combat.

      http://en.wikipedia.org/wiki/Enigma_machine [wikipedia.org]

      • Re: (Score:3, Interesting)

        Bah, instead, I am still using an Enigma machine that my grandfather brought me back. He stole it from the ennemy while in combat.

        http://en.wikipedia.org/wiki/Enigma_machine [wikipedia.org]

        You should count yourself lucky that Alan Turing died all those years ago, otherwise your data could be compromised.

        • by ls671 ( 1122017 ) * on Saturday January 09, 2010 @06:37AM (#30706178) Homepage

          > otherwise your data could be compromised.

          With this ? :

          http://en.wikipedia.org/wiki/File:Bombe-rebuild.jpg [wikipedia.org]

          Too complex to maintain in good working order. ;-))

          • while they don't showcase a running Bombe replica, I saw at Bletchley Park a fully running rebuild of the Colossus machine [wikipedia.org], which was used at WW2 to break the more complex encryption utilized by the Lorenz SZ40 machine.
            Since it has programmable boolean functions I bet it can be used to break Enigma code as well.
            • by ls671 ( 1122017 ) *

              Seriously ?

              I think simply implementing the breaking algorithm in your favorite language on your PC would be more convenient and also give results much faster ;-))

              But you would need to know that I am using an Enigma machine in the first place, oh wait...

              • by MK_CSGuy ( 953563 ) on Saturday January 09, 2010 @07:24AM (#30706352)

                I think simply implementing the breaking algorithm in your favorite language on your PC would be more convenient and also give results much faster ;-))

                You are right [bbc.co.uk] of course [wikipedia.org]:

                Nevertheless the victor's 1.4 GHz laptop, running his own code, took less than a minute to find the settings for all 12 wheels... 240 times faster than Colossus. If you scale the CPU frequency by that factor, you get an equivalent clock of 5.8 MHz for Colossus. That is a remarkable speed for a computer built in 1944.

                You still get massive geek cred. either way :)

                • Re: (Score:3, Informative)

                  by evilviper ( 135110 )

                  you get an equivalent clock of 5.8 MHz for Colossus. That is a remarkable speed for a computer built in 1944.

                  It would be incredible if true, but it's not. Special-purpose hardware can perform certain types of computations far faster than general-purpose processors. Hardware that could decode 1080i MPEG-2 (HDTV) could easily (though not inexpensively) have been made a decade before Intel/AMD CPUs were up to the task. That doesn't mean we had 2GHz+ CPUs in the early 1990s, it just means we had special-purp

        • Turing isn't dead as long as a single geek is alive.

      • Re: (Score:3, Funny)

        by PopeRatzo ( 965947 ) *

        I am still using an Enigma machine that my grandfather brought me back. He stole it from the ennemy while in combat.

        You're the one who's got my grandad's Enigma machine!

        Give it back. You can send it to me here in Argentina.
         

    • by snemarch ( 1086057 ) on Saturday January 09, 2010 @06:47AM (#30706202)

      Not really applicable to a hardware device.

      Also, keep in mind that RSA by itself is much too slow to encrypt large amounts of data; thus, PGP and other solutions only use RSA to encrypt a symmetric cipher, which is then used for the bulk encryption.

      Standard AES-256 is actually just fine, problem with these devices is that the manufacturers screwed up the implementation *majorly* (as I understand it, use the same key for every device and depend on a usermode app to say GOOD_GUY/BAD_GUY to the hardware) - but that's covered elsewhere.

      • by TubeSteak ( 669689 ) on Saturday January 09, 2010 @10:51AM (#30707232) Journal

        Standard AES-256 is actually just fine, problem with these devices is that the manufacturers screwed up the implementation *majorly* (as I understand it, use the same key for every device and depend on a usermode app to say GOOD_GUY/BAD_GUY to the hardware) - but that's covered elsewhere.

        The fact that so many major companies have the same exact flaw in their product suggests (to me) that there is only one manufacturer and multiple vendors who just rebadged the item.

        I think it's less likely that multiple companies independantly managed to screw up their products in exactly the same way.

        • Sounds plausible; the thing that irks me is that if these USB devices really do have AES-256 capabilites and the contents of the flashram is encrypted... why the hell have they made such a terrible flaw implementing the rest of the system? :-s

        • The bug is not in the key itself. It's in software that runs on the host system (which is a bug in itself) They may all have licensed the same software.

  • by Anonymous Coward on Saturday January 09, 2010 @06:16AM (#30706128)

    Encryption algorithm's aren't the weak link, its the implementation. But most people just look at how big the key is not who implemented it.

    • Re: (Score:3, Funny)

      by Anonymous Coward

      Put it in a way /. understands, please.

      "It's like having a really huge penis but never leaving your mother's basement."

    • by Kjella ( 173770 ) on Saturday January 09, 2010 @06:30AM (#30706166) Homepage

      Encryption algorithm's aren't the weak link, its the implementation.

      What's more usually the case is that the implementation of the algorithm is just fine, but you fail at using it in the right way. Usually because then you've handed it off from the cryptography experts and to the general team that's building the rest of the system. Kinda like a door that has a great lock but is easy to take off its hinges, won't do you much good.

      • by Anonymous Coward on Saturday January 09, 2010 @08:18AM (#30706564)

        My understanding of these devices puts the analogy at:

        A super solid, near uncrackable/breakable safe, but all models use the same 12345 passcode, and the owners cannot change this.

        To make matters worse, the door of the safe has been mounted on a normal house door that only has a sign saying 'do not open if this is not yours'.

        This way, the safe owners dont need to rember such a complex code as '12345', and you get all the security of the full safe. Unless a intruder happens to have a brain, then they will just open the house door that holds the safe door thus negating the entire system.

        If anyone could design a worse system, they are truely rube goldberg masters.

        • Re: (Score:1, Insightful)

          by Anonymous Coward

          A slightly better analogy is the following:

          A super solid uncrackable safe which has a 12345 passcode for all models. This safe however, has a removeable front panel door that covers the passcode entry input. This front panel has it's own combination input that can be user set, and when the user inputs the correct user set passcode, the front panel automatically inputs the 12345 into the underlying safe door.

          So all is well and good until someone comes along and just removes the removeable front panel.

          This

      • Or an expensive bike lock you can open with the end of a BIC pen [wired.com]...
      • Re: (Score:1, Interesting)

        by Anonymous Coward

        What's more usually the case is that the implementation of the algorithm is just fine, but you fail at using it in the right way. Usually because then you've handed it off from the cryptography experts and to the general team that's building the rest of the system. Kinda like a door that has a great lock but is easy to take off its hinges, won't do you much good.

        That's a problem with software in general, not just encryption.

        Often once the coders have solved the "interesting" problems, they get bored with the mundane implementation details. If your software does everything it is supposed to, but the user experience sucks, then you have still failed. Coding isn't about creating great code. The user doesn't care about your code. They want to solve their problem. Programs are only a tool to achieve that.

      • Comment removed based on user account deletion
    • by Joce640k ( 829181 ) on Saturday January 09, 2010 @07:39AM (#30706426) Homepage

      The weak link is in the apostrophe.

    • Re: (Score:2, Funny)

      by jd2112 ( 1535857 )
      Just use Quantum Encryption, They'll never crack that.

      Oh, nevermind. http://it.slashdot.org/story/09/12/30/2118250/Quantum-Encryption-Implementation-Broken?art_pos=1 [slashdot.org]
  • by Snotboble_ ( 13797 ) <aje@snotboble.STRAWnet minus berry> on Saturday January 09, 2010 @07:15AM (#30706296) Homepage Journal
    This is pretty major as so many vendors are affected by it. However, until there's an update or complete recall & replacement, I'd recommend using Truecrypt [truecrypt.org]. Certified by NIST (see HERE [law2point0.com]. Cross platform. Free (as in spoken beer ;o). Of course, one can only hope that its implementation is better than the devices currently uncovered :P
    • by AHuxley ( 892839 )
      vendors are affected because they all outsourced to the same sub set of firms in 2nd and 3rd world for cheap code.
      Buzz word for the box, MS skills on the back end.
      For what they save in hiring a few smart American experts they will now lose many times over in PR spin to reinvent their failed brands.
      Never forget any junk crypto code news over the next decade or two will bring back quotes about your pathetic company.
      • For what they save in hiring a few smart American experts they will now lose many times over in PR spin to reinvent their failed brands.

        The problem is.. they won't. This kind of thing should be a good for competitors who did things correctly, but in the end, no lessons will be learned, and any competitors that actually do have good products will not gain significant market share.

        They might get a fine, and a junior executive might have to go work somewhere else (for more money, though...). The will see very little downside. Nothing compared to the unlimited liability for complications due to loosing of users' data that they should have to

    • Where's this guy [slashdot.org] when we need him?

    • My one issue with Truecrypt is that it doesn't support multiply keys. That's almost essential in a real world environment. Why the hell don;t they add that?

    • This is pretty major as so many vendors are affected by it. However, until there's an update or complete recall & replacement, I'd recommend using Truecrypt [truecrypt.org]. Certified by NIST (see HERE [law2point0.com]. Cross platform. Free (as in spoken beer ;o). Of course, one can only hope that its implementation is better than the devices currently uncovered :P

      I remember a recent Slashdot article about a vulnerability in TrueCrypt. Don't remember much more than that. Does anyone else know what I'm talking about?

    • TC isn't really portable. I mean, yes, the program itself can be installed portably, but you still need admin permission to install the driver to make it actually work.
  • "Kingston Technology Inc. ---- ....is warning customers about a potential security threat posed by a flaw in the hardware-based AES 256-bit encryption on their USB flash drives."
  • by advocate_one ( 662832 ) on Saturday January 09, 2010 @08:09AM (#30706534)

    IronKey was among a number of companies to issue statements reassuring customers that their devices were safe from the same attacks. Jevans said that's because the password and authentication process is contained on the USB drive itself and has nothing to do with the host system.
    "We don't trust the computer at all," he said. "The computer could have malware on it or have hackers accessing it. In our security design, we said we have to assume the computer is completely untrustworthy. That's where we started our threat modeling."

    • Re: (Score:3, Funny)

      by AHuxley ( 892839 )
      Why not just say "Microsoft"?
    • You need buttons on the device. Without that, your password could be swiped by PC malware.

      A no-frills minimal device comes with 10 buttons. The password is a 10-digit number printed on a card hidden in the packaging. To avoid having the password revealed by button wear, none of the digits repeats. You put the device in, press buttons, and then it shows up to the OS.

      A better device has a config setup. Press an extra recessed button, and the device appears as a USB netword device with a DHCP server and all. G

      • And this is "not trusting the computer" how?
        • by r00t ( 33219 )

          Of course, a certain level of trust is implied by the use of the device. At minimum you're making one filesystem readable; it could be copied.

          That doesn't mean you must make the filesystem writable unless actually needed.

          That doesn't mean you must give up a password which you might (wrongly...) be using elsewhere.

          That doesn't mean that you must make all filesystems available. Virtual devices allow you to make a per-computer choice.

      • by ascari ( 1400977 )
        And after all that, once your encrypted is duly decrypted by you it is displayed on the screen in cleartext. The hypothetical malware running on your untrustworthy PC simply swipes is out of the frame buffer and sends it to your enemy. (It might even encrypt it before sending it, sweet irony. ;-)

        Hypotehtical perhaps, but quite doable. Bottom line is if you ever look at your data in cleartext at any time on any device I've already pwned you. Sorry to be the bearer of bad news.

        • by r00t ( 33219 )

          You got the current data, and maybe (if a read-write password is used) you got to mangle it.

          You don't get future access. Stealing the device next year won't do you any good.

          You don't get some password that I might be improperly (humans are lazy) using for something more important.

          If virtual devices are in use, you only got access to the one I activated.

  • Back in the nineties there was a teapot tempest around an issue with Intuit's Quicken personal finance software. Users could "encrypt" their Quicken data files, but the password they created to access it was easily recoverable by simply looking at the plain text raw data stored on hard drive sectors. I had Norton Utilities which allowed me to examine drive sectors, and sure enough, there was my Quicken password (altho the actual data was encrypted).

    That's what this reminds me of. Is this a valid -if pe
  • by Gnavpot ( 708731 ) on Saturday January 09, 2010 @09:48AM (#30706898)

    I really do not understand this part:
    "The National Institute of Standards and Technology said it's investigating whether it needs to modify [CC] its standards to include password authentication software on host systems."

    This has already been proven to be very unsafe hardware. The fact that you can access the data without using the original software and without knowing the user's password should leave no doubt. As long as you have some software which says "Open Sesame" in the same way as the original software, you will get access.

    So why was this not discovered during the NIST certification process? And why do NIST state that they may need to approve the software too to protect against this?

    It seems to me that NIST blames the software so they will not have to take blame for their faulty certification of the hardware.

    • Re: (Score:3, Insightful)

      by Anonymous Coward

      NIST doesn't actually certify things as "secure", it makes very specific certification of particular tests that may or may not represent what you think of as "secure". It's only the marketing that makes you think that if the words "NIST", "Certified" and "Security" appear in the same sentence that someone has done a proper end-to-end review.

      It's like the way that the auditors' certificate in financial reports makes people think that the auditor is guaranteeing that there cannot be any fraud in the company a

    • by Jaime2 ( 824950 ) on Saturday January 09, 2010 @12:02PM (#30707618)

      So why was this not discovered during the NIST certification process?

      Because the certification the hardware received only verifies that the algorithm strength is sufficient and that the device is hardened against physical tampering.

      It seems to me that NIST blames the software so they will not have to take blame for their faulty certification of the hardware.

      Nope, it seems that the NIST has recognized that the certification, as currently written, isn't sufficient and is looking into making it more robust. Had they audited the software, they would have discovered that the software-to-hardware interface is poorly designed and not granted the certification.

  • Best part of TFA (Score:3, Insightful)

    by dr2chase ( 653338 ) on Saturday January 09, 2010 @10:18AM (#30707038) Homepage
    is how everything is carefully run through the make-nice factory. The memory chip makers ucked fup. NIST ucked fup. Yet, NIST cannot say, "whoa, we blew it, we have to fix that standard immediately" (else it will be completely worthless). No, they're organizing a committee to appoint a task force to propose revisions to the standard, pending who-knows-what. And even the guys who got it right, try to make nice with a handy excuse for how this came about -- "difficult to administer with all those different passwords". You set two passwords for each device, duh, and let either access the bits. Vendors provide them with a customer-specified admin password, or vendor supplies a chip initialization utility where customer may bake in an admin password.
  • by Old Man Kensey ( 5209 ) on Saturday January 09, 2010 @02:11PM (#30708456) Homepage
    Any flash drive whose "security" involves a required app running on the host system will not be suitable for cross-platform use even if the app is well-written. The only right way to do it is to encrypt the data written to the drive, using well-known secure encryption algorithms run on the host. And for that purpose a cheap, dumb drive works just as well as a super-expensive "secure, tamper-proof" drive.
    • by ascari ( 1400977 )
      Unfortunately, not even that is enough. The current issue of 2600 magazine has a "how-to" article on breaking full disk security. Won't spoil the story by giving it away, but it does raise some question about depending on encryption for security in general.
  • Yep. (Score:3, Funny)

    by ScrewMaster ( 602015 ) * on Saturday January 09, 2010 @04:28PM (#30709438)

    Security specialist Bruce Schneier was blunt in his characterization of the flaw: 'It's a stupid crypto mistake and they screwed up and they should be rightfully embarrassed for making it.'"

    That's our Bruce.

  • If all the drives have identical challenge response interchanges, why wasn't this noticed the first time someone using his own machine/software/password successfully accessed the wrong drive?

  • I would never even bother to encrypt data on a flash drive.

    1) If your data is so important that other people shouldn't read it, why are you putting it on a flash drive which can be stolen?

    2) If, for some reason, you are storing your data on a flash drive, why are you leaving it where others can steal it? You can have my flash drive when you pry it from my cold, dead hands!

    3) If you do put your data on a flash drive and somebody kills you to get it, it stands to reason that they will be able to work at t

Get hold of portable property. -- Charles Dickens, "Great Expectations"

Working...