Hardware

Big, Beautiful Boxes From Computer History 238

Slatterz writes "We might sometimes complain about the limitations of today's technology, but there's nothing like seeing photos of a 27Kg hard drive with a capacity of 5MB to put things into perspective. PC Authority has toured the Computer History Museum in California, and has posted these fascinating photos, including monster 27Kg and 60Kg drives, and a SAGE air-defense system. Each SAGE housed an A/N FSQ-7 computer, which had around 60,000 vacuum tubes. IBM constructed the hardware, and each computer occupied a huge amount of space. From its completion in 1954 it analyzed radar data in real-time, to provide a complete picture of US Airspace during the cold war. Other interesting photos and trivia include some giant early IBM disc platters, and pics of a curvaceous Cray-1 supercomputer, built in 1972. It was the fastest machine in the world until 1977 and an icon for decades. It cost a mere $6 million, and could perform at 160MFLOPS — which your phone can now comfortably manage."
Math

Pi Calculated To Record 2.5 Trillion Digits 432

Joshua writes "Researchers from Japan have calculated Pi to over 2.5 trillion decimals using the T2K Open Supercomputer (which is currently ranked 47th in the world according to a June, 2009 report from Top500.org). This new number more than doubles the previous record of about 1.2 trillion decimals set in 2002 by another Japanese research team. Unfortunately, there still seems to be no pattern."
Supercomputing

US Supercomputer Lead Sparks Russian Govt's Competitive Drive 74

CWmike writes "Russia's launch of Sputnik in 1957 triggered a crisis of confidence in the US that helped drive the creation of a space program. Now, Russia is comparing the US's achievements in supercomputing with theirs, and they don't like what they see. In a speech on Tuesday, Russia's President, Dmitry Medvedev, criticized his country's IT industry almost to the point of sarcasm for failing to develop supercomputing technology, and urged a dramatic change in Russia's use of high-performance computing. Medvedev, at the opening address of a Security Council Meeting on Supercomputers in Moscow, told attendees that 476 out of the 500 supercomputers on the Top500 list were manufactured in the United States. 'Therefore, in general, our situation is very difficult,' he said."
Security

Sandia Studies Botnets In 1M OS Digital Petri Dish 161

Ponca City, We love you writes "The NY Times has the story of researchers at Sandia National Laboratories creating what is in effect a vast digital petri dish able to hold one million operating systems at once in an effort to study the behavior of botnets. Sandia scientist Ron Minnich, the inventor of LinuxBIOS, and his colleague Don Rudish have converted a Dell supercomputer to simulate a mini-Internet of one million computers. The researchers say they hope to be able to infect their digital petri dish with a botnet and then gather data on how the system behaves. 'When a forest is on fire you can fly over it, but with a cyber-attack you have no clear idea of what it looks like,' says Minnich. 'It's an extremely difficult task to get a global picture.' The Dell Thunderbird supercomputer, named MegaTux, has 4,480 Intel microprocessors running Linux virtual machines with Wine, making it possible to run 1 million copies of a Windows environment without paying licensing fees to Microsoft. MegaTux is an example of a new kind of computational science, in which computers are used to simulate scientific instruments that were once used in physical world laboratories. In the past, the researchers said, no one has tried to program a computer to simulate more than tens of thousands of operating systems."
Supercomputing

BOINC Exceeds 2 Petaflop/s Barrier 114

Myrrh writes "Though an official announcement has not yet been made, it would appear that the BOINC project as a whole has exceeded two petaflop/s performance. The top page features this legend: '24-hour average: 2,793.53 TeraFLOPS.' According to last month's Top500 list of supercomputers, BOINC's performance is now beating that of the fastest supercomputer, RoadRunner, by more than a factor of two (with the caveat that BOINC has not been benchmarked on Linpack)."
Supercomputing

DARPA Wants a 19" Super-Efficient Supercomputer 200

coondoggie writes "If you can squish all the processing power of, say, an IBM Roadrunner supercomputer inside a 19-inch box and make it run on about 60 kilowatts of electricity, the government wants to talk to you. The extreme scientists at the Defense Advanced Research Projects Agency this week issued a call for research that might develop a super-small, super-efficient super beast of a computer. Specifically, DARPA's desires for Ubiquitous High Performance Computing (UHPC) will require a new system-wide technology approach including hardware and software co-design to minimize energy dissipation per operation and maximize energy efficiency, with a 50GFLOPS per watt goal."
Space

Aussie Scientists Build a Cluster To Map the Sky 58

Tri writes "Scientists at the Siding Spring Observatory have built a new system to map and record over 1 billion objects in the southern hemisphere sky. They collect 700 GB of data every night, which they then crunch down using some perl scripts and make available to other scientists through a web interface backed on Postgresql. 'Unsurprisingly, the Southern Sky Survey will result in a large volume of raw data — about 470 terabytes ... when complete. ... the bulk of the analysis of the SkyMapper data will be done on a brand new, next generation Sun supercomputer kitted out with 12,000 cores. Due to be fully online by December, the supercomputer will offer a tenfold increase in performance over the facility's current set up of two SGI machines, each with just under 3500 cores in total.'"
Hardware Hacking

Hydraulic Analog Computer From 1949 184

mbone writes "In the New York Times, there is an interesting story about a hydraulic analog computer from 1949 used to model the feedback loops in the economy. According to the article, 'copies of the 'Moniac,' as it became known in the United States, were built and sold to Harvard, Cambridge, Oxford, Ford Motor Company and the Central Bank of Guatemala, among others.' There is a cool video of the computer in operation at Cambridge University. I remember that the Instrumentation Lab at MIT still had an analog computer in its computer center in the mid-1970s. Even then, it seemed archaic, and now this form of computation is largely forgotten. With 14 machines built, it must have been one of the more successful analog computers — a supercomputer of its day. Of course, you have to wonder if it could have been used to predict our current economic difficulties."
Supercomputing

Flu Models Predict Pandemic, But Flu Chips Ready 216

An anonymous reader writes "Supercomputer software models predict that swine flu will likely go pandemic sometime next week, but flu chips capable of detecting the virus within four hours are already rolling off the assembly line. The U.S. Department of Health and Human Services (HHS), which has designated swine flu as the '2009 H1N1 flu virus,' is modeling the spread of the virus using modeling software designed by the Department of Defense back when avian flu was a perceived threat. Now those programs are being run on cluster supercomputers and predict that officials are not implementing enough social distancing--such as closing all schools--to prevent a pandemic. Companies that designed flu-detecting chips for avian flu, are quickly retrofitting them to detect swine flu, with the first flu chips being delivered to labs today." Relatedly, at least one bio-surveillance firm is claiming they detected and warned the CDC and the WHO about the swine flu problem in Mexico over two weeks before the alert was issued.
Supercomputing

Supercomputer As a Service 78

gubm writes "Nearly one and a half years after making a stunning entry into the global supercomputer list with Eka, ranked as the fourth-fastest supercomputer in the world, Computational Research Laboratories (CRL), a Tata Sons' subsidiary, has succeeded in creating a new market for supercomputers — that of offering supercomputing power on rent to enterprises in India. For now, for want of a better word, let us call it 'Supercomputer as a Service.'"
Supercomputing

Collaborative Map-Reduce In the Browser 188

igrigorik writes "The generality and simplicity of Google's Map-Reduce is what makes it such a powerful tool. However, what if instead of using proprietary protocols we could crowd-source the CPU power of millions of users online every day? Javascript is the most widely deployed language — every browser can run it — and we could use it to push the job to the client. Then, all we would need is a browser and an HTTP server to power our self-assembling supercomputer (proof of concept + code). Imagine if all it took to join a compute job was to open a URL."
Hardware Hacking

DIY 1980s "Non-Von" Supercomputer 135

Brietech writes "Ever wanted to own your own supercomputer? This guy recreated a 31-processor SIMD supercomputer from the early 1980s called the 'Non-Von 1' in an FPGA. It uses a 'Non-Von Neumann' architecture, and was intended for extremely fast database searches and artificial intelligence applications. Full-scale models were intended to have more than a million processors. It's a cool project for those interested in 'alternative' computer architectures, and yes, full source code (Verilog) is available, along with a python library to program it with." Hope the WIPO patent has expired.
Supercomputing

IBM Building 20 Petaflop Computer For the US Gov't 248

eldavojohn writes "When it's built, 'Sequoia' will outshine every super computer on the top 500 list today. The specs on this 96 rack beast are a bit hard to comprehend as it consists of 1.6 million processors and some 1.6TB of memory. That's 1.6 million processors — not cores. Its purpose? Primarily to keep track of nuclear waste & simulate explosions of nuclear munitions, but also for research into astronomy, energy, the human genome, and climate change. Hopefully the government uses this magnificent tool wisely when it gets it in 2012."
AMD

AMD Plans 1,000-GPU Supercomputer For Games, Cloud 148

arcticstoat writes "AMD is planning to use over 1,000 Radeon HD 4870 GPUs to create a supercomputer capable of processing one petaflop, which the company says will make 'cloud' computing a reality. When it's built later this year, the Fusion Render Cloud will be available as an online powerhorse for a variety of people, from gamers to 3D animators. The company claims that it could 'deliver video games, PC applications and other graphically-intensive applications through the Internet "cloud" to virtually any type of mobile device with a web browser.' The idea is that the Fusion Render Cloud will do all the hard work, so all you need is a machine capable of playing back the results, saving battery life and the need for ever greater processing power. AMD also says that the supercomputer will 'enable remote real-time rendering of film and visual effects graphics on an unprecedented scale.' Meanwhile, game developers would be able to use the supercomputer to quickly develop games, and also 'serve up virtual world games with unlimited photo-realistic detail.' The supercomputer will be powered by OTOY software, which allows you to render 3D visuals in your browser via streaming, compressed online data."
Supercomputing

How To Build a Homebrew PS3 Cluster Supercomputer 211

eldavojohn writes "UMass Dartmouth Physics Professor Gaurav Khanna and UMass Dartmouth Principal Investigator Chris Poulin have created a step-by-step guide designed to show you how to build your own supercomputer for about $4,000. They are also hoping that by publishing this guide they will bring about a new kind of software development targeting this architecture & grid (I know a few failed NLP projects of my own that could use some new hardware). If this catches on for research institutions it may increase Sony's sales, but they might not be seeing the corresponding sale of games spike (where they make the most profit)."
Supercomputing

Inside Tsubame, Japan's GPU-Based Supercomputer 75

Startled Hippo writes "Japan's Tsubame supercomputer was ranked 29th-fastest in the world in the latest Top 500 ranking with a speed of 77.48T Flops (floating point operations per second) on the industry-standard Linpack benchmark. Why is it so special? It uses NVIDIA GPUs. Tsubame includes hundreds of graphics processors of the same type used in consumer PCs, working alongside CPUs in a mixed environment that some say is a model for future supercomputers serving disciplines like material chemistry." Unlike the GPU-based Tesla, Tsubame definitely won't be mistaken for a personal computer.
Supercomputing

NVIDIA's $10K Tesla GPU-Based Personal Supercomputer 236

gupg writes "NVIDIA announced a new category of supercomputers — the Tesla Personal Supercomputer — a 4 TeraFLOPS desktop for under $10,000. This desktop machine has 4 of the Tesla C1060 computing processors. These GPUs have no graphics out and are used only for computing. Each Tesla GPU has 240 cores and delivers about 1 TeraFLOPS single precision and about 80 GigaFLOPS double-precision floating point performance. The CPU + GPU is programmed using C with added keywords using a parallel programming model called CUDA. The CUDA C compiler/development toolchain is free to download. There are tons of applications ported to CUDA including Mathematica, LabView, ANSYS Mechanical, and tons of scientific codes from molecular dynamics, quantum chemistry, and electromagnetics; they're listed on CUDA Zone."
IBM

DARPA's IBM-Led Neural Network Project Seeks To Imitate Brain 170

An anonymous reader writes "According to an article in the BBC, IBM will lead an ambitious DARPA-funded project in 'cognitive computing.' According to Dharmendra Modha, the lead scientist on the project, '[t]he key idea of cognitive computing is to engineer mind-like intelligent machines by reverse engineering the structure, dynamics, function and behaviour of the brain.' The article continues, 'IBM will join five US universities in an ambitious effort to integrate what is known from real biological systems with the results of supercomputer simulations of neurons. The team will then aim to produce for the first time an electronic system that behaves as the simulations do. The longer-term goal is to create a system with the level of complexity of a cat's brain.'"
Supercomputing

E=mc^2 Verified In Quantum Chromodynamic Calculation 268

chirishnique and other readers sent in a story in AFP about a heroic supercomputer computation that has verified Einstein's most famous equation at the level of subatomic particles for the first time. "A brainpower consortium led by Laurent Lellouch of France's Centre for Theoretical Physics, using some of the world's mightiest supercomputers, have set down the calculations for estimating the mass of protons and neutrons, the particles at the nucleus of atoms. ... [T]he mass of gluons is zero and the mass of quarks is only five per cent. Where, therefore, is the missing 95 per cent? The answer, according to the study published in the US journal Science on Thursday, comes from the energy from the movements and interactions of quarks and gluons. ... [E]nergy and mass are equivalent, as Einstein proposed in his Special Theory of Relativity in 1905." Update: 11/21 15:50 GMT by KD : New Scientist has a slightly more technical look at the accomplishment.
Supercomputing

Windows Breaks Into Supercomputer Top 10 294

yanx0016 writes "Wow, that's some news this week at SuperComputing 08. Apparently Microsoft Windows HPC Server 2008, with a Chinese hardware OEM (Dawning), made #10 on the Top500 list, edging out #11 by only 600 Gflops. Folks were shocked to see Microsoft getting so serious around HPC; I think we are only beginning to see a glimpse of Microsoft in the HPC field."

Slashdot Top Deals