Big IT for CERN's particle smashing experiment
CERN's Large Hadron Collider isn't just the largest experiment in history, it's also a massive IT project linking hundreds of thousands of computers - and could change how we share and analyse data.
Nestled in the Alps near lovely Geneva is a feat of engineering and technological ingenuity on par with few others created in the name of science, and it's about to awaken.
When the Large Hadron Collider (LHC) goes live at CERN along the Swiss-French border in just a few days, it will produce an awesome amount of data which will be fed to thousands of scientists around the world assuming the thing doesn't create a massive black hole and kill us all in a horribly cool sci-fi sort of way.
CERN's big baby isn't just an overwhelming science experiment; it's a grand IT project, too. And the last time CERN's band of boffins had tech trouble, some guy named Tim invented the web.
This time, CERN's tech team is focusing on distributed computing, also known as grid computing, leaving some to wonder if the LHC could lead to innovation on the scale of the web project pioneered by Sir Tim Berners-Lee. Given the scale of this project, it's hard not to think something new and shiny will come out of it.
Indeed, trying to describe the project both the IT and physics in numbers is mind-boggling. They're all too huge to get your head around: 100,000 CPUs crunching 15 petabytes of data at hundreds of sites around the world what does that even mean?
For the past few years, engineers at CERN the French acronym for the European Organisation for Nuclear Research have been putting together a 16.77 mile tunnel, lined with sensors and 1,600 superconducting magnets across eight sectors, which are held at an operating temperature some 271 degrees below zero. The accelerator, which is accurate to a nanosecond, will be used to slam particles into each other to try to recreate the conditions at the beginning of time as in, the Big Bang.
While cynics have suggested this will create a black hole that will destroy the universe, CERN's (hopefully-not-evil) geniuses have said that's just not very likely.
Even without the peril of death, it's pretty exciting. CERN's geeks are looking for the so-called god particle, dubbed Higgs Boson, which will offer clues about dark matter and other missing links of physics.
This week, despite 10 September being pushed as the official start-up, no collisions will take place. That date simply marks the first time a beam will be sent all the way around the track; the smashing is set to start up on 21 October. As the wee particles are slammed into each other, sensors will take all that data, filter it for useful bits, and then start spitting it out to scientists to work with.
"It's not the flip of a switch, but a long-planned process," said David Britton, project manager of GridPP, the UK side of the distributed computing system which will allow all this data to move around and be analysed. The collider itself has been cooling for some six weeks, and the data testing has preceded this week, too. "We have been turning on the grid all this year," he said, adding that it's been tested with fake data to ensure success.
"It's an ongoing process, and next week, we will start to see a trickle of real data," he said. While that data will be more interesting than the fake stuff, it doesn't start to become really important until the beams collide, he said.
An avalanche of data
And when the real data does start coming from the Alpine lab, it will be an avalanche. The data output about 15 petabytes a year will require some 100,000 CPUs across 50 countries with some tens of thousands of those in the UK, which will also hold 700TB of disk storage across 17 sites.
The data sharing for the project is tiered and spread out over a grid managed by the LHC Computing Grid (LCG) project. The LHC's detectors will stream data straight to tier zero', which is CERN's own computing centre. The farm' features processing power equal to 30,000 computers, including storage to five million gigabytes of spinning disk space and 16 million on tape.
Raw and processed data will be sent from there over 10GB/sec lines to tier one' academic bodies, which will act as back-up to CERN and also help disseminate the data to 150 tier two' organisations and tier three' computers which could be a basic PC sitting on a scientist's desk.
In This Article
Preparing for long-term remote working after COVID-19
Learn how to safely and securely enable your remote workforceDownload now
Cloud vs on-premise storage: What’s right for you?
Key considerations driving document storage decisions for businessesDownload now
Staying ahead of the game in the world of data
Create successful marketing campaigns by understanding your customers betterDownload now
Solutions that facilitate work at full speedDownload now