The Worldwide LHC Computing Grid (WLCG), formerly (until 2006) the LHC Computing Grid (LCG), is an international collaborative project that consists of a grid-based
computer network
A computer network is a collection of communicating computers and other devices, such as printers and smart phones. In order to communicate, the computers and devices must be connected by wired media like copper cables, optical fibers, or b ...
infrastructure incorporating over 170 computing centers in 42 countries, . It was designed by
CERN
The European Organization for Nuclear Research, known as CERN (; ; ), is an intergovernmental organization that operates the largest particle physics laboratory in the world. Established in 1954, it is based in Meyrin, western suburb of Gene ...
to handle the prodigious volume of data produced by
Large Hadron Collider
The Large Hadron Collider (LHC) is the world's largest and highest-energy particle accelerator. It was built by the CERN, European Organization for Nuclear Research (CERN) between 1998 and 2008, in collaboration with over 10,000 scientists, ...
(LHC) experiments.
[
]

By 2012, data from over 300 trillion (3×10
14) LHC proton-proton collisions had been analyzed,
[Hunt for Higgs boson hits key decision point](_blank)
/ref> and LHC collision data was being produced at approximately 25 petabyte
The byte is a unit of digital information that most commonly consists of eight bits. Historically, the byte was the number of bits used to encode a single character of text in a computer and for this reason it is the smallest addressable un ...
s per year. the LHC Computing Grid is the world's largest computing grid comprising over 170 computing facilities in a worldwide network across 42 countries scattered around the world that produce a massive distributed computing infrastructure with about 1,000,000 CPU cores, providing more than 10,000 physicists around the world with near real-time access to the LHC data, and the power to process it.
According to th
WLCG Website
as of 2024: "WLCG combines about 1.4 million computer cores and 1.5 exabytes of storage from over 170 sites in 42 countries ..It runs over 2 million tasks per day and ..global transfer rates exceeded 260 GB/s." Indicating substantial upgrades to WLCG over time beyond its initial release.
Background
The Large Hadron Collider at CERN was designed to test the existence of the Higgs boson
The Higgs boson, sometimes called the Higgs particle, is an elementary particle in the Standard Model of particle physics produced by the excited state, quantum excitation of the Higgs field,
one of the field (physics), fields in particl ...
, an important but elusive piece of knowledge that had been sought by particle physicists
Particle physics or high-energy physics is the study of fundamental particles and forces that constitute matter and radiation. The field also studies combinations of elementary particles up to the scale of protons and neutrons, while the stud ...
for over 40 years. A very powerful particle accelerator
A particle accelerator is a machine that uses electromagnetic fields to propel electric charge, charged particles to very high speeds and energies to contain them in well-defined particle beam, beams. Small accelerators are used for fundamental ...
was needed, because Higgs bosons might not be seen in lower energy experiments, and because vast numbers of collisions would need to be studied. Such a collider would also produce unprecedented quantities of collision data
Data ( , ) are a collection of discrete or continuous values that convey information, describing the quantity, quality, fact, statistics, other basic units of meaning, or simply sequences of symbols that may be further interpreted for ...
requiring analysis. Therefore, advanced computing facilities were needed to process the data.
Description
A design report was published in 2005.
It was announced to be ready for data on 3 October 2008.
A popular 2008 press article predicted "the internet could soon be made obsolete" by its technology.
CERN had to publish its own articles trying to clear up the confusion.
It incorporates both private fiber-optic cable links and existing high-speed portions of the public Internet
The Internet (or internet) is the Global network, global system of interconnected computer networks that uses the Internet protocol suite (TCP/IP) to communicate between networks and devices. It is a internetworking, network of networks ...
. At the end of 2010, the Grid consisted of some 200,000 processing cores and 150 petabytes of disk space, distributed across 34 countries.
The data stream from the detectors provides approximately 300 GByte/s of data, which after filtering for "interesting events", results in a data stream of about 300 MByte/s. The CERN computer center, considered "Tier 0" of the LHC Computing Grid, has a dedicated 10 Gbit/s connection to the counting room.
The project was expected to generate multiple TB of raw data and event summary data, which represents the output of calculations done by the CPU farm at the CERN data center. This data is sent out from CERN to thirteen Tier 1 academic institutions in Europe, Asia, and North America, via dedicated links with 10 Gbit/s or higher of bandwidth. This is called the LHC Optical Private Network.
More than 150 Tier 2 institutions are connected to the Tier 1 institutions by general-purpose national research and education network
A national research and education network (NREN) is a specialised internet service provider dedicated to supporting the needs of the research and education communities within a country.
It is usually distinguished by support for a high-speed backb ...
s.[final-draft-4-key](_blank)
The data produced by the LHC on all of its distributed computing grid is expected to add up to 200 PB of data each year. In total, the four main detectors at the LHC produced 13 petabytes of data in 2010.
The Tier 1 institutions receive specific subsets of the raw data, for which they serve as a backup repository for CERN. They also perform reprocessing when recalibration is necessary. The primary configuration for the computers used in the grid is based on CentOS
CentOS (, from Community Enterprise Operating System; also known as CentOS Linux) is a discontinued Linux distribution that provided a free and open-source community-supported computing platform, functionally compatible with its upstream (softw ...
. In 2015, CERN switched away from Scientific Linux to CentOS.
In 2023 CERN switched to a mix of RHEL9 for high level operation and Debian
Debian () is a free and open-source software, free and open source Linux distribution, developed by the Debian Project, which was established by Ian Murdock in August 1993. Debian is one of the oldest operating systems based on the Linux kerne ...
for controller hardware that would no longer be supported by RedHat. Since the controller hardware is considered functional for the near to mid-range future by falling back to Debian CERN can save on spending on new hardware only to be compatible with RHEL9.
Distributed computing
Distributed computing is a field of computer science that studies distributed systems, defined as computer systems whose inter-communicating components are located on different networked computers.
The components of a distributed system commu ...
resources for analysis by end-user physicists are provided by multiple federations across the Europe, Asia Pacific and the Americas.
See also
* CERN openlab
References
External links
*
{{CERN
Cyberinfrastructure
E-Science
Information technology organizations based in Europe
CERN