Recent from talks
Nothing was collected or created yet.
Data center
View on Wikipedia

A data center is a building, a dedicated space within a building, or a group of buildings[1] used to house computer systems and associated components, such as telecommunications and storage systems.[2][3]
Since IT operations are crucial for business continuity, it generally includes redundant or backup components and infrastructure for power supply, data communication connections, environmental controls (e.g., air conditioning, fire suppression), and various security devices. A large data center is an industrial-scale operation using as much electricity as a medium town.[4] Estimated global data center electricity consumption in 2022 was 240–340 TWh, or roughly 1–1.3% of global electricity demand. This excludes energy used for cryptocurrency mining, which was estimated to be around 110 TWh in 2022, or another 0.4% of global electricity demand.[5] The IEA projects that data center electric use could double between 2022 and 2026.[6] High demand for electricity from data centers, including by cryptomining and artificial intelligence, has also increased strain on local electric grids and increased electricity prices in some markets.
Data centers can vary widely in terms of size, power requirements, redundancy, and overall structure. Four common categories used to segment types of data centers are onsite data centers, colocation facilities, hyperscale data centers, and edge data centers.[7] In particular, colocation centers often host private peering connections between their customers, internet transit providers, cloud providers,[8][9] meet-me rooms for connecting customers together[10] Internet exchange points,[11][12] and landing points and terminal equipment for fiber optic submarine communication cables,[13] connecting the internet.[14]
History
[edit]
Data centers have their roots in the huge computer rooms of the 1940s, typified by ENIAC, one of the earliest examples of a data center.[15][note 1] Early computer systems, complex to operate and maintain, required a special environment in which to operate. Many cables were necessary to connect all the components, and methods to accommodate and organize these were devised such as standard racks to mount equipment, raised floors, and cable trays (installed overhead or under the elevated floor). A single mainframe required a great deal of power and had to be cooled to avoid overheating. Security became important – computers were expensive, and were often used for military purposes.[15][note 2] Basic design guidelines for controlling access to the computer room were therefore devised.
During the microcomputer industry boom of the 1980s, users started to deploy computers everywhere, in many cases with little or no care about operating requirements. However, as information technology (IT) operations started to grow in complexity, organizations grew aware of the need to control IT resources. The availability of inexpensive networking equipment, coupled with new standards for the network structured cabling, made it possible to use a hierarchical design that put the servers in a specific room inside the company. The use of the term data center, as applied to specially designed computer rooms, started to gain popular recognition about this time.[15][note 3]
A boom of data centers came during the dot-com bubble of 1997–2000.[16][note 4] Companies needed fast Internet connectivity and non-stop operation to deploy systems and to establish a presence on the Internet. Installing such equipment was not viable for many smaller companies. Many companies started building very large facilities, called internet data centers (IDCs),[17] which provide enhanced capabilities, such as crossover backup: "If a Bell Atlantic line is cut, we can transfer them to ... to minimize the time of outage."[17]
The term cloud data centers (CDCs) has been used.[18] Increasingly, the division of these terms has almost disappeared and they are being integrated into the term data center.[19]
The global data center market saw steady growth in the 2010s, with a notable acceleration in the latter half of the decade. According to Gartner, worldwide data center infrastructure spending reached $200 billion in 2021, representing a 6% increase from 2020 despite the economic challenges posed by the COVID-19 pandemic.[20]
The latter part of the 2010s and early 2020s saw a significant shift towards AI and machine learning applications, generating a global boom for more powerful and efficient data center infrastructure. As of March 2021, global data creation was projected to grow to more than 180 zettabytes by 2025, up from 64.2 zettabytes in 2020.[21]
The United States is currently the foremost leader in data center infrastructure, hosting 5,381 data centers as of March 2024, the highest number of any country worldwide.[22] According to global consultancy McKinsey & Co., U.S. market demand is expected to double to 35 gigawatts (GW) by 2030, up from 17 GW in 2022.[23] As of 2023, the U.S. accounts for roughly 40 percent of the global market.[23] In 2025, it was estimated that the U.S. GDP growth was only 0.1% without the investments in data centers.[24]
A study published by the Electric Power Research Institute (EPRI) in May 2024 estimates U.S. data center power consumption could range from 4.6% to 9.1% of the country's generation by 2030.[25] As of 2023, about 80% of U.S. data center load was concentrated in 15 states, led by Virginia and Texas.[25]
Requirements for modern data centers
[edit]
Modernization and data center transformation enhances performance and energy efficiency.[26]
Information security is also a concern, and for this reason, a data center has to offer a secure environment that minimizes the chances of a security breach. A data center must, therefore, keep high standards for assuring the integrity and functionality of its hosted computer environment.
Industry research company International Data Corporation (IDC) puts the average age of a data center at nine years old.[26] Gartner, another research company, says data centers older than seven years are obsolete.[27] The growth in data (163 zettabytes by 2025[28]) is one factor driving the need for data centers to modernize.
Focus on modernization is not new: concern about obsolete equipment was decried in 2007,[29] and in 2011 Uptime Institute was concerned about the age of the equipment therein.[note 5] By 2018 concern had shifted once again, this time to the age of the staff: "data center staff are aging faster than the equipment."[30]
Meeting standards for data centers
[edit]The Telecommunications Industry Association's Telecommunications Infrastructure Standard for Data Centers[31] specifies the minimum requirements for telecommunications infrastructure of data centers and computer rooms including single tenant enterprise data centers and multi-tenant Internet hosting data centers. The topology proposed in this document is intended to be applicable to any size data center.[32]
Telcordia GR-3160, NEBS Requirements for Telecommunications Data Center Equipment and Spaces,[33] provides guidelines for data center spaces within telecommunications networks, and environmental requirements for the equipment intended for installation in those spaces. These criteria were developed jointly by Telcordia and industry representatives. They may be applied to data center spaces housing data processing or Information Technology (IT) equipment. The equipment may be used to:
- Operate and manage a carrier's telecommunication network
- Provide data center based applications directly to the carrier's customers
- Provide hosted applications for a third party to provide services to their customers
- Provide a combination of these and similar data center applications
Data center transformation
[edit]Data center transformation takes a step-by-step approach through integrated projects carried out over time. This differs from a traditional method of data center upgrades that takes a serial and siloed approach.[34] The typical projects within a data center transformation initiative include standardization/consolidation, virtualization, automation and security.
Data center consolidation consists in reducing the number of data centers[35][36] and avoiding server sprawl (both physical and virtual),[37] often includes replacing aging data center equipment. Likewise, this process is aided by standardization which makes these systems follow a uniform set of configurations in order to simplify and improve efficiency.[36] Virtualization, on the other hand, lowers capital, operational expenses,[38] and reduces energy consumption.[39] Virtualized desktops can be hosted in data centers and rented out on a subscription basis.[40] Investment bank Lazard Capital Markets estimated that in 2008, 48 percent of enterprise operations will be virtualized by 2012. Gartner views virtualization as a catalyst for modernization.[41] Automating tasks such as provisioning, configuration, patching, release management, and compliance are other ways in which data centers can be upgraded.These changes are needed not just when facing fewer skilled IT workers.[42] Lastly, security initiatives integrates the protection of virtual systems with existing security of physical infrastructures.[43]
Raised floor
[edit]
A raised floor standards guide named GR-2930 was developed by Telcordia Technologies, a subsidiary of Ericsson.[44]
The first raised floor computer room was made by IBM in 1956,[45] and they have "been around since the 1960s";[46] it was during the 1970s that it became more common for computer centers to thereby allow cool air to circulate more efficiently.[47][48]
The first purpose of the raised floor was to allow access for wiring.[45]
Lights out
[edit]The lights-out[49] data center, also known as a darkened or a dark data center, is a data center that, ideally, has all but eliminated the need for direct access by personnel, except under extraordinary circumstances. Because of the lack of need for staff to enter the data center, it can be operated without lighting. All of the devices are accessed and managed by remote systems, with automation programs used to perform unattended operations. In addition to the energy savings, reduction in staffing costs and the ability to locate the site further from population centers, implementing a lights-out data center reduces the threat of malicious attacks upon the infrastructure.[50][51]
Noise levels
[edit]Generally speaking, local authorities prefer noise levels at data centers to be "10 dB below the existing night-time background noise level at the nearest residence."[52]
OSHA regulations require monitoring of noise levels inside data centers if noise exceeds 85 decibels.[53] The average noise level in server areas of a data center may reach as high as 92-96 dB(A).[54]
Residents living near data centers have described the sound as "a high-pitched whirring noise 24/7", saying "It's like being on a tarmac with an airplane engine running constantly ... Except that the airplane keeps idling and never leaves."[55][56][57][58]
External sources of noise include HVAC equipment and energy generators.[59][60]
Data center design
[edit]The field of data center design has been growing for decades in various directions, including new construction big and small along with the creative re-use of existing facilities, like abandoned retail space, old salt mines and war-era bunkers.
- a 65-story data center has already been proposed[61]
- the number of data centers as of 2016 had grown beyond 3 million USA-wide, and more than triple that number worldwide[16]
Local building codes may govern the minimum ceiling heights and other parameters. Some of the considerations in the design of data centers are:

- Size - one room of a building, one or more floors, or an entire building;
- Capacity - can hold up to or past 1,000 servers;[62]
- Other considerations - Space, power, cooling, and costs in the data center;[63]
- Mechanical engineering infrastructure - heating, ventilation and air conditioning (HVAC); humidification and dehumidification equipment; pressurization;[64]
- Electrical engineering infrastructure design - utility service planning; distribution, switching and bypass from power sources; uninterruptible power source (UPS) systems; and more.[64][65]
- Screening to improve design perception during permitting[66]

Design criteria and trade-offs
[edit]- Availability expectations: The costs of avoiding downtime should not exceed the cost of the downtime itself[67]
- Site selection: Location factors include proximity to power grids, telecommunications infrastructure, networking services, transportation lines and emergency services. Other considerations should include flight paths, neighboring power drains, geological risks, and climate (associated with cooling costs).[68]
- Often, power availability is the hardest to change.
High availability
[edit]Various metrics exist for measuring the data-availability that results from data-center availability beyond 95% uptime, with the top of the scale counting how many nines can be placed after 99%.[69]
Modularity and flexibility
[edit]Modularity and flexibility are key elements in allowing for a data center to grow and change over time. Data center modules are pre-engineered, standardized building blocks that can be easily configured and moved as needed.[70]
A modular data center may consist of data center equipment contained within shipping containers or similar portable containers.[71] Components of the data center can be prefabricated and standardized which facilitates moving if needed.[72]
Electrical power
[edit]

Backup power consists of one or more uninterruptible power supplies, battery banks, diesel or gas turbine generators.[73]
To prevent single points of failure, all elements of the electrical systems, including backup systems, are typically given redundant copies, and critical servers are connected to both the A-side and B-side power feeds. This arrangement is often made to achieve N+1 redundancy in the systems. Static transfer switches are sometimes used to ensure instantaneous switchover from one supply to the other in the event of a power failure.
Low-voltage cable routing
[edit]Options for low voltage cable routing might include; Data cabling that is routed through overhead cable trays;[74] Raised floor cabling, both for security reasons and to avoid the extra cost of cooling systems over the racks; Smaller/less expensive data centers may use anti-static tiles instead for a flooring surface.
Airflow and environmental control
[edit]Airflow management is the practice of achieving data center cooling efficiency by preventing the recirculation of hot exhaust air and by reducing bypass airflow. Common approaches include hot-aisle/cold-aisle containment and the deployment of in-row cooling units which position cooling directly between server racks to intercept exhaust heat before it mixes with room air.[75]
Maintaining suitable temperature and humidity levels is critical to preventing equipment damage caused by overheating. Overheating can cause components, usually the silicon or copper of the wires or circuits to melt, causing connections to loosen, causing fire hazards. Typical control methods include:
- Air conditioning
- Indirect cooling, such as the use of outside air,[76][77][note 6] Indirect Evaporative Cooling (IDEC) units, and seawater cooling.
Humidity control not only prevents moisture-related issues: importantly, excess humidity can cause dust to adhere more readily to fan blades and heat sinks, impeding air cooling leading to higher temperatures.[78]
Aisle containment
[edit]Cold aisle containment is done by exposing the rear of equipment racks, while the fronts of the servers are enclosed with doors and covers. This is similar to how large-scale food companies refrigerate and store their products.

Computer cabinets/Server farms are often organized for containment of hot/cold aisles. Proper air duct placement prevents the cold and hot air from mixing. Rows of cabinets are paired to face each other so that the cool and hot air intakes and exhausts do not mix air, which would severely reduce cooling efficiency.
Alternatively, a range of underfloor panels can create efficient cold air pathways directed to the raised-floor vented tiles. Either the cold aisle or the hot aisle can be contained.[79]
Another option is fitting cabinets with vertical exhaust duct chimneys.[80] Hot exhaust pipes/vents/ducts can direct the air into a Plenum space above a Dropped ceiling and back to the cooling units or to outside vents. With this configuration, traditional hot/cold aisle configuration is not a requirement.[81]
Fire protection
[edit]
Data centers feature fire protection systems, including passive and Active Design elements, as well as implementation of fire prevention programs in operations. Smoke detectors are usually installed to provide early warning of a fire at its incipient stage.
Although the main room usually does not allow Wet Pipe-based Systems due to the fragile nature of Circuit-boards, there still exist systems that can be used in the rest of the facility or in cold/hot aisle air circulation systems that are closed systems, such as:[82]
- Sprinkler systems
- Misting, using high pressure to create extremely small water droplets, which can be used in sensitive rooms due to the nature of the droplets.
However, there also exist other means to put out fires, especially in Sensitive areas, usually using Gaseous fire suppression, of which Halon gas was the most popular, until the negative effects of producing and using it were discovered.[1]
Security
[edit]Physical access is usually restricted. Layered security often starts with fencing, bollards and mantraps.[83] Video camera surveillance and permanent security guards are almost always present if the data center is large or contains sensitive information. Fingerprint recognition mantraps are starting to be commonplace.
Logging access is required by some data protection regulations; some organizations tightly link this to access control systems. Multiple log entries can occur at the main entrance, entrances to internal rooms, and at equipment cabinets. Access control at cabinets can be integrated with intelligent power distribution units, so that locks are networked through the same appliance.[84]
Energy use
[edit]
Energy use is a central issue for data centers. Power draw ranges from a few kW for a rack of servers in a closet to several tens of MW for large facilities. Some facilities have power densities more than 100 times that of a typical office building.[85] For higher power density facilities, electricity costs are a dominant operating expense and account for over 10% of the total cost of ownership (TCO) of a data center.[86]
Greenhouse gas emissions
[edit]In 2020, data centers (excluding cryptocurrency mining) and data transmission each used about 1% of world electricity.[87] Although some of this electricity was low carbon, the IEA called for more "government and industry efforts on energy efficiency, renewables procurement and RD&D",[87] as some data centers still use electricity generated by fossil fuels.[88] They also said that lifecycle emissions should be considered, that is including embodied emissions, such as in buildings.[87] Data centers are estimated to have been responsible for 0.5% of US greenhouse gas emissions in 2018.[89] Some Chinese companies, such as Tencent, have pledged to be carbon neutral by 2030, while others such as Alibaba have been criticized by Greenpeace for not committing to become carbon neutral.[90] Google and Microsoft now each consume more power than some fairly big countries, surpassing the consumption of more than 100 countries.[91]
Energy efficiency and overhead
[edit]The most commonly used energy efficiency metric for data centers is power usage effectiveness (PUE), calculated as the ratio of total power entering the data center divided by the power used by IT equipment.
PUE measures the percentage of power used by overhead devices (cooling, lighting, etc.). The average USA data center has a PUE of 2.0,[92] meaning two watts of total power (overhead + IT equipment) for every watt delivered to IT equipment. State-of-the-art data centers are estimated to have a PUE of roughly 1.2.[93] Google publishes quarterly efficiency metrics from its data centers in operation.[94] PUEs of as low as 1.01 have been achieved with two phase immersion cooling.[95]
The U.S. Environmental Protection Agency has an Energy Star rating for standalone or large data centers. To qualify for the ecolabel, a data center must be within the top quartile in energy efficiency of all reported facilities.[96] The Energy Efficiency Improvement Act of 2015 (United States) requires federal facilities—including data centers—to operate more efficiently. California's Title 24 (2014) of the California Code of Regulations mandates that every newly constructed data center must have some form of airflow containment in place to optimize energy efficiency.
The European Union also has a similar initiative: EU Code of Conduct for Data Centres.[97]
Energy use analysis and projects
[edit]The focus of measuring and analyzing energy use goes beyond what is used by IT equipment; facility support hardware such as chillers and fans also use energy.[98]
In 2011, server racks in data centers were designed for more than 25 kW and the typical server was estimated to waste about 30% of the electricity it consumed. The energy demand for information storage systems is also rising. A high-availability data center is estimated to have a 1 megawatt (MW) demand and consume $20,000,000 in electricity over its lifetime, with cooling representing 35% to 45% of the data center's total cost of ownership. Calculations show that in two years, the cost of powering and cooling a server could be equal to the cost of purchasing the server hardware.[99] Research in 2018 has shown that a substantial amount of energy could still be conserved by optimizing IT refresh rates and increasing server utilization.[100] Research for optimizing task scheduling is also underway, with researchers looking to implement energy-efficient scheduling algorithms that could reduce energy consumption by anywhere between 6% to 44%.[101]
In 2011, Facebook, Rackspace and others founded the Open Compute Project (OCP) to develop and publish open standards for greener data center computing technologies. As part of the project, Facebook published the designs of its server, which it had built for its first dedicated data center in Prineville. Making servers taller left space for more effective heat sinks and enabled the use of fans that moved more air with less energy. By not buying commercial off-the-shelf servers, energy consumption due to unnecessary expansion slots on the motherboard and unneeded components, such as a graphics card, was also saved.[102] In 2016, Google joined the project and published the designs of its 48V DC shallow data center rack. This design had long been part of Google data centers. By eliminating the multiple transformers usually deployed in data centers, Google had achieved a 30% increase in energy efficiency.[103] In 2017, sales for data center hardware built to OCP designs topped $1.2 billion and are expected to reach $6 billion by 2021.[102]
Power and cooling analysis
[edit]
Power is the largest recurring cost to the user of a data center.[104] Cooling at or below 70 °F (21 °C) wastes money and energy.[104] Furthermore, overcooling equipment in environments with a high relative humidity can expose equipment to a high amount of moisture that facilitates the growth of salt deposits on conductive filaments in the circuitry.[105]
A power and cooling analysis, also referred to as a thermal assessment, measures the relative temperatures in specific areas as well as the capacity of the cooling systems to handle specific ambient temperatures.[106] A power and cooling analysis can help to identify hot spots, over-cooled areas that can handle greater power use density, the breakpoint of equipment loading, the effectiveness of a raised-floor strategy, and optimal equipment positioning (such as AC units) to balance temperatures across the data center. Power cooling density is a measure of how much square footage the center can cool at maximum capacity.[107] The cooling of data centers is the second largest power consumer after servers. The cooling energy varies from 10% of the total energy consumption in the most efficient data centers and goes up to 45% in standard air-cooled data centers.
Energy efficiency analysis
[edit]An energy efficiency analysis measures the energy use of data center IT and facilities equipment. A typical energy efficiency analysis measures factors such as a data center's Power Use Effectiveness (PUE) against industry standards, identifies mechanical and electrical sources of inefficiency, and identifies air-management metrics.[108] However, the limitation of most current metrics and approaches is that they do not include IT in the analysis. Case studies have shown that by addressing energy efficiency holistically in a data center, major efficiencies can be achieved that are not possible otherwise.[109]
Computational Fluid Dynamics (CFD) analysis
[edit]This type of analysis uses sophisticated tools and techniques to understand the unique thermal conditions present in each data center—predicting the temperature, airflow, and pressure behavior of a data center to assess performance and energy consumption, using numerical modeling.[110] By predicting the effects of these environmental conditions, CFD analysis of a data center can be used to predict the impact of high-density racks mixed with low-density racks[111] and the onward impact on cooling resources, poor infrastructure management practices, and AC failure or AC shutdown for scheduled maintenance.
Thermal zone mapping
[edit]Thermal zone mapping uses sensors and computer modeling to create a three-dimensional image of the hot and cool zones in a data center.[112]
This information can help to identify optimal positioning of data center equipment. For example, critical servers might be placed in a cool zone that is serviced by redundant AC units.
Green data centers
[edit]
Data centers use a lot of power, consumed by two main usages: The power required to run the actual equipment and then the power required to cool the equipment. Power efficiency reduces the first category.
Cooling cost reduction through natural means includes location decisions: When the focus is avoiding good fiber connectivity, power grid connections, and people concentrations to manage the equipment, a data center can be miles away from the users. Mass data centers like Google or Facebook do not need to be near population centers. Arctic locations that can use outside air, which provides cooling, are becoming more popular.[113]
Renewable electricity sources are another plus. Thus countries with favorable conditions, such as Canada,[114] Finland,[115] Sweden,[116] Norway,[117] and Switzerland[118] are trying to attract cloud computing data centers.
Singapore lifted a three-year ban on new data centers in April 2022. A major data center hub for the Asia-Pacific region,[119] Singapore lifted its moratorium on new data center projects in 2022, granting 4 new projects, but rejecting more than 16 data center applications from over 20 new data centers applications received. Singapore's new data centers shall meet very strict green technology criteria including "Water Usage Effectiveness (WUE) of 2.0/MWh, Power Usage Effectiveness (PUE) of less than 1.3, and have a "Platinum certification under Singapore's BCA-IMDA Green Mark for New Data Centre" criteria that clearly addressed decarbonization and use of hydrogen cells or solar panels.[120][121][122][123]
Direct current data centers
[edit]Direct current data centers are data centers that produce direct current on site with solar panels and store the electricity on site in a battery storage power station. Computers run on direct current and the need for inverting the AC power from the grid would be eliminated. The data center site could still use AC power as a grid-as-a-backup solution. DC data centers could be 10% more efficient and use less floor space for inverting components.[124][125]
Energy reuse
[edit]It is very difficult to reuse the heat which comes from air-cooled data centers. For this reason, data center infrastructures are more often equipped with heat pumps.[126]
An alternative to heat pumps is the adoption of liquid cooling throughout a data center. Different liquid cooling techniques are mixed and matched to allow for a fully liquid-cooled infrastructure that captures all heat with water. Different liquid technologies are categorized in 3 main groups, indirect liquid cooling (water-cooled racks), direct liquid cooling (direct-to-chip cooling) and total liquid cooling (complete immersion in liquid, see server immersion cooling). This combination of technologies allows the creation of a thermal cascade as part of temperature chaining scenarios to create high-temperature water outputs from the data center.[citation needed]
Impact on electricity prices
[edit]Cryptomining and the artificial intelligence boom of the 2020s has also led to increased demand for electricity,[127][128] that the IEA expects could double global overall data center demand for electricity between 2022 and 2026.[6] The US could see its share of the electricity market going to data centers increase from 4% to 6% over those four years.[6] Bitcoin used up 2% of US electricity in 2023.[129] This has led to increased electricity prices in some regions,[130] particularly in regions with lots of data centers like Santa Clara, California[131] and upstate New York.[132] Data centers have also generated concerns in Northern Virginia about whether residents will have to foot the bill for future power lines.[129] It has also made it harder to develop housing in London.[133] A Bank of America Institute report in July 2024 found that the increase in demand for electricity due in part to AI has been pushing electricity prices higher and is a significant contributor to electricity inflation.[134][135][136] A Harvard Law School report in March 2025 found that because utilities are increasingly in competition to attract data center contracts from big tech companies, they are likely hiding subsidies to those trillion-dollar companies in power prices by raising costs for American consumers.[137]
Water consumption and environmental impact
[edit]The rapid expansion of AI data centers has raised significant concerns over their water consumption, particularly in drought-prone regions. Cooling According to the International Energy Agency (IEA), a single 100-megawatt data center can use up to 2,000,000 litres (530,000 US gal) of water per day—equivalent to the daily consumption of 6,500 households.[138][139] Its water usage can be divided into three categories, on-site (direct usage from data centers), off-site (indirect usage from electricity), and supply-chain (water usage from manufacturing processes).[140]
On-site water use refers to the direct water consumed by the data center for the cooling of its equipment.[140] Water is used specifically for space humidification (adds moisture to the air), evaporative cooling systems (air is cooled before entering server rooms),[138] and cooling towers (water is used to remove heat from the facility). All of which consume a vast amount of clean water.[141]
Off-site is the indirect water usage from the electricity generated in data centers. It is estimated that 56% of U.S. data centers' electricity comes from fossil fuels, this process require water to operate the power plants and produce energy.[142]
Since 2022, more than two-thirds of new data centers have been built in water-stressed areas, including Texas, Arizona, Saudi Arabia, and India, where freshwater scarcity is already a critical issue. The global water footprint of data centers is estimated at 560 billion litres (150×109 US gal) annually, a figure projected to double by 2030 due to increasing AI demand.[143][144]
In regions like Aragon, Spain, Amazon's planned data centers are licensed to withdraw 755,720 cubic metres (612.67 acre⋅ft) of water per year, sparking conflicts with farmers who rely on the same dwindling supplies. Similar tensions have arisen in Chile, the Netherlands, and Uruguay, where communities protest the diversion of water for tech infrastructure.[143][145]
Tech companies, including Microsoft, Google, and Amazon, have pledged to become "water positive" by 2030, aiming to replenish more water than they consume. However, critics argue that such commitments often rely on water offsetting, which does not address acute local shortages.[143][145]
With at least 59 additional data centers planned for water-stressed U.S. regions by 2028, and AI's projected global water demand reaching 6.6 billion cubic metres (1,700×109 US gal) by 2027, experts warn of an unsustainable trajectory. As Arizona State University water policy expert Kathryn Sorensen asked: "Is the increase in tax revenue and the relatively paltry number of jobs worth the water?"[146][144]
Dynamic infrastructure
[edit]Dynamic infrastructure[147] provides the ability to intelligently, automatically and securely move workloads within a data center[148] anytime, anywhere, for migrations, provisioning,[149] to enhance performance, or building co-location facilities. It also facilitates performing routine maintenance on either physical or virtual systems all while minimizing interruption. A related concept is Composable Infrastructure, which allows for the dynamic reconfiguration of the available resources to suit needs, only when needed.[150]
Side benefits include
- reducing cost
- facilitating business continuity and high availability
- enabling cloud and grid computing.[151]
Network infrastructure
[edit]

Communications in data centers today are most often based on networks running the Internet protocol suite. Data centers contain a set of routers and switches that transport traffic between the servers and to the outside world[152] which are connected according to the data center network architecture. Redundancy of the internet connection is often provided by using two or more upstream service providers (see Multihoming).
Some of the servers at the data center are used for running the basic internet and intranet services needed by internal users in the organization, e.g., e-mail servers, proxy servers, and DNS servers.
Network security elements are also usually deployed: firewalls, VPN gateways, intrusion detection systems, and so on. Also common are monitoring systems for the network and some of the applications. Additional off-site monitoring systems are also typical, in case of a failure of communications inside the data center.
Software/data backup
[edit]Non-mutually exclusive options for data backup are:
- Onsite
- Offsite
Onsite is traditional,[153] and one of its major advantages is immediate availability.
Offsite backup storage
[edit]Data backup techniques include having an encrypted copy of the data offsite. Methods used for transporting data are:[154]
Modular data center
[edit]
For quick deployment or IT disaster recovery, several large hardware vendors have developed mobile/modular solutions that can be installed and made operational in a very short amount of time.
Micro data center
[edit]Micro data centers (MDCs) are access-level data centers which are smaller in size than traditional data centers but provide the same features.[157] They are typically located near the data source to reduce communication delays, as their small size allows several MDCs to be spread out over a wide area.[158][159] MDCs are well suited to user-facing, front end applications.[160] They are commonly used in edge computing and other areas where low latency data processing is needed.[161]
Data centers in space
[edit]
Data centers in space is a proposed idea to place a data center in outer space in low Earth orbit. The theoretical advantages are that of space-based solar power, in addition to aiding in weather forecasting and weather prediction computation from weather satellites,[162] and the ability to freely scale up.[163]
Challenges include temperature fluctuations, cosmic rays, and micrometeorites.[162]
References
[edit]- ^ "Cloud Computing Brings Sprawling Centers, but Few Jobs". The New York Times. August 27, 2016. Archived from the original on July 21, 2023. Retrieved December 21, 2018.
data center .. a giant .. facility .. 15 of these buildings, and six more .. under construction
- ^ "From Manhattan to Montvale". The New York Times. April 20, 1986. Archived from the original on December 16, 2022. Retrieved May 14, 2019.
- ^ Ashlee Vance (December 8, 2008). "Dell Sees Double With Data Center in a Container". The New York Times. Archived from the original on August 2, 2021. Retrieved May 14, 2019.
- ^ James Glanz (September 22, 2012). "Power, Pollution and the Internet". The New York Times. Archived from the original on 2019-05-16. Retrieved 2012-09-25.
- ^ "Data centres & networks". IEA. Archived from the original on 2023-10-06. Retrieved 2023-10-07.
- ^ a b c Calma, Justine (2024-01-24). "AI and crypto mining are driving up data centers' energy use". The Verge. Retrieved 2024-08-21.
- ^ "Types of Data Centers | How do you Choose the Right Data Center?". Maysteel Industries, LLC. Archived from the original on 2023-06-01. Retrieved 2023-10-07.
- ^ "Touring the PhoenixNAP Data Center". 22 June 2021.
- ^ "The rise and rebirth of carrier hotels". 6 October 2023.
- ^ Dave Bullock (April 3, 2008). "A Lesson in Internet Anatomy: The World's Densest Meet-Me Room". Wired. Retrieved 2022-06-16.
- ^ "AWASR, AMS-IX, and Alliance Networks launch Internet exchange in Oman". 13 September 2023.
- ^ "Data Center Tours: Equinix DC12, Ashburn, Virginia". 16 July 2024.
- ^ "How the Internet works: Submarine fiber, brains in jars, and coaxial cables". 26 May 2016.
- ^ "Equinix Expands Miami Data Center Key to Latin American Connectivity".
- ^ a b c Angela Bartels (August 31, 2011). "Data Center Evolution: 1960 to 2000". Archived from the original on October 24, 2018. Retrieved October 24, 2018.
- ^ a b Cynthia Harvey (July 10, 2017). "Data Center". Datamation. Archived from the original on December 16, 2022. Retrieved October 28, 2018.
- ^ a b John Holusha (May 14, 2000). "Commercial Property/Engine Room for the Internet; Combining a Data Center With a 'Telco Hotel'". The New York Times. Archived from the original on June 28, 2024. Retrieved June 23, 2019.
- ^ H Yuan (2015). "Workload-aware request routing in cloud data center using software-defined networking". Journal of Systems Engineering and Electronics. 26: 151–160. doi:10.1109/JSEE.2015.00020. S2CID 59487957.
- ^ Quentin Hardy (October 4, 2011). "A Data Center Power Solution". The New York Times. Archived from the original on December 16, 2022. Retrieved June 23, 2019.
- ^ "Gartner Says Worldwide Data Center Infrastructure Spending to Grow 6% in 2021". Gartner, Inc. Archived from the original on 2024-07-06. Retrieved 6 July 2024.
- ^ Petroc, Taylor. "Volume of data/information created, captured, copied, and consumed worldwide from 2010 to 2020, with forecasts from 2021 to 2025 (in zettabytes)". Statista. Archived from the original on 30 August 2024. Retrieved 6 July 2024.
- ^ Petroc, Taylor. "Leading countries by number of data centers as of March 2024". Statista. Archived from the original on 2022-12-16. Retrieved 2024-07-06.
- ^ a b "Investing in the rising data center economy". McKinsey & Co. Archived from the original on 30 August 2024. Retrieved 6 July 2024.
- ^ Lichtenberg, Nick (8 October 2025). "Without data centers, GDP growth was 0.1% in the first half of 2025, Harvard economist says". Yahoo Finance. Retrieved 19 October 2025.
- ^ a b "Powering Intelligence: Analyzing Artificial Intelligence and Data Center Energy Consumption". Electric Power Research Institute (EPRI). Retrieved 6 July 2024.
- ^ a b "Mukhar, Nicholas. "HP Updates Data Center Transformation Solutions," August 17, 2011". Archived from the original on August 12, 2012. Retrieved September 9, 2011.
- ^ "Sperling, Ed. "Next-Generation Data Centers," Forbes, March 15. 2010". Forbes.com. Archived from the original on 2023-12-10. Retrieved 2013-08-30.
- ^ "IDC white paper, sponsored by Seagate" (PDF). Archived (PDF) from the original on 2017-12-08. Retrieved 2018-01-11.
- ^ "Data centers are aging, unsuited for new technologies". December 10, 2007.
- ^ "Data center staff are aging faster than the equipment". Network World. August 30, 2018. Archived from the original on December 7, 2023. Retrieved December 21, 2018.
- ^ "TIA-942 Certified Data Centers - Consultants - Auditors - TIA-942.org". www.tia-942.org.
- ^ "Telecommunications Standards Development". Archived from the original on November 6, 2011. Retrieved November 7, 2011.
- ^ "GR-3160 - Telecommunications Data Center - Telcordia". telecom-info.njdepot.ericsson.net. Archived from the original on 2022-10-07. Retrieved 2021-01-19.
- ^ "Tang, Helen. "Three Signs it's time to transform your data center," August 3, 2010, Data Center Knowledge". Data Center Knowledge. 3 August 2010. Archived from the original on August 10, 2011. Retrieved September 9, 2011.
- ^ Darrow, Barb. "Welcome to the Great Data Center Consolidation". Fortune. Retrieved 2025-10-14.
- ^ a b "This Wave of Data Center Consolidation is Different from the First One". DataCenterKnowledge. Retrieved 2025-10-14.
- ^ "IBM Systems Magazine - Stop Virtual Server Sprawl". ibmsystemsmag.com. Archived from the original on 2018-10-23. Retrieved 2025-10-07.
- ^ "Carousel's Expert Walks Through Major Benefits of Virtualization". technews.tmcnet.com. Retrieved 2025-10-14.
- ^ "The New Urgency For Server Virtualization - Government - Enterprise Architecture - Informationweek". www.informationweek.com. Archived from the original on 2012-04-02. Retrieved 2025-10-14.
- ^ "UK Government Web Archive" (PDF). webarchive.nationalarchives.gov.uk. Archived from the original (PDF) on 2023-04-26. Retrieved 2025-10-14.
- ^ "Gartner: Virtualization Disrupts Server Vendors". DataCenterKnowledge. Retrieved 2025-10-14.
- ^ "Complexity: Growing Data Center Challenge". DataCenterKnowledge. Retrieved 2025-10-07.
- ^ "Securing the Data-Center Transformation Aligning Security and Data-Center Dynamics". Lippis Report. Archived from the original on 2017-06-25. Retrieved 2025-10-14.
- ^ "GR-2930 - NEBS: Raised Floor Requirements".
- ^ a b "Data Center Raised Floor History" (PDF).
- ^ "Raised Floor Info | Tips for Ordering Replacement Raised Floor Tiles". www.accessfloorsystems.com.
- ^ Hwaiyu Geng (2014). Data Center Handbook. John Wiley & Sons. ISBN 978-1-118-43663-9.
- ^ Steven Spinazzola (2005). "HVAC: The Challenge And Benefits of Under Floor Air Distribution Systems". FacilitiesNet.com. Archived from the original on 2017-03-29. Retrieved 2018-10-28.
- ^ "Premier 100 Q&A: HP's CIO sees 'lights-out' data centers". Informationweek. March 6, 2006. Archived from the original on July 12, 2019.
- ^ Victor Kasacavage (2002). Complete book of remote access: connectivity and security. The Auerbach Best Practices Series. CRC Press. p. 227. ISBN 0-8493-1253-1.
- ^ Roxanne E. Burkey; Charles V. Breakfield (2000). Designing a total data solution: technology, implementation and deployment. Auerbach Best Practices. CRC Press. p. 24. ISBN 0-8493-0893-3.
- ^ Clarke, Renaud (2020-07-01). "Acoustic Barriers for Data Centres". IAC Acoustics. Retrieved 2023-02-11.
- ^ Thibodeau, Patrick (2007-07-31). "That sound you hear? The next data center problem". Computerworld. Retrieved 2023-02-11.
- ^ Sensear. "Data Center Noise Levels". Sensear. Retrieved 2023-02-11.
- ^ Weisbrod, Katelyn (2023-02-10). "In Northern Virginia, a Coming Data Center Boom Sounds a Community Alarm". Inside Climate News. Retrieved 2023-02-11.
- ^ Judge, Peter (2022-07-19). "Prince William residents complain of "catastrophic noise" from data centers". DCD. Retrieved 2023-02-11.
- ^ Judge, Peter (2022-07-27). "Chicago residents complain of noise from Digital Realty data center". DCD. Retrieved 2023-02-11.
- ^ Phillips, Mark (2021-11-30). "Chandler to consider banning data centers amid noise complaints". ABC15 Arizona in Phoenix (KNXV). Retrieved 2023-02-11.
- ^ "Data Center Soundproofing and Noise Control- Reduce Server Noise". DDS Acoustical Specialties. Retrieved 2023-02-11.
- ^ Bosker, Bianca (2019-12-06). "Your "cloud" data is making noise on the ground". Marketplace. Retrieved 2023-02-11.
- ^ Patrick Thibodeau (April 12, 2016). "Envisioning a 65-story data center". Computerworld. Archived from the original on October 28, 2018. Retrieved October 28, 2018.
- ^ "Google container data center tour (video)". YouTube. 7 April 2009. Archived from the original on 2021-11-04.
- ^ "Romonet Offers Predictive Modeling Tool For Data Center Planning". June 29, 2011. Archived from the original on August 23, 2011. Retrieved February 8, 2012.
- ^ a b "BICSI News Magazine - May/June 2010". www.nxtbook.com. Archived from the original on 2019-04-20. Retrieved 2012-02-08.
- ^ "Hedging Your Data Center Power". Archived from the original on 2024-05-17. Retrieved 2012-02-08.
- ^ Nelson, Michael (2025-10-17). "Architectural Specialty Metals for Data Center Design". AMICO Architectural Metals. Retrieved 2025-10-17.
- ^ Clark, Jeffrey. "The Price of Data Center Availability—How much availability do you need?", Oct. 12, 2011, The Data Center Journal "Data Center Outsourcing in India projected to grow according to Gartner". Archived from the original on 2011-12-03. Retrieved 2012-02-08.
- ^ "Five tips on selecting a data center location".
- ^ "IBM zEnterprise EC12 Business Value Video". YouTube. Archived from the original on 2012-08-29.
- ^ Niles, Susan. "Standardization and Modularity in Data Center Physical Infrastructure," 2011, Schneider Electric, page 4. "Standardization and Modularity in Data Center Physical Infrastructure" (PDF). Archived from the original (PDF) on 2012-04-16. Retrieved 2012-02-08.
- ^ "Strategies for the Containerized Data Center". September 8, 2011.
- ^ Niccolai, James (2010-07-27). "HP says prefab data center cuts costs in half".
- ^ Detailed explanation of UPS topologies "EVALUATING THE ECONOMIC IMPACT OF UPS TECHNOLOGY" (PDF). Archived from the original (PDF) on 2010-11-22.
- ^ "Cable tray systems support cables' journey through the data center". Cabling Installation & Maintenance. 2016-04-01. Retrieved 2025-10-14.
- ^ Mike Fox (15 February 2012). "Stulz announced it has begun manufacturing In Row server cooling units under the name "CyberRow"". DataCenterFix. Archived from the original on 1 March 2012. Retrieved 27 February 2012.
- ^ "tw telecom and NYSERDA Announce Co-location Expansion". Reuters. 14 September 2009. Archived from the original on 26 September 2009.
- ^ "Air to air combat – indirect air cooling wars".
- ^ "The Importance of Humidity Management in Data Centres and comms Rooms". Treske Pty Limited. 10 February 2024.
- ^ Hot-Aisle vs. Cold-Aisle Containment for Data Centers, John Niemann, Kevin Brown, and Victor Avelar, APC by Schneider Electric White Paper 135, Revision 1
- ^ "US Patent Application for DUCTED EXHAUST EQUIPMENT ENCLOSURE Patent Application (Application #20180042143 issued February 8, 2018) - Justia Patents Search". patents.justia.com. Retrieved 2018-04-17.
- ^ "Airflow Management Basics – Comparing Containment Systems • Data Center Frontier". Data Center Frontier. 2017-07-27. Archived from the original on 2019-02-19. Retrieved 2018-04-17.
- ^ "Data Center Fire Suppression Systems: What Facility Managers Should Consider". Facilitiesnet. Archived from the original on 2024-05-22. Retrieved 2018-10-28.
- ^ Sarah D. Scalet (2005-11-01). "19 Ways to Build Physical Security Into a Data Center". Csoonline.com. Archived from the original on 2008-04-21. Retrieved 2013-08-30.
- ^ Systems and methods for controlling an electronic lock for a remote device, 2016-08-01, archived from the original on 2023-03-06, retrieved 2018-04-25
- ^ "Data Center Energy Consumption Trends". U.S. Department of Energy. Archived from the original on 2010-06-05. Retrieved 2010-06-10.
- ^ J. Koomey, C. Belady, M. Patterson, A. Santos, K.D. Lange: Assessing Trends Over Time in Performance, Costs, and Energy Use for Servers Released on the web August 17th, 2009.
- ^ a b c "Data Centres and Data Transmission Networks – Analysis". IEA. Archived from the original on 2023-07-05. Retrieved 2022-03-06.
- ^ Kantor, Alice (2021-05-18). "Big Tech races to clean up act as cloud energy use grows". Financial Times. Archived from the original on 2022-12-10. Retrieved 2022-03-06.
- ^ Siddik, Md Abu Bakar; Shehabi, Arman; Marston, Landon (2021-05-21). "The environmental footprint of data centers in the United States". Environmental Research Letters. 16 (6): 064017. Bibcode:2021ERL....16f4017S. doi:10.1088/1748-9326/abfba1. hdl:10919/109747. ISSN 1748-9326. S2CID 235282419.
- ^ James, Greg (2022-03-01). "Tencent pledges to achieve carbon neutrality by 2030". SupChina. Archived from the original on 2022-07-11. Retrieved 2022-03-06.
- ^ Craig Hale (2024-07-15). "Google and Microsoft now each consume more power than some fairly big countries". TechRadar. Retrieved 2024-07-18.
- ^ "Report to Congress on Server and Data Center Energy Efficiency" (PDF). U.S. Environmental Protection Agency ENERGY STAR Program.
- ^ "Data Center Energy Forecast" (PDF). Silicon Valley Leadership Group. Archived from the original (PDF) on 2011-07-07. Retrieved 2010-06-10.
- ^ "Efficiency: How we do it – Data centers". Archived from the original on 2019-09-30. Retrieved 2015-01-19.
- ^ "Immersion cooling firm LiquidStack launches as a stand-alone company". Archived from the original on 2024-02-28. Retrieved 2024-02-28.
- ^ Commentary on introduction of Energy Star for Data Centers "Introducing EPA ENERGY STAR for Data Centers". Jack Pouchet. 2010-09-27. Archived from the original (Web site) on 2010-09-25. Retrieved 2010-09-27.
- ^ "EU Code of Conduct for Data Centres". iet.jrc.ec.europa.eu. Archived from the original on 2013-08-11. Retrieved 2013-08-30.
- ^ "Reducing Data Center Power and Energy Consumption: Saving Money and "Going Green"" (PDF). www.gtsi.com. Archived from the original (PDF) on 2012-12-03. Retrieved 2012-02-08.
- ^ Daniel Minoli (2011). Designing Green Networks and Network Operations: Saving Run-the-Engine Costs. CRC Press. p. 5. ISBN 978-1-4398-1639-4.
- ^ Rabih Bashroush (2018). "A Comprehensive Reasoning Framework for Hardware Refresh in Data Centres". IEEE Transactions on Sustainable Computing. 3 (4): 209–220. Bibcode:2018ITSC....3..209B. doi:10.1109/TSUSC.2018.2795465. S2CID 54462006.
- ^ Kaur, Rupinder; Kaur, Gurjinder; Goraya, Major Singh (2025-05-05). "EESF: Energy-Efficient Scheduling Framework for Deadline-Constrained Workflows with Computation Speed Estimation Method in Cloud". Parallel Computing. 124 103139. doi:10.1016/j.parco.2025.103139. ISSN 0167-8191.
- ^ a b Peter Sayer (March 28, 2018). "What is the Open Compute Project?". NetworkWorld. Archived from the original on November 29, 2023. Retrieved February 3, 2019.
- ^ Peter Judge (March 9, 2016). "OCP Summit: Google joins and shares 48V tech". DCD Data center Dynamics. Archived from the original on February 3, 2019. Retrieved February 3, 2019.
- ^ a b Joe Cosmano (2009), Choosing a Data Center (PDF), Disaster Recovery Journal, retrieved 2012-07-21[permanent dead link]
- ^ David Garrett (July 9, 2004), "Heat Of The Moment", Processor, 26 (28), archived from the original on 2013-01-31, retrieved 2012-07-21
- ^ Needle, David (25 July 2007). "HP's Green Data Center Portfolio Keeps Growing". InternetNews. Archived from the original on Oct 25, 2020.
- ^ "How to Choose a Data Center", Inc., Nov 29, 2010, archived from the original on Mar 8, 2013, retrieved 2012-07-21
- ^ Kathryn, Siranosian (April 5, 2011). "HP Shows Companies How to Integrate Energy Management and Carbon Reduction". TriplePundit. Archived from the original on August 22, 2018. Retrieved February 8, 2012.
- ^ Rabih Bashroush; Eoin Woods (2017). "Architectural Principles for Energy-Aware Internet-Scale Applications". IEEE Software. 34 (3): 14–17. Bibcode:2017ISoft..34c..14B. doi:10.1109/MS.2017.60. S2CID 8984662.
- ^ Bullock, Michael. "Computation Fluid Dynamics - Hot topic at Data Center World," Transitional Data Services, March 18, 2010. Archived January 3, 2012, at the Wayback Machine
- ^ "Bouley, Dennis (editor). "Impact of Virtualization on Data Center Physical Infrastructure," The Green grid, 2010" (PDF). Archived from the original (PDF) on 2014-04-29. Retrieved 2012-02-08.
- ^ "HP Thermal Zone Mapping plots data center hot spots". SearchDataCenter. Archived from the original on 2021-01-26. Retrieved 2012-02-08.
- ^ "Fjord-cooled DC in Norway claims to be greenest". 23 December 2011. Retrieved 23 December 2011.
- ^ Canada Called Prime Real Estate for Massive Data Computers - Globe & Mail Retrieved June 29, 2011.
- ^ Finland - First Choice for Siting Your Cloud Computing Data Center.. Retrieved 4 August 2010.
- ^ "Stockholm sets sights on data center customers". Archived from the original on 19 August 2010. Retrieved 4 August 2010.
- ^ In a world of rapidly increasing carbon emissions from the ICT industry, Norway offers a sustainable solution Archived 2020-10-29 at the Wayback Machine Retrieved 1 March 2016.
- ^ Swiss Carbon-Neutral Servers Hit the Cloud. Archived 2017-07-03 at the Wayback Machine. Retrieved 4 August 2010.
- ^ Baxtel. "Singapore Data Centers & Colocation". baxtel.com. Retrieved 2024-09-18.
- ^ "Singapore authorities invite applications for new data centers". 20 July 2022.
- ^ "BCA-IMDA Green Mark for Data Centres Scheme". Infocomm Media Development Authority. Retrieved 2024-09-18.
- ^ "Singapore to free up 300MW for data centres". Capacity Media. 2024-05-30. Retrieved 2024-09-18.
- ^ "4 Proposals Selected from Data Centre Application". Infocomm Media Development Authority. Retrieved 2024-09-18.
- ^ Could DC win the new data center War of the Currents?
- ^ "Direct Current (DC) Power | Center of Expertise for Energy Efficiency in Data Centers".
- ^ "Data Center Cooling with Heat Recovery" (PDF). StockholmDataParks.com. January 23, 2017. Archived (PDF) from the original on 2017-03-25. Retrieved 2018-11-16.
- ^ Halper, Evan (2024-03-07). "Amid explosive demand, America is running out of power". Washington Post. Retrieved 2024-08-19.
- ^ Rogers, Reece (July 11, 2024). "AI's Energy Demands Are Out of Control. Welcome to the Internet's Hyper-Consumption Era". Wired. ISSN 1059-1028. Retrieved 2024-08-19.
- ^ a b Chow, Andrew R. (2024-06-12). "How AI Is Fueling a Boom in Data Centers and Energy Demand". TIME. Retrieved 2024-08-21.
- ^ Halper, Evan; O'Donovan, Caroline (November 1, 2024). "As data centers for AI strain the power grid, bills rise for everyday customers". Washington Post. Archived from the original on November 12, 2024. Retrieved November 1, 2024.
- ^ Petersen, Melody (2024-08-12). "Power-hungry AI data centers are raising electric bills and blackout risk". Los Angeles Times. Retrieved 2024-08-19.
- ^ Benetton, Matteo; Compiani, Giovanni; Morse, Adair (2023-08-12). "When cryptomining comes to town: High electricity use spillovers to the local economy". VoxEU. Retrieved 2024-08-20.
- ^ Vincent, James (2022-07-28). "The electricity demands of data centers are making it harder to build new homes in London". The Verge. Retrieved 2024-08-21.
- ^ Walton, Robert (July 8, 2024). "US electricity prices rise again as AI, onshoring may mean decades of power demand growth: BofA". Utility Dive. Retrieved 2024-08-19.
- ^ Mott, Filip De. "Utility bills are getting cheaper, but AI could spoil the party". Markets Insider. Retrieved 2024-08-19.
- ^ "A.I. Power Shortage. Plus, Oil & Gas Stock Picks - Barron's Streetwise Barron's Podcasts". Barron's. May 17, 2024. Retrieved 2024-08-21.
- ^ "Extracting Profits from the Public: How Utility Ratepayers Are Paying for Big Tech's Power – Environmental and Energy Law Program". eelp.law.harvard.edu. Retrieved 2025-08-26.
- ^ a b Abdullahi, Aminu (2025-05-09). "AI Data Centers Boom is Draining Water From Drought-Prone Areas". TechRepublic. Retrieved 2025-10-13.
- ^ "The AI Boom Is Draining Water From the Areas That Need It Most". Bloomberg.com. 2025-05-08. Retrieved 2025-10-13.
- ^ a b Li, Pengfei; Yang, Jianyi; Islam, Mohammad A.; Ren, Shaolei (2025-03-26), Making AI Less "Thirsty": Uncovering and Addressing the Secret Water Footprint of AI Models, arXiv:2304.03271
- ^ Lei, Nuoa; Lu, Jun; Shehabi, Arman; Masanet, Eric (2025-06-01). "The water use of data center workloads: A review and assessment of key determinants". Resources, Conservation and Recycling. 219 108310. Bibcode:2025RCR...21908310L. doi:10.1016/j.resconrec.2025.108310. ISSN 0921-3449.
- ^ Environmental and Energy Study Institute (EESI). "Data Centers and Water Consumption | Article | EESI". www.eesi.org. Retrieved 2025-10-14.
- ^ a b c Abdullahi, Aminu (9 May 2025). "AI Data Centers Boom is Draining Water From Drought-Prone Areas – Sustainability Tipping Point?". TechRepublic.
- ^ a b "AI Draining Water From Areas That Need It Most". Water Education Foundation. 8 May 2025.
- ^ a b Gordon, Cindy (25 February 2024). "AI Is Accelerating the Loss of Our Scarcest Natural Resource: Water". Forbes.
- ^ "AI's Thirst: How Data Centers Are Draining Water in Vulnerable Regions". Bloomberg. 2025.
- ^ "Method for Dynamic Information Technology Infrastructure Provisioning".
- ^ Meyler, Kerrie (April 29, 2008). "The Dynamic Datacenter". Network World.
- ^ "Computation on Demand: The Promise of Dynamic Provisioning".[permanent dead link]
- ^ "Just What the Heck Is Composable Infrastructure, Anyway?". IT Pro. July 14, 2016.
- ^ Montazerolghaem, Ahmadreza (2020-07-13). "Software-defined load-balanced data center: design, implementation and performance analysis" (PDF). Cluster Computing. 24 (2): 591–610. doi:10.1007/s10586-020-03134-x. ISSN 1386-7857. S2CID 220490312.
- ^ Mohammad Noormohammadpour; Cauligi Raghavendra (July 16, 2018). "Datacenter Traffic Control: Understanding Techniques and Tradeoffs". IEEE Communications Surveys & Tutorials. 20 (2): 1492–1525. arXiv:1712.03530. doi:10.1109/comst.2017.2782753. S2CID 28143006.
- ^ "Protecting Data Without Blowing The Budget, Part 1: Onsite Backup". Forbes. October 4, 2018.
- ^ "Iron Mountain vs Amazon Glacier: Total Cost Analysis" (PDF). Archived from the original (PDF) on 2018-10-28. Retrieved 2018-10-28.
- ^ What IBM calls "PTAM: Pickup Truck Access Method." "PTAM - Pickup Truck Access Method (disaster recovery slang)".
- ^ "Iron Mountain introduces cloud backup and management service". Network world. September 14, 2017. Archived from the original on February 18, 2024. Retrieved October 28, 2018.
- ^ Ibrahim, Rosdiazli; Porkumaran, K.; Kannan, Ramani; Nor, Nursyarizal Mohd; Prabakar, S. (2022-11-13). International Conference on Artificial Intelligence for Smart Community: AISC 2020, 17–18 December, Universiti Teknologi Petronas, Malaysia. Springer Nature. p. 461. ISBN 978-981-16-2183-3.
- ^ Guo, Song; Qu, Zhihao (2022-02-10). Edge Learning for Distributed Big Data Analytics: Theory, Algorithms, and System Design. Cambridge University Press. pp. 12–13. ISBN 978-1-108-83237-3.
- ^ Research Anthology on Edge Computing Protocols, Applications, and Integration. IGI Global. 2022-04-01. p. 55. ISBN 978-1-6684-5701-6.
- ^ Furht, Borko; Escalante, Armando (2011-12-09). Handbook of Data Intensive Computing. Springer Science & Business Media. p. 17. ISBN 978-1-4614-1414-8.
- ^ Srivastava, Gautam; Ghosh, Uttam; Lin, Jerry Chun-Wei (2023-06-24). Security and Risk Analysis for Intelligent Edge Computing. Springer Nature. p. 17. ISBN 978-3-031-28150-1.
- ^ a b "Projections and feasibility of data centers in space | TechTarget".
- ^ "This Startup Wants to Tackle AI Energy Demands with Data Centers in Space". 5 September 2024.
See also
[edit]- Colocation centre
- Computer cooling
- Data center management
- Dynamic infrastructure
- Electrical network
- Internet exchange point
- Internet hosting service
- Microsoft underwater data center
- Neher–McGrath method
- Network operations center
- Open Compute Project, by Facebook
- Peering
- Server farm
- Server room
- Server Room Environment Monitoring System
- Telecommunications network
- Utah Data Center
- Web hosting service
Notes
[edit]- ^ Old large computer rooms that housed machines like the U.S. Army's ENIAC, which were developed pre-1960 (1945), are now referred to as data centers.
- ^ Until the early 1960s, it was primarily the government that used computers, which were large mainframes housed in rooms that today we call data centers.
- ^ In the 1990s, network-connected minicomputers (servers) running without input or display devices were housed in the old computer rooms. These new "data centers" or "server rooms" were built within company walls, co-located with low-cost networking equipment.
- ^ There was considerable construction of data centers during the early 2000s, in the period of expanding dot-com businesses.
- ^ In May 2011, data center research organization Uptime Institute reported that 36 percent of the large companies it surveyed expect to exhaust IT capacity within the next 18 months. James Niccolai. "Data Centers Turn to Outsourcing to Meet Capacity Needs". CIO magazine. Archived from the original on 2011-11-15. Retrieved 2011-09-09.
- ^ Indirect systems can reduce or eliminate the need for mechanical chillers or conventional air conditioners, resulting in energy savings.
External links
[edit]- Lawrence Berkeley Lab - Research, development, demonstration, and deployment of energy-efficient technologies and practices for data centers
- DC Power For Data Centers Of The Future - FAQ: 380VDC testing and demonstration at a Sun data center.
- White Paper - Property Taxes: The New Challenge for Data Centers
- The European Commission H2020 EURECA Data Centre Project Archived 2021-08-25 at the Wayback Machine - Data centre energy efficiency guidelines, extensive online training material, case studies/lectures (under events page), and tools.
Data center
View on GrokipediaHistory
Origins in computing infrastructure
The infrastructure for data centers originated in the specialized facilities required to house and operate early electronic computers during the 1940s and 1950s, when computing hardware demanded substantial electrical power, cooling, and physical space to function reliably. The ENIAC, the first general-purpose electronic computer, completed in 1945 by the U.S. Army and the University of Pennsylvania, occupied a 1,800-square-foot room in Philadelphia, consumed up to 150 kilowatts of power, and generated immense heat from its 18,000 vacuum tubes, necessitating dedicated electrical distribution and rudimentary air conditioning systems to prevent failures.[1][10] Similar installations, such as the UNIVAC I delivered to the U.S. Census Bureau in 1951, required controlled environments with raised floors for underfloor cabling and ventilation, marking the initial shift from ad-hoc setups to purpose-built computing rooms focused on uptime and maintenance access.[3] In the 1950s, the proliferation of mainframe systems for military and commercial data processing amplified these requirements, as machines like the IBM 701 (1952) and IBM 704 (1954) processed batch jobs in centralized locations, often consuming tens of kilowatts and producing heat loads equivalent to dozens of households.[11] These early computer rooms incorporated features such as backup generators, electromagnetic shielding, and specialized HVAC to mitigate vacuum tube fragility and power fluctuations, laying the groundwork for modern data center redundancies; for instance, the SAGE system deployed in 1958 across 23 sites featured modular computing nodes with continuous operation mandates, driving innovations in fault-tolerant infrastructure.[3] Industry standards began emerging, with organizations like the American Standards Association publishing guidelines in the late 1950s for computer room design, emphasizing fire suppression, humidity control, and seismic bracing to ensure operational continuity.[12] By the early 1960s, transistorization reduced size and power needs but increased density and data volumes, prompting the consolidation of computing resources into "data processing departments" within corporations, equipped with tape libraries, printers, and operator consoles in climate-controlled spaces.[11] IBM's System/360 announcement in 1964 standardized architectures, accelerating the build-out of dedicated facilities that integrated power conditioning, diesel backups, and structured cabling—elements persisting in contemporary data centers—while shifting focus from scientific computation to enterprise transaction processing.[3] This era's infrastructure emphasized scalability through modular racking and environmental monitoring, directly influencing the evolution toward formalized data centers as computing became integral to business operations.[10]Growth during the internet era
The proliferation of the internet in the 1990s shifted data centers from enterprise-focused installations to hubs supporting public-facing digital services, as businesses and ISPs required reliable infrastructure for web hosting, email, and early e-commerce. Prior to this, data processing was largely siloed within organizations, but the commercialization of the World Wide Web—following its public debut in 1991—drove demand for shared facilities capable of handling network traffic and storage at scale. This era saw the emergence of colocation centers, enabling smaller entities to rent rack space, power, and connectivity without building proprietary sites.[13][14] The dot-com boom of the late 1990s accelerated this expansion dramatically, with internet startups fueling a construction frenzy to accommodate anticipated surges in online activity. Investments poured into new builds and retrofits, including the conversion of landmark structures into data centers to meet urgent needs for server capacity.[15][16] Colocation providers proliferated, offering tenants redundant power and cooling amid rapid scaling; for instance, facilities in key internet exchange points like Northern Virginia began clustering to minimize latency. However, speculative overbuilding—driven by projections of exponential traffic growth—resulted in excess capacity, as evidenced by billions spent on underutilized sites.[17][18] The 2000–2001 bust exposed vulnerabilities, with many operators facing bankruptcy due to unmet revenue expectations, yet it consolidated the industry by weeding out inefficient players and paving the way for sustained growth. Broadband adoption post-bust, coupled with Web 2.0 applications like social networking from the mid-2000s, sustained demand for enhanced processing and storage, leading to more efficient, carrier-neutral facilities. In the United States, this period mirrored broader trends, as federal agencies expanded from 432 data centers in 1998 to 2,094 by 2010 to support networked government operations.[19][3] The internet era thus established data centers as foundational to digital economies, transitioning from ad-hoc responses to strategic, high-reliability infrastructure.[20]Rise of cloud and hyperscale facilities
![Google data center in The Dalles, Oregon][float-right] The rise of cloud computing fundamentally reshaped data center architecture and ownership, shifting from siloed enterprise facilities to vast, shared infrastructures managed by a handful of dominant providers. Amazon Web Services (AWS) pioneered modern public cloud services with the launch of Simple Storage Service (S3) in March 2003 and Elastic Compute Cloud (EC2) in August 2006, enabling on-demand access to scalable computing resources over the internet.[22] This model rapidly gained traction as businesses sought to avoid the capital-intensive burden of maintaining proprietary data centers, leading to exponential growth in cloud adoption; by 2010, competitors like Microsoft Azure and Google App Engine had entered the market, intensifying competition and innovation in distributed computing.[23] Hyperscale data centers emerged as a direct response to the demands of cloud services, characterized by their immense scale—typically comprising thousands of servers across facilities exceeding 10,000 square feet—and engineered for rapid elasticity to handle massive workloads like web-scale applications and big data processing. The term "hyperscale" gained prominence in the early 2010s as companies such as Amazon, Google, Microsoft, and Meta invested heavily in custom-built campuses optimized for efficiency and low-latency global distribution.[24] These facilities consolidated computing power, achieving economies of scale unattainable by traditional enterprise setups, with hyperscalers capturing over 68% of cloud workloads by 2020 through modular designs and advanced automation.[25] Global proliferation accelerated post-2015, driven by surging data volumes from mobile internet, streaming, and e-commerce; the number of tracked hyperscale data centers grew at an average annual rate of 12% from 2018 onward, reaching 1,136 facilities by early 2025, with 137 new ones coming online in 2024 alone.[26] The United States dominates with 54% of total hyperscale capacity, fueled by tech hubs in Virginia and Oregon, while emerging markets saw expansions to support localized latency needs.[26] Market analyses project a compound annual growth rate (CAGR) of 9.58% for hyperscale infrastructure through 2030, underpinned by investments approaching $7 trillion globally by that decade's end to meet escalating compute demands.[27][28] This evolution reduced the number of organizations directly operating data centers, as cloud providers assumed the role of primary builders and operators, leasing capacity to end-users via APIs and shifting industry focus toward specialization in power efficiency, redundancy, and interconnectivity.[29] Hyperscalers' vertical integration—from hardware design to software orchestration—enabled unprecedented resource utilization, though it concentrated control among a few entities, raising questions about dependency and resilience that empirical data on uptime metrics (often exceeding 99.99%) has largely mitigated through redundant architectures.[24]AI-driven expansion since 2020
![Google data center in The Dalles][float-right] The surge in artificial intelligence applications, particularly large language models and generative AI following the release of models like GPT-3 in 2020 and ChatGPT in November 2022, has profoundly accelerated data center construction and capacity expansion. Training and inference for these models require vast computational resources, predominantly graphics processing units (GPUs) from NVIDIA, which consume significantly more power than traditional servers. This demand prompted hyperscale operators to prioritize AI-optimized facilities, shifting from general-purpose cloud infrastructure to specialized high-density racks supporting exaflop-scale computing.[30][31] Hyperscale providers such as Alphabet, Amazon, Microsoft, and Meta committed over $350 billion in 2025 to data center infrastructure, with projections exceeding $400 billion in 2026, largely to accommodate AI workloads. Globally, capital expenditures on data centers are forecasted to reach nearly $7 trillion by 2030, driven by the need for AI-ready capacity expected to grow at 33% annually from 2023 to 2030. In the United States, primary market supply hit a record 8,155 megawatts in the first half of 2025, reflecting a 43.4% year-over-year increase, while worldwide an estimated 10 gigawatts of hyperscale and colocation projects are set to break ground in 2025. The hyperscale data center market alone is projected to reach $106.7 billion in 2025, expanding at a 24.5% compound annual growth rate to $319 billion by 2030.[32][28][33][34][35][36] Power consumption has emerged as a critical bottleneck, with AI data centers driving a projected 165% increase in global electricity demand from the sector by 2030, according to Goldman Sachs estimates. Data centers accounted for 4% of U.S. electricity use in 2024, with demand expected to more than double by 2030; worldwide, electricity use by data centers is set to exceed 945 terawatt-hours by 2030, more than doubling from prior levels. In the U.S., AI-specific demand could reach 123 gigawatts by 2035, while new computational needs may add 100 gigawatts by 2030. Notably, 80-90% of AI computing power is now devoted to inference rather than training, amplifying ongoing operational demands on facilities. Global data center power capacity expanded to 81 gigawatts by 2024, with projections for 130 gigawatts by 2028 at a 16% compound annual growth rate from 2023.[37][5][38][39][40][30][41][42] This expansion has concentrated in regions with access to power and fiber connectivity, including the U.S. Midwest and Southeast, Europe, and Asia-Pacific, though grid constraints and regulatory hurdles have delayed some projects. The AI data center market is anticipated to grow from $17.73 billion in 2025 to $93.60 billion by 2032 at a 26.8% compound annual growth rate, underscoring the sector's transformation into a cornerstone of AI infrastructure. Innovations in modular designs and liquid cooling are being adopted to scale facilities faster and more efficiently for AI's dense workloads.[43][44]Design and Architecture
Site selection and operational requirements
Site selection for data centers emphasizes access to abundant, reliable electricity, as modern facilities can demand capacities exceeding 100 megawatts, with hyperscale operations scaling to gigawatts amid AI-driven growth.[45] Developers prioritize regions with stable grids, diverse utility sources, and proximity to renewable energy like hydroelectric or solar to mitigate costs and supply constraints.[46] [47] Fiber optic connectivity and closeness to internet exchange points are essential for minimizing latency, particularly for edge computing and real-time applications, often favoring established tech corridors over remote isolation.[48] [49] Sites must also offer expansive land for modular expansion, clear zoning for high-density builds, and logistical access via highways and airports for equipment delivery.[50] [51] Geohazards drive avoidance of flood-prone, seismic, or hurricane-vulnerable areas, with assessments incorporating historical data and climate projections to ensure long-term resilience; for instance, inland temperate zones reduce both disaster risk and cooling demands through natural ambient temperatures.[52] [53] Regulatory incentives, such as tax abatements, further influence choices, though operators scrutinize local policies for permitting delays that could impact timelines.[54] Operational requirements enforce redundancy in power delivery, typically via N+1 or 2N configurations with uninterruptible power supplies (UPS) and diesel generators capable of sustaining full load for hours during outages, targeting uptime exceeding 99.741% annually in Tier II facilities and higher in advanced tiers.[55] [56] Cooling infrastructure must counteract server heat densities up to 20-50 kW per rack, employing chilled water systems or air handlers to maintain inlet temperatures around 18-27°C per ASHRAE guidelines, with efficiency measured by power usage effectiveness (PUE) ratios ideally under 1.2 for leading operators.[57] [58] Physical security protocols include layered perimeters with fencing, ballistic-rated barriers, 24/7 surveillance, and biometric controls, integrated with environmental sensors for early detection of intrusions or failures.[59] [60] Fire suppression relies on clean agents like FM-200 to avoid equipment damage, complemented by compartmentalized designs and redundant HVAC for sustained habitability.[61] These elements collectively ensure operational continuity, with sites selected to support scalable integration of such systems without compromising causal dependencies like power-cooling interlocks.[62]Structural and modular design elements
Data centers employ robust structural elements to support heavy IT equipment and ensure operational stability. Standard server racks measure approximately 2 feet wide by 4 feet deep and are rated to hold up to 3,000 pounds, necessitating floors capable of distributing such loads evenly across the facility.[63] Raised access floors, a traditional structural feature, elevate the IT environment 12 to 24 inches above the subfloor, providing space for underfloor air distribution, power cabling, and data conduits while facilitating maintenance access through removable panels.[64] These floors typically consist of cement-filled steel or cast aluminum panels designed for lay-in installation, with perforated tiles offering 20-60% open area to optimize airflow for cooling.[65][66] However, raised floors face limitations in high-density environments, where modern racks can exceed 25 kW of power and require airflow volumes four times higher than legacy designs accommodate, often demanding unobstructed underfloor heights of at least 1 meter.[67] Consequently, some facilities shift to non-raised or slab-on-grade floors to support greater rack densities and heavier loads without structural constraints, though this may complicate cable management and airflow precision.[68] Overall, structural integrity also incorporates seismic bracing, fire-rated walls, and reinforced concrete slabs to withstand environmental stresses and comply with building codes.[63] Modular design elements enable scalable and rapid deployment through prefabricated components assembled on-site. Prefabricated modular data centers (PMDCs) integrate racks, power systems, and cooling into factory-built units, such as shipping container-based setups, allowing deployment in weeks rather than months compared to traditional construction.[69][70] Advantages include cost savings from reduced labor and site work, enhanced quality control via off-site fabrication, and flexibility for edge locations or temporary needs under 2 MW.[71][72] The global modular data center market, valued at $32.4 billion in 2024, is projected to reach $85.2 billion by 2030, driven by demands for quick scaling amid AI and edge computing growth.[73] These modules support incremental expansion by adding units without disrupting operations, though they may introduce integration complexities for larger hyperscale applications.[43][74]Electrical power systems
Electrical power systems in data centers deliver uninterrupted, high-reliability electricity to IT equipment, which typically consumes between 100-500 watts per server rack, scaling to megawatts for large facilities.[75] These systems prioritize redundancy to achieve uptime exceeding 99.999%, or "five nines," mitigating risks from grid failures or surges.[76] Primary power enters via utility feeds at medium voltages (e.g., 13.8 kV), stepped down through transformers to 480 V for distribution.[77] In the United States, data centers accounted for approximately 176 terawatt-hours (TWh) of electricity in 2023, representing 4.4% of national consumption, with projections indicating doubling or tripling by 2028 due to AI workloads.[78] Gigawatt-scale AI data centers, such as xAI's Colossus facility in Memphis, exemplify these demands, targeting 1 GW power capacity with challenges including grid interconnection delays and reliance on temporary gas turbines for initial operations to support rapid deployment of over 100,000 GPUs.[79] Uninterruptible power supplies (UPS) provide immediate bridging during outages, using lithium-ion battery banks—reliant on critical minerals such as lithium, cobalt, nickel, graphite, and manganese—or flywheels to sustain loads for minutes until generators activate.[80][81] These batteries are essential for managing the massive and unreliable power draws in AI data centers.[80] Diesel generators, often in N+1 configurations, offer extended backup, with capacities sized to handle full facility loads for hours or days; for instance, facilities may deploy multiple 2-3 MW units per module.[82] Redundancy architectures like N+1 (one extra component beyond minimum needs) or 2N (fully duplicated paths) ensure failover without capacity loss, as a single UPS or generator failure does not compromise operations.[75] Dual utility feeds and automatic transfer switches further enhance reliability, with systems tested under load to verify seamless transitions.[83] Power distribution occurs via switchgear, busways, and power distribution units (PDUs), which allocate conditioned electricity to racks at 208-415 V.[84] Remote power panels (RPPs) and rack PDUs enable granular metering and circuit protection, often with intelligent monitoring for real-time anomaly detection.[85] Efficiency is optimized through high-efficiency transformers and PDUs, reducing losses to under 2-3% in modern designs.[86] Global data center electricity use grew to 240-340 TWh in 2022, with annual increases of 15% projected through 2030 driven by compute-intensive applications.[87][88] Monitoring integrates sensors across transformers, UPS, and PDUs to track power quality metrics like harmonics and supraharmonics, which can degrade equipment if unmanaged.[89] Facilities often employ predictive maintenance via SCADA systems to preempt failures, aligning with Tier III/IV standards requiring concurrent maintainability.[90] As demands escalate, some operators explore on-site renewables or microgrids, though grid dependency persists for baseload stability.[91]Cooling and thermal management
Data centers generate substantial heat from IT equipment, where electrical power consumption converts to thermal output that must be dissipated to prevent hardware failure and maintain performance; cooling systems typically account for 30% to 40% of total facility energy use.[92][93] Effective thermal management relies on removing heat at rates matching rack power densities, which have risen from traditional levels of 5-10 kW per rack to over 50 kW in AI-driven workloads, necessitating advanced techniques beyond basic air handling.[94][95] In gigawatt-scale AI data centers, these densities can exceed 100 kW per rack, requiring liquid cooling innovations such as direct-to-chip and immersion systems to handle the thermal loads from dense GPU clusters.[43] Air cooling remains prevalent in lower-density facilities, employing computer room air conditioning (CRAC) units or handlers to circulate conditioned air through raised floors or overhead ducts, often with hot-aisle/cold-aisle containment to minimize mixing and improve efficiency.[96] These systems support densities up to 20 kW per rack but struggle with higher loads due to air's limited thermal capacity—approximately 1/3000th that of water—leading to increased fan power and hotspots.[97] Free cooling, leveraging external ambient air or evaporative methods when temperatures permit, can reduce mechanical cooling needs by 50-70% in suitable climates, contributing to power usage effectiveness (PUE) values as low as 1.2 in optimized setups.[98][99] Liquid cooling addresses limitations of air systems in high-density environments, particularly for AI and high-performance computing racks exceeding 50 kW, by using dielectric fluids or water loops to transfer heat directly from components like CPUs and GPUs.[100] Direct-to-chip methods pipe coolant to cold plates on processors, while immersion submerges servers in non-conductive liquids; these approaches can cut cooling energy by up to 27% compared to air and enable densities over 100 kW per rack with PUE improvements to below 1.1.[94][101] Hybrid systems, combining rear-door heat exchangers with air, offer retrofit paths for existing infrastructure, though challenges include leak risks, higher upfront costs, and the need for specialized maintenance.[102][103] Emerging innovations for AI-era demands include two-phase liquid cooling, where refrigerants boil to enhance heat absorption, and heat reuse for district heating or power generation, potentially recovering 20-30% of waste energy.[104][105] Regulatory pressures and efficiency benchmarks, such as those from the U.S. Department of Energy, drive adoption of variable-speed compressors and AI-optimized controls to dynamically match cooling to loads, reducing overall consumption amid projections of data center cooling market growth to $24 billion by 2032.[106][107] Despite air cooling's simplicity for legacy sites, liquid and advanced methods dominate hyperscale deployments for their superior causal efficacy in heat rejection at scale.[108]Networking infrastructure
Data center networking infrastructure encompasses the switches, routers, cabling systems, and protocols that interconnect servers, storage arrays, and other compute resources, facilitating low-latency, high-bandwidth data exchange essential for workload performance.[109] Traditional three-tier architectures, consisting of access, aggregation, and core layers, have historically supported hierarchical traffic flows but face bottlenecks in east-west server-to-server communication prevalent in modern cloud and AI environments.[110] In contrast, the leaf-spine (or spine-leaf) topology, based on Clos non-blocking fabrics, has become the dominant design since the mid-2010s, where leaf switches connect directly to servers at the top-of-rack level and link to spine switches for full-mesh interconnectivity, enabling scalable bandwidth and sub-millisecond latencies.[109][111] Core components include Ethernet switches operating at speeds from 100 Gbps to 400 Gbps per port in current deployments, with transitions to 800 Gbps using 112 Gbps electrical lanes for denser fabrics supporting AI training clusters.[112] Leaf switches typically feature 32 to 64 ports for server downlinks, while spine switches provide equivalent uplink capacity to maintain non-oversubscribed throughput across hundreds of racks.[113] Cabling relies heavily on multimode or single-mode fiber optics for inter-switch links, supplemented by direct-attach copper (DAC) or active optical cables (AOC) for shorter distances under 100 meters, ensuring signal integrity amid dense port counts.[114] Structured cabling systems, adhering to TIA-942 standards, organize pathways in underfloor trays or overhead ladders to minimize latency and support future upgrades.[115] Ethernet remains the standard protocol due to its cost-effectiveness, interoperability, and enhancements like RDMA over Converged Ethernet (RoCE) for low-overhead data transfer, increasingly supplanting InfiniBand in non-hyperscale AI back-end networks despite the latter's native advantages in remote direct memory access (RDMA) and zero-copy semantics.[116][117] InfiniBand, with speeds up to NDR 400 Gbps, persists in high-performance computing (HPC) and large-scale AI facilities for its sub-microsecond latencies and lossless fabric via adaptive routing, though Ethernet's ecosystem maturity drives projected dominance in enterprise AI data centers by 2030.[118][119] Software-defined networking (SDN) overlays, such as those using OpenFlow or BGP-EVPN, enable dynamic traffic orchestration and virtualization, optimizing for bursty AI workloads while integrating with external WAN links via border routers.[115] Recent advancements, including co-packaged optics in Nvidia's Spectrum-X Ethernet, promise further density improvements for 1.6 Tbps fabrics by reducing power and latency in optical-electrical conversions.[120]Physical and cybersecurity measures
Data centers employ layered physical security protocols to deter unauthorized access and protect critical infrastructure. Perimeter defenses typically include reinforced fencing, bollards to prevent vehicle ramming, and monitored entry gates with 24/7 surveillance cameras and security patrols.[121] [122] Facility-level controls extend to mantraps—dual-door vestibules that prevent tailgating—and biometric authentication systems such as fingerprint scanners or facial recognition for high-security zones.[123] [124] Inside server rooms, cabinet-level measures involve locked racks with individual access logs and intrusion detection sensors that trigger alarms upon tampering.[125] These protocols align with standards like ISO/IEC 27001, which emphasize defense-in-depth to minimize risks from physical breaches, as evidenced by reduced incident rates in compliant facilities.[126] Professional security personnel operate continuously, conducting patrols and verifying identities against pre-approved lists, with all access events logged for auditing.[127] [128] Visitor management requires escorted access and temporary badges, often integrated with video surveillance covering 100% of interior spaces without blind spots.[129] Motion detectors and environmental sensors further enhance detection, linking to central command centers for rapid response, as implemented in major providers' facilities since at least 2020.[59] Cybersecurity measures complement physical protections through logical controls and network defenses tailored to data centers' high-value assets. Firewalls, intrusion detection/prevention systems (IDS/IPS), and endpoint protection platforms form the core, segmenting networks to isolate operational technology (OT) from IT systems and mitigate ransomware threats, which surged 72% in reported cyber risks by 2025.[130] [131] Zero-trust architectures enforce continuous verification, requiring multi-factor authentication (MFA) and role-based access for all users, reducing unauthorized data exfiltration risks as per NIST SP 800-53 guidelines.[132] [133] Encryption at rest and in transit, alongside security information and event management (SIEM) tools for real-time monitoring, addresses evolving threats like phishing and supply-chain attacks, with best practices updated in 2023 to include AI-driven anomaly detection.[134] [135] Incident response plans, mandated under frameworks like NIST Cybersecurity Framework 2.0 (released 2024), incorporate regular penetration testing and employee training to counter human-error vulnerabilities, which account for over 70% of breaches in audited data centers.[136] [137] Compliance with SOC 2 and HIPAA further verifies these layered defenses, prioritizing empirical threat modeling over unverified vendor claims.[126]Operations and Reliability
High availability and redundancy
![Datacenter Backup Batteries showing UPS systems for power redundancy][float-right]High availability in data centers refers to the design and operational practices that minimize downtime, targeting uptime levels such as 99.99% or higher, which equates to no more than 52.6 minutes of annual outage.[138] This is achieved through redundancy, which involves duplicating critical components and pathways to eliminate single points of failure, enabling seamless failover during faults. Redundancy configurations include N (minimum required capacity without spares), N+1 (one additional unit for backup), 2N (fully duplicated systems), and 2N+1 (duplicated plus extra spares), with higher levels providing greater fault tolerance at increased cost.[139] The Uptime Institute's Tier Classification System standardizes these practices across four tiers, evaluating infrastructure for expected availability and resilience to failures. Tier I offers basic capacity without redundancy, susceptible to any disruption; Tier II adds partial redundancy for planned maintenance; Tier III requires N+1 redundancy for concurrent maintainability, allowing repairs without shutdown; and Tier IV demands 2N or equivalent for fault tolerance against multiple simultaneous failures, achieving 99.995% uptime or better.[140] [82] Many enterprise and hyperscale data centers operate at Tier III or IV, with certification verifying compliance through rigorous modeling and on-site audits.[141] Power systems exemplify redundancy implementation, featuring dual utility feeds, uninterruptible power supplies (UPS) with battery banks for seconds-to-minutes bridging, and diesel generators for extended outages. In an N+1 setup for a 1 MW load, five 250 kW UPS modules serve the requirement, tolerating one failure; 2N doubles the infrastructure for independent operation.[139] Generators typically follow N+1, with automatic transfer switches ensuring sub-10-second failover, though fuel storage and testing mitigate risks like wet stacking.[142] Cooling redundancy mirrors power, using multiple computer room air conditioners (CRACs) or chillers in N+1 arrays to prevent thermal shutdowns from unit failures or maintenance. Best practices recommend one spare unit per six active cooling units in large facilities, supplemented by diverse methods like air-side economizers or liquid cooling loops to enhance resilience without over-reliance on any single technology.[143] Network infrastructure employs redundant switches, fiber optic paths, and protocols like Border Gateway Protocol (BGP) for dynamic routing failover, advertising multiple prefixes to reroute traffic upon link or node failure within seconds.[144] At the IT layer, high availability incorporates server clustering, RAID storage arrays, and geographic distribution across facilities for disaster recovery, with metrics like mean time between failures (MTBF) and mean time to repair (MTTR) guiding designs. While redundancy raises capital expenditures—2N systems can double costs—empirical data from certified facilities shows it reduces outage frequency, prioritizing causal reliability over efficiency trade-offs in mission-critical environments.[83]
Automation and remote management
Data center automation encompasses software-driven processes that minimize manual intervention in IT operations, including server provisioning, network configuration, and resource allocation. These systems leverage orchestration tools such as Ansible, Puppet, and Chef to execute scripts across infrastructure, enabling rapid deployment and consistent configurations.[145] Adoption of automation has accelerated with the growth of hyperscale facilities, where manual management proves inefficient for handling thousands of servers. The global data center automation market expanded from $10.7 billion in 2024 to an estimated $12.45 billion in 2025, reflecting demand driven by cloud and AI workloads.[146] Remote management systems facilitate oversight and control of data center assets from off-site locations, often through out-of-band access methods that operate independently of primary networks. Technologies like IPMI (Intelligent Platform Management Interface) and vendor-specific solutions, such as Dell's iDRAC or HPE's iLO, allow administrators to monitor hardware status, reboot systems, and apply firmware updates remotely via secure protocols.[147] Console servers and KVM-over-IP switches provide serial console access and virtual keyboard-video-mouse control, essential for troubleshooting during network outages.[148] Data Center Infrastructure Management (DCIM) software integrates automation and remote capabilities by aggregating data from power, cooling, and IT equipment sensors to enable predictive analytics and automated responses. For instance, DCIM tools can trigger cooling adjustments based on real-time thermal data or alert on power anomalies, improving operational efficiency and reducing downtime.[149] Federal assessments indicate DCIM implementations enhance metering accuracy and Power Usage Effectiveness (PUE) tracking, with capabilities for capacity planning and asset management.[150] In practice, these systems support high availability by automating failover processes and integrating with monitoring platforms like Prometheus for anomaly detection.[151] Automation reduces human error in repetitive tasks, with studies showing up to 95% data storage optimization through deduplication integrated in automated workflows, though implementation requires robust integration to avoid silos.[152] Remote management mitigates risks in distributed environments, such as edge computing, by enabling centralized control, but demands secure protocols to counter vulnerabilities like unauthorized access.[153] Overall, these technologies underpin scalable operations, with market projections estimating the sector's growth to $23.80 billion by 2030 at a 17.83% CAGR.[154]Data management and backup strategies
Data management in data centers encompasses the systematic handling of data throughout its lifecycle, including storage, access, integrity verification, and retention to ensure availability and compliance with regulatory requirements. Storage technologies commonly employed include hard disk drives (HDDs) for high-capacity archival needs and solid-state drives (SSDs) for faster access in performance-critical applications, with hybrid arrays balancing cost and speed.[155] Redundancy mechanisms such as RAID configurations protect against single-drive failures by striping data with parity, though they incur higher overhead in large-scale environments compared to erasure coding, which fragments data into systematic chunks and generates parity blocks for reconstruction, enabling tolerance of multiple failures with lower storage overhead—typically 1.25x to 2x versus RAID's 2x or more.[155] [156] Backup strategies prioritize the creation of multiple data copies to mitigate loss from hardware failure, cyberattacks, or disasters, adhering to the 3-2-1 rule: maintaining three copies of data on two different media types, with one stored offsite or in a geographically separate location.[157] Full backups capture entire datasets periodically, while incremental and differential approaches copy only changes since the last full or prior backup, respectively, optimizing bandwidth and storage but requiring careful sequencing for restoration.[158] Replication techniques, including synchronous mirroring for zero data loss or asynchronous for cost efficiency, distribute data across nodes or sites, enhancing resilience in distributed architectures.[159] Disaster recovery planning integrates backup with defined metrics: Recovery Point Objective (RPO), the maximum acceptable data loss measured as time elapsed since the last backup, and Recovery Time Objective (RTO), the targeted duration to restore operations post-incident.[160] For mission-critical systems, RPOs often target under 15 minutes via continuous replication, while RTOs aim for hours or less through automated failover to redundant sites.[161] Best practices include regular testing of recovery procedures, automation of backups to prevent oversight, and integration with geographically distributed storage to counter regional outages, as demonstrated in frameworks handling petabyte-scale data across facilities.[162] [163] Compliance-driven retention policies, such as those mandated by regulations like GDPR or HIPAA, further dictate immutable backups to withstand ransomware, with erasure coding aiding efficient long-term archival by minimizing reconstruction times from parity data.[155]Energy Consumption
Trends in power demand
Global data center electricity consumption reached approximately 683 terawatt-hours (TWh) in 2024, representing about 2-3% of worldwide electricity use.[164] This figure has grown steadily, with U.S. data centers alone consuming 4.4% of national electricity in 2023, up from lower shares in prior decades amid expansions in cloud computing and hyperscale facilities. In regions with high concentrations like Virginia, data centers accounted for about 26% of state electricity consumption in 2023, rivaling the total electricity use of smaller states and straining local power grids, which has led to utility rate hikes including a 13% price spike in Virginia.[5][165][166] Load growth for data centers has tripled over the past decade, driven by increasing server densities and computational demands.[6] Projections indicate accelerated demand, primarily fueled by artificial intelligence workloads, where training and inference consume huge amounts of electricity due to their high energy intensity and sustained computational requirements, with global data center power demand expected to grow over 50% in coming years. This is exemplified by the need for high-performance accelerators like GPUs, which elevate power densities per rack from traditional levels of 5-10 kilowatts to 50-100 kilowatts or more.[87] The International Energy Agency forecasts global data center electricity use to more than double to 945 TWh by 2030, growing at 15% annually—over four times the rate of overall electricity demand—equivalent to Japan's current total consumption.[38] Goldman Sachs Research similarly projects a 165% increase in global data center power demand by 2030, with a 50% rise by 2027, attributing this to AI training and inference scaling with larger models and datasets.[9] This surge is exemplified by gigawatt-scale AI facilities, such as xAI's Colossus supercomputer cluster in Memphis, Tennessee, which is expanding to over 1 GW of power capacity to support hundreds of thousands of GPUs for AI training and inference.[79] In the United States, data centers are expected to account for 6.7-12% of total electricity by 2028, with demand potentially doubling overall by 2030 from 2024 levels.[6] Regional spikes are evident, such as in Texas where utility power demand from data centers is projected to reach 9.7 gigawatts (GW) in 2025, up from under 8 GW in 2024, influenced by cryptocurrency mining alongside AI.[167] By 2035, U.S. AI-specific data center demand could hit 123 GW, per Deloitte estimates, straining grid capacity and prompting shifts toward on-site generation and renewable integration.[39] These trends reflect causal drivers like exponential growth in data processing needs, rather than efficiency offsets alone, though improvements in power usage effectiveness (PUE) mitigate some escalation.[87]Efficiency metrics and benchmarks
Power Usage Effectiveness (PUE) serves as the predominant metric for evaluating data center energy efficiency, calculated as the ratio of total facility power consumption to the power utilized solely by information technology (IT) equipment, with a theoretical ideal value of 1.0 indicating no overhead losses.[168] Developed by The Green Grid Association, PUE quantifies overhead from cooling, power distribution, and lighting but excludes IT workload productivity or server utilization rates, limiting its scope to infrastructure efficiency rather than overall operational effectiveness.[169] A complementary metric, Data Center Infrastructure Efficiency (DCiE), expresses the same ratio inversely as a percentage (DCiE = 100 / PUE), where higher values denote better efficiency.[170] Industry benchmarks reveal significant variation by facility type, scale, and age. Hyperscale operators like Google achieved a fleet-wide annual PUE of 1.09 in 2024, reflecting advanced cooling and power systems that reduced overhead energy by 84% compared to the broader industry average of 1.56.[171] Enterprise data centers typically range from 1.5 to 1.8, while newer colocation facilities trend toward 1.3 or lower; overall averages have stabilized around 1.5-1.7 in recent years, with improvements concentrated in larger, modern builds rather than legacy sites.[172][43] Uptime Institute surveys indicate that PUE levels have remained largely flat for five years through 2024, masking gains in hyperscale segments amid rising power demands from AI workloads.[173] Emerging metrics address PUE's limitations by incorporating broader resource factors. The Green Grid's Data Center Resource Effectiveness (DCRE), introduced in 2025, integrates energy, water, and carbon usage into a holistic assessment, enabling comparisons of total environmental impact beyond power alone.[174] Water Usage Effectiveness (WUE), measured in liters per kWh, averages 1.9 across U.S. data centers, highlighting cooling-related demands that PUE overlooks.[8] Carbon Usage Effectiveness (CUE) further benchmarks emissions intensity, with efficient facilities targeting values near 0 by sourcing renewable energy.[175] These expanded indicators underscore that while PUE drives infrastructure optimization, true efficiency requires balancing power, water, and emissions in context of workload density and grid carbon intensity.[176]| Facility Type | Typical PUE Range | Notes |
|---|---|---|
| Hyperscale | 1.09–1.20 | Leaders like Google report 1.09 fleet-wide in 2024.[171][172] |
| Colocation | 1.3–1.5 | Newer facilities approach lower end.[172] |
| Enterprise | 1.5–1.8 | Older sites often higher; averages ~1.6 industry-wide.[43][177] |
Power distribution innovations
Data centers traditionally rely on alternating current (AC) power distribution, which necessitates multiple AC-to-DC and DC-to-AC conversions to power IT equipment, resulting in efficiency losses of up to 10-15% from transformation stages.[178] Innovations in direct current (DC) power distribution address these inefficiencies by reducing conversion steps, enabling higher overall system efficiency—potentially up to 30% gains in end-to-end power delivery—and facilitating denser rack configurations with lower cooling demands due to minimized heat generation from conversions.[179][178] High-voltage DC (HVDC) architectures represent a prominent advancement, distributing power at voltages like 800V to IT loads, which cuts transmission losses compared to low-voltage AC systems and improves voltage stability for high-density AI workloads.[180] NVIDIA's 800V HVDC design, announced in May 2025, exemplifies this shift, optimizing for AI factories by integrating seamlessly with renewable sources and battery storage while reducing cabling weight and space requirements by avoiding bulky transformers.[180] Similarly, Delta Electronics demonstrated HVDC/DC power shelves in October 2025 capable of supporting both legacy AC-48V and native HVDC racks, enhancing scalability for hyperscale facilities where power demands exceed 100 MW per site.[181] Medium-voltage DC distribution directly to the IT space, coupled with solid-state transformers, emerges as another key innovation to handle surging AI-driven loads, projected to double data center electricity demand by 2028, by enabling finer-grained power control and fault isolation without traditional step-down infrastructure.[182][6] These systems leverage semiconductor-based transformation for higher reliability and efficiency, mitigating risks from grid fluctuations in regions with intermittent renewables integration.[183] Adoption remains challenged by the need for standardized components and retrofitting costs, though pilot deployments in 2024-2025 hyperscale projects demonstrate 5-10% reductions in power usage effectiveness (PUE) metrics.[184][182]Environmental Impact
Water usage realities
Data centers primarily consume water for cooling systems, particularly through evaporative cooling towers that dissipate heat by evaporating water, a process essential for maintaining equipment temperatures below failure thresholds in high-density computing environments.[8] This consumptive use—where water is lost to evaporation rather than discharged—accounts for the majority of water withdrawal in water-cooled facilities, distinguishing it from non-consumptive industrial uses.[185] Air-cooled or closed-loop systems exist but are less prevalent in warm climates due to lower efficiency, as evaporative methods achieve higher heat rejection per unit of energy.[186] In the United States, data centers withdrew approximately 17 billion gallons (64 billion liters) of water in 2023, predominantly for cooling, according to estimates from Lawrence Berkeley National Laboratory, with hyperscale operators like Google, Microsoft, and Meta accounting for a significant share.[187] Globally, the International Energy Agency projects data center water consumption could reach 1.2 billion cubic meters (317 billion gallons) annually by 2030, driven by AI workload expansion, equivalent to the annual household water usage of 6 to 10 million people, though this remains a fraction of total sectoral water use dominated by agriculture.[188][189] Per-facility figures vary: a medium-sized data center may use up to 110 million gallons yearly, while large hyperscale sites can exceed 5 million gallons daily, comparable to the annual supply for 10,000–50,000 residents.[8] For instance, Google's Council Bluffs, Iowa facility consumed 1.3 billion gallons of potable water in 2024, or about 3.7 million gallons daily.[190] Water usage intensity is often measured in gallons or liters per megawatt (MW) of IT load: a 1 MW facility using direct evaporative cooling can consume over 25 million liters (6.7 million gallons) annually, scaling to roughly 2 million liters daily for a 100 MW site.[191] [192] These rates are site-specific, influenced by local humidity, temperature, and workload; facilities in arid regions like Arizona or Nevada face amplified stress, as evaporative demands peak during heat waves when municipal supplies are strained, with roughly two-thirds of data centers built or in development since 2022 located in high water-stress areas, potentially exacerbating local shortages in regions like the US Southwest.[193][192] Conversely, northern or coastal sites leverage free air cooling or seawater, minimizing freshwater draw—Equinix reported consuming 60% of its 2023 withdrawals (3,580 megaliters globally) via evaporation, with the rest recycled or discharged.[194] Despite growth, data center water footprints are modest relative to broader economies: U.S. totals equate to less than 0.1% of national freshwater withdrawals, overshadowed by irrigation and manufacturing.[8] Operators mitigate impacts through metrics like Water Usage Effectiveness (WUE), targeting reductions via hybrid cooling, wastewater reuse, or dry coolers; Google averaged 550,000 gallons daily per data center in recent years but has piloted air-cooled designs in water-scarce areas.[191] Projections indicate AI-driven demand could double usage by 2027 without efficiencies, yet causal factors—such as denser chips generating more heat—necessitate cooling innovation over blanket restrictions, as outages from overheating would cascade economic losses far exceeding water costs.[195][196]Noise pollution
Data centers generate significant noise during operation, primarily from cooling systems (fans, chillers, air handling units), server fans, and backup generators. Internal noise levels often reach 90-100 dBA, requiring hearing protection for personnel. Externally, facilities produce a constant low-frequency hum typically at 55-85 dBA from HVAC, with higher levels from generators, which can disturb nearby residents and contribute to community opposition.[197][198]Carbon emissions and mitigation
Data centers primarily generate carbon emissions through electricity consumption for servers, cooling, and ancillary systems, with Scope 1 and 2 emissions dominated by grid-supplied power whose carbon intensity varies by region.[199] In 2024, global data center electricity use reached approximately 415 terawatt-hours, accounting for about 1.5% of worldwide electricity demand, translating to roughly 0.5% of global CO2 emissions when weighted by average grid carbon factors.[200] [201] This footprint, equivalent to 1% of energy-related greenhouse gas emissions including networks, has grown modestly due to efficiency gains offsetting rising demand, but artificial intelligence workloads are projected to drive consumption to double by 2030, potentially elevating emissions to 300-500 million tonnes annually under varying scenarios.[199] [202] Mitigation efforts center on reducing power usage effectiveness (PUE) ratios, which measure total facility energy against IT equipment energy, with leading hyperscale operators achieving averages below 1.1 through advanced cooling like liquid immersion and free air systems.[173] Energy sourcing strategies include power purchase agreements (PPAs) for renewables, direct investments in solar and wind, and site selection in low-carbon grids such as hydroelectric-heavy regions like Quebec or Scandinavia.[203] For instance, major operators like Google and Microsoft report matching over 90% of data center electricity with renewable sources via these mechanisms, though critics argue this offsets rather than directly displaces fossil generation, and Scope 3 supply-chain emissions remain substantial.[204] Actual carbon avoidance depends on grid decarbonization rates; in fossil-reliant areas, on-site natural gas backups and backup diesel generators contribute Scope 1 emissions, with one analysis estimating big tech's reported figures understate in-house data center emissions by up to 7.62 times due to unaccounted flaring and venting.[204] Emerging tactics involve demand flexibility, such as shifting non-critical workloads to off-peak hours or curtailing during high-emission periods, integrated with battery storage to support grid stability while minimizing fossil peaking plants.[205] Innovations like waste heat recovery for district heating and carbon capture at backup generators show promise but face scalability hurdles, as rapid capacity expansion—fueled by AI—often outpaces renewable buildout, necessitating hybrid grids with interim natural gas.[206] Overall, while technical efficiencies have held emissions growth below demand increases since 2020, achieving net-zero requires accelerated grid greening and policy incentives beyond voluntary corporate pledges, as embodied emissions from hardware manufacturing add 20-50% to lifecycle totals.[207]Debunking common myths
A persistent misconception holds that data centers consume electricity on the scale of entire countries, often cited as equivalent to the Netherlands' total usage. In fact, data centers accounted for about 1.5% of global electricity consumption in 2024, a figure projected to double by 2030 primarily due to AI workloads, though this growth is moderated by rapid efficiency improvements in hardware and operations that have reduced power usage effectiveness (PUE) metrics to averages below 1.5 globally.[208] [201] Such country comparisons typically rely on outdated or selective data from the early 2010s, ignoring that data centers' share remains a fraction—around 1-2%—of worldwide electricity, far less than sectors like transportation or residential heating.[209] Another fallacy claims data centers indiscriminately guzzle potable water, depleting local supplies akin to major cities. While hyperscale facilities may use 1-5 million gallons daily for evaporative cooling in some configurations, this often involves non-potable or recycled water, and many operators shift to air-based or dry cooling in water-scarce areas to minimize withdrawal; comprehensive reviews find no instances in the United States where data center operations have impaired community water access or caused shortages.[210] [211] Globally, data centers' water use totals an estimated 1-2 billion gallons per day, negligible compared to agriculture's 70% share of freshwater withdrawals, with innovations like closed-loop systems further reducing net consumption.[212] Claims that data centers' cooling systems waste the majority of their power are also overstated. Modern facilities achieve PUE ratios as low as 1.1 through liquid immersion, free air cooling, and AI-optimized airflow, meaning overheads like cooling represent under 10% of total energy in efficient setups, a stark improvement from pre-2010 averages exceeding 50%.[213] This efficiency counters narratives of inherent waste, as causal analysis shows compute demand drives innovation that lowers per-task energy needs, decoupling raw power growth from output.[214] It is erroneously asserted that data centers' carbon emissions will scale linearly with AI expansion, overwhelming mitigation efforts. Empirical data indicates that while electricity demand rises, carbon intensity declines via renewable integration—many operators match 100% of usage with clean sources—and efficiency gains prevent proportional footprint growth; data centers currently contribute about 0.5% of global CO2 from electricity, enabling broader dematerialization effects like reduced physical shipping that offset far more emissions elsewhere.[201] [214] Assertions of uncontrollable emissions often stem from models assuming static technology, disregarding historical trends where compute efficiency doubled every 2.5 years, akin to Moore's Law extensions.[215]Sustainability practices and trade-offs
Data centers implement sustainability practices aimed at reducing energy intensity and resource consumption, such as procuring renewable energy and optimizing power usage effectiveness (PUE). Operators like Google prioritize carbon-free energy matching for 24/7 operations, achieving average PUE values below 1.1 in advanced facilities through advanced cooling and server efficiencies.[216] Similarly, Meta focuses on hyperscale designs that integrate clean energy procurement, targeting net-zero emissions by 2030 via efficiency gains and renewable power purchase agreements.[217] However, industry-wide renewable adoption remains partial, with estimates indicating that only about 25% of U.S. data center electricity derives from directly procured renewables as of 2024, constrained by grid limitations and intermittency.[218] Cooling represents a core area of innovation, with liquid-based systems like immersion and cold-plate technologies reducing overall energy consumption by 15-20% and greenhouse gas emissions by up to 21% relative to air-cooled alternatives, as demonstrated in Microsoft evaluations.[219] Waste heat recovery further enhances sustainability by repurposing exhaust thermal energy for district heating; for instance, Facebook's Odense facility in Denmark recovers up to 100,000 MWh annually to supply urban hot water networks.[220] In Nordic regions, data centers in Finland and Sweden integrate with district systems to offset fossil fuel heating, capturing low-grade heat from IT equipment that would otherwise dissipate.[221] These practices have proliferated, with heat export projects like Equinix's collaborations enabling reuse in adjacent infrastructure.[222] Trade-offs inherent to these practices limit universal adoption and effectiveness. Renewable integration demands backup generation or storage to ensure uptime, as solar and wind variability can necessitate fossil fuel peakers, potentially offsetting emission reductions during peak loads; this reliability-energy nexus has slowed sustainability advances amid AI-driven demand surges in 2025.[223][224] Cooling choices exemplify resource conflicts: liquid systems, while more energy-efficient, elevate water demands through evaporative processes or direct usage, with water-cooled centers consuming about 10% less power but straining local supplies in arid regions, unlike air cooling's higher electricity footprint.[185][225] Hybrid approaches mitigate this by alternating methods, yet require site-specific engineering that increases capital costs by 20-30% upfront.[226] Heat recovery, though beneficial, confines facilities to proximate demand centers like urban districts, curtailing scalability in remote or hyperscale deployments where transmission losses erode viability.[227] Overall, these tensions—balancing efficiency, resilience, and localization—underscore that sustainability gains often yield marginal net benefits against exponential compute growth, with AI workloads projecting 44 GW additional U.S. demand by 2030.[228]Economic Role
Industry growth and major operators
The data center industry has expanded rapidly, propelled by the adoption of cloud computing and the computational demands of artificial intelligence applications. Global revenue in the data center market is projected to reach US$527.46 billion in 2025, driven by increasing data generation and processing needs.[229] Market analyses forecast a compound annual growth rate (CAGR) of approximately 11.2% from 2025 to 2030, with the sector valued at USD 347.60 billion in 2024 and expected to surpass USD 652 billion by 2030.[230] According to Dell'Oro Group reports, global data center capital expenditure (CapEx) reached $260 billion in 2023 (up 4% year-over-year), surged to $455 billion in 2024 (up 51% year-over-year) driven primarily by AI-optimized accelerated servers and hyperscaler investments, and is projected to rise more than 30% in 2025 due to sustained AI infrastructure demand and recovery in general-purpose infrastructure.[231][232] This growth manifests in physical capacity additions, including an estimated 10 gigawatts of hyperscale and colocation facilities projected to break ground worldwide in 2025, alongside 7 gigawatts reaching completion. A 150–250 MW regional facility aligns broadly with major operators' growth, representing a meaningful portion of their forward order books for hyperscale/AI-driven developments in the late 2020s.[33][35] Artificial intelligence represents a primary catalyst, with demand for AI-ready data center capacity anticipated to grow at 33% annually from 2023 to 2030 under midrange scenarios, necessitating vast expansions in high-density computing infrastructure.[33] Concurrently, overall power consumption from data centers is expected to increase by 165% by the end of the decade, reflecting the energy-intensive nature of AI training and inference workloads integrated with cloud services.[9] Hyperscale operators have accelerated this trend, shifting global capacity toward their facilities, which are projected to comprise 61% of total data center capacity by 2030, compared to 22% for on-premise enterprise setups.[233] Regional concentrations of capacity are evident in Europe, where primary hubs for premium data centers include the FLAP-D markets—Frankfurt, London, Amsterdam, Paris, and Dublin—with Zurich as another notable location due to high connectivity and security.[234] Debt financing supports these AI-driven expansions, with operators justifying longer GPU depreciation schedules of around six years based on sustained demand enabling high utilization rates and premium pricing. GPUs transition to inference workloads with useful lives extending 6–8 years, real-world data indicating minimal value drop due to ongoing profitability of older hardware.[235][236][237] This economic viability, exemplified by providers like CoreWeave applying six-year cycles since 2023, facilitates capital access amid market momentum despite rapid innovation cycles.[235] Leading operators include hyperscalers such as Amazon Web Services (AWS), Microsoft Azure, and Google Cloud, which dominate through proprietary builds optimized for their cloud platforms and AI services, collectively holding significant market influence in capacity deployment.[238] Hyperscale data centers currently account for about 35% of the overall market share, underscoring their role in scaling for large-scale tenants.[239] Supporting this expansion are key infrastructure providers such as Arista Networks (ANET), which supplies high-performance networking switches for AI workloads and cloud connectivity; Vertiv (VRT), specializing in critical power, cooling including liquid cooling for high-density AI servers, and thermal management; and Eaton (ETN), focusing on electrical power distribution, uninterruptible power supply (UPS) systems, modular solutions, and grid support for AI facilities, with data center orders growing approximately 70% year-over-year in Q3 2025.[240][241][242] These providers benefit substantially from the AI data center buildout, with VRT and ANET frequently highlighted as leading AI infrastructure stocks in 2025.[243] In the colocation segment, providers like Equinix and Digital Realty manage extensive networks of multi-tenant facilities, offering interconnection and power redundancy to enterprises, with power costs often passed through to tenants separately via reimbursement or direct billing, sometimes including a small markup that contributes to operator revenue. Revenue is primarily derived from leasing power capacity, where annual revenue per MW = lease rate per kW/month × 1,000 × 12 months. Equinix operating over 250 data centers across multiple continents as of 2025.[244][245][246][247] These operators compete and collaborate amid tightening supply, as evidenced by declining global vacancy rates to 6.6% in early 2025.[248]Contributions to employment and GDP
The data center industry in the United States contributed $727 billion to gross domestic product (GDP) in 2023, representing a 105% increase from $355 billion in 2017, encompassing direct operations, indirect supply chain effects, and induced spending.[249] This figure stems from a PwC analysis commissioned by industry groups, highlighting the sector's role in value added across information processing, construction, and supporting services.[250] Investment in data centers and related information processing equipment drove 92% of U.S. GDP growth in the first half of 2025, despite comprising only about 4% of total GDP, according to economic analyses attributing surges to hyperscaler capital expenditures nearing $400 billion annually.[251][252] Employment impacts are amplified by multipliers, with each direct data center job generating approximately six indirect or induced positions in construction, maintenance, logistics, and local services, per a PwC assessment of nationwide effects.[253] Nationwide data-center-related employment reached 3.5 million jobs by 2021, a 20% rise from 2.9 million in 2017, outpacing the 2% growth in overall U.S. employment during the period, as tracked by real estate and economic data.[254] Direct employment in data processing, hosting, and related services (NAICS 518210) grew over 60% from 2016 to 2023, though concentrated in hubs like Northern Virginia and uneven across regions, with limited expansion in rural or non-primary markets.[255] Labor income from the sector increased 74% directly and 40% in total impact between 2017 and 2021, reflecting high-wage roles in engineering, operations, and IT.[256] The rapid expansion has intensified demand for specialized skills, particularly qualified electrical engineers to design, install, and maintain the complex power systems essential for reliable operations. Operators report recruitment challenges amid this boom, leading technical colleges and universities to introduce new programs in power engineering, data center operations, and related fields. For example, Texas State Technical College launched short-term data center technician courses in 2025 to address needs in high-growth areas like Dallas.[257][258] Projections indicate further job creation from expansion, with new data center construction potentially adding nearly 500,000 positions, $40 billion in labor income, and $140 billion to GDP through direct, indirect, and induced channels, based on modeling of planned builds as of October 2025.[259] Globally, data center effects are less quantified but follow similar patterns in major markets like Europe, where the sector supports digital infrastructure integral to broader GDP contributions from ICT, though U.S. dominance in hyperscale facilities accounts for the largest share of documented impacts.[28] These contributions arise causally from demand for cloud computing, AI workloads, and digital services, driving capital-intensive builds that sustain long-term economic multipliers despite operational automation limiting per-facility headcounts.Local infrastructure effects
Data centers exert considerable pressure on local electrical grids due to their high power consumption, frequently requiring upgrades to transmission and distribution infrastructure to avoid capacity shortfalls. In Northern Virginia, which hosts the largest data center market globally with approximately 13% of worldwide operational capacity as of 2024, the rapid expansion has led to projected reliability risks, including potential blackouts totaling hundreds of hours annually without further enhancements.[260][261] For instance, utility provider Dominion Energy sought approval in 2023 to recover $63.1 million for transmission upgrades specifically driven by data center growth in the region.[262] Neighboring states have also borne costs; Maryland utility customers faced an estimated $800 million in transmission investments by mid-2025 to support Virginia's data centers via regional grid interconnections.[263] These facilities often fund or trigger infrastructure expansions, including new high-voltage lines and substations, as operators commit to connecting under utility tariffs that allocate upgrade costs. A 2025 approval by Virginia regulators for an eight-tower, 230-kilovolt transmission project costing millions directly served a single 176-megawatt hyperscale data center, illustrating how individual sites can necessitate dedicated grid reinforcements.[264] However, such developments can elevate local electricity rates; in areas like West Virginia, data center loads on the regional PJM grid contributed to higher wholesale prices passed to residential users as of October 2025.[265] In New York, state inquiries in October 2025 highlighted data center-driven demand as a factor in rising utility bills, with assembly hearings examining grid strain from AI-related facilities.[266] Beyond power, data center construction and operations impact transportation networks through increased heavy vehicle traffic for materials and equipment delivery. Projects typically require road widening, bridge reinforcements, and temporary access improvements to accommodate oversized loads, as seen in multiple U.S. developments where local governments mandate infrastructure mitigations prior to permitting.[267][268] In rural or small-town settings, such as proposed sites in Virginia's Culpeper County, construction phases have raised concerns over congestion and wear on existing roadways, prompting community opposition and regulatory delays in at least 20% of announced projects nationwide by late 2024.[269][270] These effects are compounded by the need for reliable fiber optic and water lines, though operators frequently invest in parallel utility extensions, yielding long-term enhancements to local connectivity and capacity.[271] Overall, while straining existing systems, data center proximity correlates with accelerated infrastructure modernization, albeit at the expense of short-term disruptions and fiscal burdens on ratepayers.[28]Debates on subsidies and fiscal impacts
Numerous jurisdictions have implemented tax incentives, including sales tax exemptions on equipment purchases and property tax abatements, to attract data center investments, with over 30 U.S. states offering such programs as of 2025.[272] Proponents argue these subsidies generate substantial economic benefits, such as job creation and capital investment, which outweigh initial revenue forgone; for instance, Virginia's data center sales tax exemption, enacted in 2015 and expanded thereafter, has supported an industry contributing an estimated 74,000 jobs, $5.5 billion in annual labor income, and $9.1 billion to state GDP, according to a 2024 legislative analysis.[260] Industry-commissioned studies, like a 2025 PwC report for the Data Center Coalition, quantify broader multipliers, including indirect employment in construction and services, positioning data centers as net fiscal contributors over their lifecycle through eventual property tax payments post-abatement periods.[256] Critics contend that these incentives represent a zero-sum "race to the bottom" among states, forfeiting hundreds of millions in potential revenue without commensurate public returns, as evidenced by a 2025 CNBC analysis of state-level exemptions.[273] At least 10 states forgo over $100 million annually in sales tax revenue from data centers, per Good Jobs First estimates, often with minimal job creation—typically 50-100 operational positions per facility, far fewer than promised relative to multi-billion-dollar investments.[274] In Wisconsin, a 2025 sales tax exemption projected to cost $200 million over a decade has drawn opposition for subsidizing hyperscalers like Microsoft without guaranteed long-term local benefits or clawback mechanisms for unmet commitments.[275] Such policies, critics argue, distort market-driven location choices, favoring tax havens over efficient sites and straining public budgets amid rising AI-driven demand. Fiscal impacts extend beyond direct taxes to indirect costs, including subsidized utility expansions that elevate rates for residents; a 2025 University of Michigan study found data centers impose disproportionate energy burdens on lower-income households, with Michigan facilities alone projected to increase statewide electricity demand by 8-10% by 2030, potentially adding $1-2 monthly to average bills.[276] While data centers generate billions in aggregate tax revenue—estimated at $10-15 billion nationally in 2024 from property and other levies—the debate hinges on net effects post-incentives, with some analyses questioning whether contributions fully offset exemptions and infrastructure outlays.[277] Reforms proposed include performance-based clawbacks, transparency in subsidy awards, and tying incentives to verifiable metrics like renewable energy integration, though empirical evidence on long-term fiscal neutrality remains mixed, varying by jurisdiction-specific abatement durations and enforcement.[278]Emerging Technologies
Modular and edge computing facilities
Modular data centers consist of prefabricated, standardized components assembled off-site and transported for rapid on-site deployment, enabling scalability through incremental additions of modules housing IT equipment, power, cooling, and networking systems.[279] These facilities emerged in the early 2000s as responses to demands for faster construction timelines compared to traditional builds, which can take 18-24 months, versus modular's 3-6 months for initial modules.[280] By integrating self-contained units, such as shipping container-based designs, they reduce upfront capital expenditure by up to 30% and minimize construction waste through factory-controlled assembly.[71] Edge computing facilities extend this modularity to distributed locations proximate to data generation sources, processing information locally to achieve latencies under 10 milliseconds, essential for applications like autonomous vehicles and industrial IoT.[281] Unlike centralized hyperscale centers, edge sites are smaller-scale—often 1-10 racks—and leverage modular designs for deployment in urban micro-hubs, rural areas, or temporary setups, supporting 5G networks where base stations require integrated compute.[282] The convergence of modular and edge architectures facilitates hybrid models, where core data centers orchestrate edge nodes, optimizing bandwidth by filtering only aggregated insights for central transmission, thereby cutting network traffic by 50-80% in high-volume scenarios.[283] Key technologies in these facilities include integrated liquid cooling for high-density racks exceeding 50 kW, advanced fire suppression like FM-200 agents that avoid residue damage to electronics, and prefabricated power distribution units with battery backups for uptime.[284] For edge-specific implementations, micro data centers incorporate AI-optimized orchestration software to dynamically allocate resources across nodes, enhancing fault tolerance in remote environments.[285] Challenges persist, including heightened security risks from dispersed footprints necessitating zero-trust models, and power constraints in non-grid areas, often addressed via on-site solar or generators, though scalability limits arise beyond 100 modules without custom engineering.[286] Market growth underscores adoption: the modular data center sector reached USD 29.04 billion in 2024, projected to expand at 17% CAGR to USD 75.77 billion by 2030, driven by AI workloads demanding quick provisioning.[287] Concurrently, edge data centers are forecasted to grow from USD 50.86 billion in 2025 to USD 109.20 billion by 2030 at 16.5% CAGR, fueled by IoT proliferation exceeding 75 billion devices by 2025.[288] Notable deployments include hyperscalers like Google utilizing modular pods for edge inference in telecom towers and enterprises deploying containerized units for disaster recovery, as seen in IBM's portable solutions operational since the 2010s.[289] These facilities trade centralized efficiency for distributed resilience, with empirical data showing 20-40% faster time-to-market but requiring robust supply chain verification to mitigate prefabrication defects.[290]Advanced cooling for high-density AI
Data centers optimized for AI workloads require high-density power often reaching 40-120 kW per rack to support massive GPU clusters, advanced liquid cooling systems, and scalable infrastructure for interconnecting thousands of accelerators in training and inference operations.[291] High-density AI workloads, driven by GPU clusters for training and inference, generate extreme heat loads, with rack power densities often exceeding 100 kW—far beyond the 15-20 kW limits of traditional air cooling systems.[292][100] Gigawatt-scale AI data centers amplify these challenges, as seen in xAI's Colossus supercluster in Memphis, which began operations in 2024 with over 100,000 GPUs and is expanding toward a 1-2 GW total power draw, necessitating grid-scale power solutions and advanced cooling to manage facility-wide thermal loads amid interconnection delays and supply constraints.[293][294] This necessitates advanced liquid-based cooling to maintain component temperatures below thermal throttling thresholds, prevent hardware failures, and sustain computational performance. Liquid cooling exploits the superior thermal conductivity of fluids like water or dielectric oils, which transfer heat orders of magnitude more effectively than air, enabling denser server deployments and lower overall energy consumption for cooling.[101][295] Direct-to-chip liquid cooling (DLC) delivers coolant via microchannels directly to high-heat components such as CPUs and GPUs, supporting densities up to 200 kW per rack while minimizing retrofitting needs in existing facilities.[100] Rear-door heat exchangers (RDHx) integrate liquid loops at the rack exhaust to capture hot air efficiently, often hybridized with air assist for transitional densities around 50-100 kW.[296] Immersion cooling submerges entire servers in non-conductive dielectric fluids, either single-phase (natural convection) or two-phase (boiling for phase-change heat absorption), achieving power usage effectiveness (PUE) values as low as 1.03-1.1 by eliminating fans and enabling heat reuse for applications like district heating.[297][298] In AI contexts, immersion has demonstrated up to 64% energy savings in cooling, particularly in humid or variable climates, though deployment requires fluid compatibility testing to avoid corrosion or leakage risks.[299] Hybrid systems combining liquid and air elements, augmented by AI-driven predictive controls, adapt to fluctuating AI workloads—such as bursty inference spikes—optimizing coolant flow and fan speeds in real-time to cut operational costs by 20-30% over static methods.[300][299] Major operators like hyperscalers are scaling these technologies; for instance, facilities supporting NVIDIA's high-end GPUs increasingly mandate DLC or immersion to handle 60-100 kW racks without excessive water use, contrasting with air-cooled baselines that consume 1-2 liters of water per kWh via evaporative towers.[301][8] While promising, challenges include higher upfront costs (2-3x air systems) and supply chain dependencies on specialized manifolds and pumps, though long-term efficiency gains—evidenced by PUE reductions—justify adoption for sustainable AI scaling.[302][303]Novel deployment concepts
One prominent experimental approach involves submerging data centers underwater to leverage natural ocean cooling and reduce land use. Microsoft's Project Natick initiative deployed a sealed, nitrogen-filled pod containing 12 server racks off the coast of Scotland in 2018, which operated autonomously for over two years until retrieval in 2020; failure rates were one-eighth those of terrestrial counterparts, attributed to the absence of human interference and stable temperatures around 4°C.[304] Phase 2 scaled to 864 servers in a larger pod off California's coast in 2020, demonstrating faster deployment (under 90 days) and economic viability in manufacturing, but the project was discontinued in 2024 due to logistical challenges in scaling maintenance and retrieval, rendering it impractical for widespread adoption despite environmental benefits like lower carbon footprints from reduced construction.[305] [306] In contrast, China operationalized a commercial underwater data center in Hainan by October 2025, utilizing seawater for cooling and integrating it into national infrastructure strategies, though independent verification of long-term reliability remains limited.[307] Floating data centers on barges or vessels represent another innovative strategy to bypass terrestrial land constraints and tap coastal power grids or renewable sources. Nautilus Data Technologies commissioned the 7 MW Stockton1 facility on a barge at the Port of Stockton, California, in 2021, employing seawater for cooling and achieving operational status within months, with expansions planned for additional port sites leveraging existing fiber connectivity.[308] Karpowership's Kinetics unit announced in July 2025 plans for barge-based facilities in shipyards, targeting AI workloads by avoiding lengthy land permitting while using onboard or port-supplied power, potentially deployable in under a year.[309] These designs offer mobility for relocation to optimal sites but face risks from marine weather, corrosion, and regulatory hurdles in international waters, with real-world uptime data still emerging from pilot scales.[310] Orbital data centers in space have been proposed to exploit continuous solar power and vacuum radiative cooling, potentially slashing energy costs by up to 90% compared to Earth-based systems through uninterrupted sunlight exposure.[311] Jeff Bezos endorsed the concept in October 2025, citing orbital facilities as a solution to terrestrial resource strains from AI-driven demand, while startups like Starcloud project deployments using satellite constellations for processing space-generated data or low-latency Earth links.[312] However, fundamental challenges persist: space's vacuum hinders convective heat dissipation, requiring advanced radiative systems; cosmic radiation accelerates hardware degradation; launch costs exceed $10,000 per kg; and communication latency (minimum 120 ms round-trip to geostationary orbit) limits viability for real-time applications, confining prospects to niche uses like astronomical data processing rather than general-purpose computing.[313] No operational orbital data centers exist as of 2025, with experts questioning scalability due to these physics-based barriers outweighing theoretical efficiencies.[314] Underground deployments in repurposed mines, bunkers, or excavated sites capitalize on geothermal stability for passive cooling and enhanced physical security against attacks or disasters. Facilities like Bluebird Fiber's data center, buried 85 feet (26 meters) underground, benefit from natural insulation reducing HVAC needs by up to 40% and protection from surface threats, with construction leveraging existing subsurface infrastructure for faster rollout.[315] Converted Cold War-era bunkers in Europe and the U.S., such as those operated by Cyberfort, provide bomb-proof enclosures for cloud storage, minimizing electromagnetic interference and enabling heat reuse via adjacent geothermal systems.[316] Drawbacks include higher initial excavation costs, limited scalability for high-density racks due to access constraints, and vulnerability to flooding or seismic events, though empirical data from operational sites confirm energy savings of 20-30% over above-ground equivalents in temperate climates.[317] These concepts collectively address densification pressures from AI but hinge on site-specific economics, with adoption tempered by unproven long-term resilience at hyperscale.Integration with alternative energy sources
Data centers have increasingly pursued integration with renewable energy sources to address high electricity demands and reduce reliance on fossil fuels, driven by corporate sustainability targets and regulatory pressures. Hyperscale operators such as Google, Microsoft, and Amazon have committed to achieving 100% renewable energy matching, often through power purchase agreements (PPAs) and renewable energy certificates (RECs), though actual grid-supplied power frequently includes fossil fuel components despite these offsets. For instance, Google announced in December 2024 a $20 billion investment plan to develop colocated renewable energy and storage assets alongside data centers by 2030, aiming for 24/7 carbon-free energy supply to mitigate intermittency issues.[318] Similarly, Microsoft has pursued direct integrations, including nuclear small modular reactors (SMRs) offering high energy density from compact footprints, continuous 24/7 baseload power exceeding 90% capacity factor without intermittency risks, minimal land use compared to large-scale renewables, and reduced transmission losses via on-site or co-located deployment, as announced in partnership deals in 2024-2025 to power AI workloads reliably.[319][320] Grid approval bottlenecks and interconnection delays, often spanning 2-5 years, have made on-site nuclear reactors and battery energy storage systems necessary for direct, distributed energy connections, bypassing utility queues.[321] On-site and nearby renewable installations include solar photovoltaic arrays and wind turbines, supplemented by battery energy storage systems (BESS) to handle variable output. A 2023 analysis highlighted data centers in regions with abundant hydro resources, such as the Pacific Northwest, achieving up to 90% renewable sourcing via hydroelectric dams, reducing carbon intensity compared to coal-dependent grids. Amazon Web Services (AWS) expanded solar integrations in 2023-2024, deploying over 500 MW of on-site or adjacent solar capacity across U.S. facilities to offset peak loads, though full operational matching remains limited by transmission constraints. Geothermal and biomass co-generation have seen pilot implementations in Iceland and Nordic sites, leveraging natural heat for both power and cooling, with facilities reporting power usage effectiveness (PUE) improvements to below 1.1.[322] Despite progress, integration faces causal challenges from the intermittent nature of solar and wind, which cannot reliably provide the continuous, high-density power data centers require for uptime exceeding 99.999%. Studies indicate that without sufficient storage or hybrid systems, renewables alone lead to curtailment risks and higher costs, with one 2024 review estimating that U.S. data centers' projected 100 GW demand by 2030 exceeds scalable intermittent capacity without nuclear or gas backups. Critics note that REC-based claims often overstate direct impact, as evidenced by a September 2024 report finding hyperscaler emissions 662% higher than self-reported due to unaccounted grid emissions and Scope 3 supply chain effects.[204][323] Hybrid approaches, combining renewables with nuclear or hydrogen fuel cells, emerge as pragmatic solutions for causal reliability, as pure intermittent reliance risks operational failures during low-generation periods.[324][325]Regulations and Challenges
Certification standards
Data center certification standards evaluate infrastructure reliability, operational resilience, security, and environmental sustainability, often serving as benchmarks for regulatory compliance and customer assurance. These standards typically involve third-party audits and can apply to design, construction, or ongoing operations phases.[140] The Uptime Institute's Tier Classification System, established over 30 years ago, defines four levels of data center performance based on redundancy, fault tolerance, and maintainability. Tier I provides basic non-redundant capacity suitable for low-criticality operations, while Tier II adds redundant components for partial fault tolerance; Tier III enables concurrent maintainability without downtime for planned activities, and Tier IV offers full fault tolerance against multiple failures. Certifications are issued separately for topology (design and construction) and operational sustainability, with over 2,000 facilities certified globally as of 2023, though operational ratings remain rarer due to rigorous requirements.[138][326][327] Information security certifications, such as ISO/IEC 27001:2022, outline requirements for an information security management system (ISMS) to protect data confidentiality, integrity, and availability in data centers handling sensitive workloads. Compliance demands risk assessments, implementation of 93 controls across 14 domains (including physical security and access controls), and annual surveillance audits by accredited bodies, with data centers often extending scope to cover physical infrastructure like cooling and power systems. As of 2024, ISO 27001 adoption in data centers mitigates cyber risks but does not guarantee zero vulnerabilities, as evidenced by ongoing breaches in certified facilities.[328][329][330] Energy efficiency and sustainability standards address the sector's high power consumption, which exceeded 200 terawatt-hours globally in 2022. LEED BD+C: Data Centers, tailored for hyperscale facilities, awards points for metrics like power usage effectiveness (PUE) below 1.5, renewable energy integration, and water-efficient cooling, with certification levels (Certified, Silver, Gold, Platinum) based on total credits earned through verified performance data. Similarly, ISO 50001 certifies energy management systems for continuous improvement in metrics such as PUE and carbon intensity. These standards promote verifiable reductions—LEED-certified centers have demonstrated up to 25% lower energy use—but face criticism for overlooking lifecycle emissions from hardware sourcing.[331][332][333] Sector-specific compliance certifications include SOC 2 Type II for trust services criteria (security, availability, processing integrity, confidentiality, privacy), audited over 6-12 months to validate controls for cloud and colocation providers, and PCI DSS for facilities processing payment data, requiring quarterly vulnerability scans and annual assessments. HIPAA and GDPR alignments often necessitate these alongside ISO standards for regulated industries. While certifications signal adherence, discrepancies between design intent and operational reality—such as Tier III facilities experiencing outages due to human error—underscore the need for independent verification beyond initial awards.[334][335]Grid and supply chain constraints
Data centers' escalating electricity demands, driven primarily by artificial intelligence workloads—particularly training of large models requiring sustained high power and inference for continuous real-time processing—have imposed significant strains on electrical grids worldwide, with AI expected to drive over 50% growth in data center power demand by 2027.[336] These demands arise due to transmission bottlenecks that limit power delivery to high-demand areas, challenges posed by intermittent renewable energy sources in fulfilling consistent baseload demands of AI operations, and delays in utility infrastructure upgrades and permitting that prevent supply from matching surging needs; power availability has become the primary constraint for expansion, exacerbated by grid capacity shortages, lengthy infrastructure approvals, construction delays, and supply chain pressures on generators, transformers, and cooling systems, with projections indicating that global data center power consumption could reach 20% of total electricity use by 2030-2035.[337][338][339] In the United States, data centers consumed 2.2 gigawatts (GW) of power capacity in the first half of 2025 alone, concentrated in key regions like Northern Virginia, exacerbating local grid limitations and leading to multi-year backlogs for interconnection approvals.[340] Utility providers reported spending $178 billion on grid upgrades in 2024, with forecasts for $1.1 trillion in capital investments through 2029 to accommodate surging demand, yet 92% of data center operators identify grid constraints as a major barrier to expansion.[341][342] Interconnection queues have lengthened due to the rapid scaling of hyperscale facilities, with over 100 GW of data center capacity slated to come online between 2024 and subsequent years, often clashing with aging infrastructure and regulatory hurdles.[91] In regions like the PJM Interconnection, proposed data centers are the primary driver of recent electricity bill increases for residential customers, as grid operators prioritize reliability amid load spikes that could double data centers' share of U.S. electricity by 2035.[343][344] A 2025 survey found 44% of data center firms facing utility wait times exceeding four years, compounded by geographic concentrations that amplify localized strains and delay project timelines.[342] Supply chain bottlenecks further hinder data center deployment, particularly for critical grid components like power and distribution transformers, where U.S. shortages are projected to reach 30% for power transformers and 10% for distribution units by 2025 due to manufacturing constraints and raw material limitations.[345] The surge in data center builds has driven transformer delivery wait times to years, inflating costs and stemming from policy-induced shifts, such as subsidies favoring renewables that disrupt traditional supply chains reliant on specialized steel and insulation.[346][347] Additional shortages affect switchgear, gas turbines, and cabling, with global disruptions from outdated production practices and weather events exacerbating delays for facilities requiring high-voltage equipment to handle megawatt-scale loads.[91][348] These constraints have prompted some operators to explore on-site generation or modular solutions, though scalability remains limited by the same upstream bottlenecks.[349]Public opposition and project hurdles
Public opposition to data center developments has surged globally, driven by concerns over resource consumption, environmental disruption, and quality-of-life impacts, resulting in $64 billion worth of U.S. projects blocked or delayed since 2023.[269] Local activism, involving 142 groups across 24 states, has transcended partisan lines, with 55% of opposing public officials identified as Republicans and 45% as Democrats.[350][351] Common grievances include massive electricity demands—often equivalent to those of mid-sized cities—that overload grids and raise utility rates, alongside water-intensive cooling systems exacerbating scarcity in drought-prone areas, incessant noise from fans and generators, and the industrialization of rural or residential landscapes.[352][353] In the United States, NIMBY-style resistance has manifested in protests, moratoriums, and legal challenges. Virginia's Loudoun and Prince William counties, hubs for data center growth, have seen resident-led campaigns against noise pollution and farmland loss, with yard signs in Chesapeake declaring "NO DATA" amid fears of infrastructure strain.[354] In Prince George's County, Maryland, demonstrations prompted County Executive Aisha Braveboy to suspend data center permitting on September 18, 2025, citing inadequate community input.[355] Microsoft abandoned a facility in Racine County, Wisconsin, after sustained local pushback over energy and economic costs, while in Franklin Township, Indiana, over 100 protesters rallied against a Google campus on September 8, 2025, highlighting water depletion risks in already stressed aquifers.[356][357] Bastrop, Texas, residents organized to stall projects amid grid reliability worries, and a Michigan township faced lawsuits from developers after rejecting a site due to projected hikes in power bills and water use.[358][353] A community group filed suit on October 20, 2025, to block a $165 billion OpenAI complex in rural New Mexico, alleging flawed environmental reviews.[359] Opposition continued into 2026, with big tech's data center expansions facing stiff community resistance, including surging demands for moratoriums and protests in Trenton, Ohio, where residents packed city council meetings against a $7.7 million land sale to Prologis for a proposed data center on 141 acres, citing concerns over electricity costs, grid strain, and local infrastructure burdens.[360][361] Internationally, similar hurdles have emerged. Ireland, once a data center magnet, experienced a policy reversal by 2025, with capacity caps imposed after centers consumed 18% of national electricity despite representing under 1% of GDP contribution, sparking protests over emissions and grid failures.[362] In the Netherlands, public outcry over energy imports and heat waste led to a 2024 moratorium on new builds in Amsterdam, extended amid lawsuits from residents.[363] These cases illustrate project delays averaging 12-24 months, escalated costs from redesigns or relocations, and occasional outright cancellations, as developers navigate zoning battles, environmental impact assessments, and ballot initiatives that prioritize local burdens over broader technological imperatives.[364][365]Policy incentives versus regulatory burdens
Governments worldwide have implemented policy incentives to attract data center investments, primarily through tax abatements, sales tax exemptions on equipment and energy, and expedited permitting processes, aiming to stimulate economic growth, job creation, and technological infrastructure development. In the United States, 36 states authorize such tax incentives, often tailored to large-scale projects meeting investment thresholds, such as Georgia's up to 30-year property tax abatements for facilities investing at least $400 million and creating 20 jobs with average salaries exceeding $40,000. Similarly, 42 states offer full or partial sales tax exemptions for data center construction and operations, with Illinois providing approximately $370 million in exemptions covering equipment and electricity costs as of 2025. Federally, executive actions in July 2025 directed agencies to accelerate permitting for data centers and associated high-voltage transmission lines, prioritizing reductions in regulatory delays to support AI infrastructure expansion. These incentives are justified by proponents as essential for competitiveness in a global market dominated by hyperscale operators, potentially generating billions in capital investment and thousands of construction and operational jobs per facility. Despite these incentives, data centers face substantial regulatory burdens stemming from their intensive resource demands, including electricity consumption equivalent to over 4% of total U.S. usage in 2024, with 56% derived from fossil fuels, alongside significant water usage for cooling and potential contributions to grid strain. Environmental regulations, such as emissions reporting under frameworks like California's SB 253 and the EU's Corporate Sustainability Reporting Directive, mandate disclosure of Scope 1, 2, and 3 greenhouse gases, imposing compliance costs and scrutiny on operators. State-level measures, including New York's 2025 legislation requiring annual energy consumption disclosures and prohibiting incentives tied to fossil fuel power purchase agreements, exemplify efforts to align data centers with climate goals, though critics argue these add layers of bureaucratic oversight that delay projects by months or years. Permitting challenges, including federal environmental reviews and local zoning restrictions on land use and noise, further exacerbate interconnection queues to the grid, with U.S. Energy Secretary directives in October 2025 urging regulators to streamline approvals amid surging demand. The tension between incentives and burdens manifests in policy debates where fiscal benefits—such as increased property tax bases post-exemption periods—are weighed against long-term externalities like elevated energy rates for consumers and infrastructure overloads. Some analyses highlight that uncapped tax exemptions can erode state revenues without proportional local benefits, as data centers often import specialized labor and yield limited ongoing employment relative to upfront subsidies. In response, states like Virginia and Ohio have faced legislative pushes in 2024-2025 to pause or reform incentives, conditioning them on efficiency standards or renewable energy commitments to mitigate environmental impacts. Internationally, regulatory hurdles in regions like the EU, encompassing electricity grid access, water abstraction limits, and urban planning consents, have prompted moratoriums on new builds in energy-constrained areas such as Ireland and the Netherlands, contrasting with U.S. pro-development stances but underscoring a broader causal trade-off: incentives accelerate deployment at the risk of unaddressed resource depletion, while stringent regulations safeguard sustainability yet risk ceding economic advantages to less-regulated jurisdictions. Empirical evidence from state experiences suggests that balanced approaches, such as performance-based incentives tied to low-emission operations, may optimize outcomes by internalizing externalities without stifling innovation.References
- ./assets/Google_Data_Center%252C_The_Dalles.jpg