Top 4 AI Predictions

Whitepaper
Premium Content

Structure Research and Iron Mountain Data Centers discuss the top predictions of AI impacts on global data infrastructure by 2030

Iron Mountain Data Centers top 4 AI predictions

Exclusive Preview

AI impacts on global data infrastructure by 2030

In the three years since ChatGPT launched natural language processing (NLP), agentic and generative AI have become ubiquitous, driven by successive generations of Large Language Models (LLMs), and image, audio, video and code generation tools.

To power the AI journey, investment in the GPUs and data centers that enable it has soared, accelerating and transforming global digital infrastructure and it is now becoming clearer where significant changes are most likely to occur.

This five minutes-to-read e-book looks at existing trends and uses the latest market and trend analysis from Structure Research to make four bold predictions about AI’s impact on digital infrastructure over the next five years.

While there is still plenty of scope for shifts, investment, development, and design trends for AI are becoming clearer, giving us strong indications of the potential shape of the market over the next three to five years.
Jabez TanHead of Research at Structure Research

Demand

Data center demand will exceed supply by over 500% by 2030

Hyperscaler annual capital expenditures are projected to reach $375 billion this year, a 36% increase from $275 billion in 2024. 2024 saw a 64% increase in spending compared to 2023. The top four hyperscalers alone are responsible for approximately 90% of this spend, and so far in H2 2025 these capex commitments are being built on rather than revised downwards.

Around half of this investment is being spent on infrastructure needs, such as self-build and leased data center capacity, and half on servers, including CPUs, TPUs and GPUs, for both cloud and AI related needs.

AI colocation revenue for training and inference is forecast to grow at an extraordinary 77% CAGR from 2025-2030, reaching $134 billion by 2030. By 2030, AI is expected to account for approximately 44% of total global data center colocation market revenue.

The steepness of the demand curve has huge implications for the data center industry. Demand for data center capacity is projected to outstrip supply from 2027-2030, with annual global demand reaching nearly 90 90 GW by 2030, exceeding supply by as much as 500%.

This imbalance is already evident. Colocation prices rose by an average of 35% between 2020 and 2023, and vacancy rates in Northern Virginia dropped to less than 1% in 2024. About half of all net new data centre capacity starting in 2025/26 is likely to be AI/ ML GPU clusters.

Typical training clusters are escalating in size, from approximately 40MW in 2020 to projected 1GW to 5GW systems before 2030. Overall, global digital infrastructure is expected to grow by a factor of 4 or 5 over the next 10 years, requiring a capacity increase of 130-150GW in the US alone.

Source: Structure Research

Training versus Inference

There will be 4x more inference infrastructure than training infrastructure by 2030

This is another immense shift. The proportion of Inference to Training is switching as training models mature and the move to revenue generation (i.e. AI inference) takes off.

In terms of critical IT load capacity, AI inference is projected to grow at a 79% CAGR through 2030. This significantly outpaces AI training’s 25% CAGR over the same period. During 2026, inference capacity will overtake training capacity. By 2030, AI inference is expected to account for 80% of total AI critical IT load capacity, a reversal of the balance in 2023.

The shift of balance from training centers which are not so latency sensitive and can therefore go where power and space are plentiful, to inference centers, which need the lowest possible latencies, will put immense development pressure on existing densely populated “eyeball-rich” hubs, as well as accelerating the growth of new AI hubs in densely populated areas where data sovereignty is key and demand for services is high.

Source: Structure Research

 

AI mega hubs

2 GW + data hubs will emerge in every global region

North America

In the US, traditional hubs like Northern Virginia will scale to 8.5 GW by 2030, Dallas (2.8 GW), and Phoenix (2.7 GW) will also see massive absorption.

Europe

London (2.7 GW), Frankfurt (2.68 GW), and Paris (2 GW) are forecast to lead demand by 2030. Many new hubs are growing fast in Europe in particular, driven to some extent by data sovereignty regulations as well as development and latency factors. Also look out for steep growth in Oslo, Barcelona, Madrid, Zaragoza, Lisbon, Berlin, and Dusseldorf.

Asia Pacific

In Asia Pacific (excluding China, where growth is expected to be as or more rapid), Tokyo (2.8 GW), Sydney (2.4 GW), and Johor (2.2 GW) will have the largest capacities by 2030. New APAC AI hubs will also grow fast in Busan, Kyushu, Auckland, and Perth.

The cost of artificial intelligence will drop exponentially

While demand for high-performance (and very expensive) GPUs like the H100 is unprecedented, and there will continue to be fluctuations in the cost of different training models, the averaged cost of artificial intelligence itself is declining rapidly.

The cost of the cheapest Large Language Model (LLM) with a minimum Massive Multitask Language Understanding (MMLU) score of 83 has decreased by approximately 10x every year from 2022 (GPT-3) to 2024 (Llama 3.2 3b). This reduction in the cost of

AI training and inference does not, however, lead to reduced consumption. Instead, like improvements in automobile fuel efficiency that led to more car sales and larger cars, the declining price of intelligence is expected to fuel an exponential increase in the demand curve for AI.

The scale of investment is accelerating this cost- per-unit benefit. Hyperscalers are making enormous investments in GPUs and infrastructure because, in the race to deliver services and achieve Artificial General Intelligence (AGI) they understand that the risk of under-investing is dramatically greater than the risk of over-investing.

The resulting explosion in demand will yield substantial returns, with increasing efficiency driving rather than limiting overall utilization and innovation. This counter-intuitive phenomenon is the Jevons Paradox.

Data Centers top 4 AI predictions whitepaper - AI mega hubs prediction

Source: Structure Research