Fact-checked by Grok 2 weeks ago

Zettascale computing

Zettascale computing refers to systems capable of executing at least 10^{21} double-precision 64-bit floating-point operations per second (), representing a thousandfold increase in computational power over exascale systems that achieve 10^{18} . This scale enables unprecedented simulations in fields such as climate modeling, , , and , where massive datasets and complex algorithms demand extreme parallelism and efficiency. Achieving zettascale performance faces significant challenges, including power consumption projected to reach up to 500 megawatts for a single system—equivalent to the output of several nuclear reactors—necessitating breakthroughs in energy-efficient architectures and cooling technologies. Interconnects must evolve to optical networks for low-latency transfer across millions of processing elements, while systems require petabyte-scale capacities with sub-millisecond access times to handle the data deluge. Software paradigms, including new programming models for heterogeneous combining CPUs, GPUs, and specialized accelerators, are essential to exploit this scale without bottlenecks from or synchronization overheads. Reliability at this level demands fault-tolerant designs to mitigate failures in vast node counts, potentially exceeding one million. Global efforts are accelerating toward zettascale milestones, with Japan's launching the FugakuNEXT project in 2025 in partnership with and , targeting over 600 exaFLOPS in FP8 precision (approaching 0.6 zettaFLOPS) for hybrid -HPC workloads by around 2030. 's OCI Zettascale10 cluster, unveiled in October 2025, delivers up to 16 zettaFLOPS peak performance for training using up to 800,000 GPUs interconnected via a three-tier RoCE network, marking a commercial cloud-based approach available in the second half of 2025. In a 2018 study, researchers from China's proposed a full zettascale by 2035 under the National Key Technology R&D Program, emphasizing heterogeneous architectures and advanced memory stacking. anticipates that such systems will require approximately 500 megawatts, underscoring the need for 2140 gigaFLOPS per watt efficiency to make deployment feasible. These initiatives highlight zettascale's role in addressing grand challenges like and , though distinctions between double-precision HPC and lower-precision metrics remain critical for performance claims.

Fundamentals

Definition

Zettascale computing refers to systems capable of achieving at least 10^{21} floating-point operations per second (zettaFLOPS), specifically using double-precision (64-bit) arithmetic for general-purpose scientific simulations and computations. This metric emphasizes the system's ability to perform a vast number of precise numerical calculations, enabling breakthroughs in fields like climate modeling, , and that require high-fidelity . Unlike related terms in modern and hardware, where "zettaFLOPS" often describes peak performance in lower-precision formats such as FP8 or FP16 tensor operations optimized for large neural networks, zettascale computing in the traditional (HPC) context denotes sustained performance at the double- level. This distinction ensures compatibility with legacy scientific codes that rely on 64-bit accuracy, avoiding the trade-offs in that lower-bit formats introduce for specialized workloads. The term "zetta" derives from the International System of Units (SI) prefix for 10^{21}, formally adopted in 1991 as part of the extension to include prefixes up to yotta (10^{24}). In computing, this prefix was incorporated to describe performance milestones following peta- (10^{15} FLOPS) and exa- (10^{18} FLOPS) scales, with exascale serving as the immediate precursor achieved in systems like the U.S. Department of Energy's Frontier supercomputer.

Performance Scales

The hierarchy of supercomputing performance scales follows the (SI) prefixes, denoting orders of magnitude in floating-point operations per second (). Terascale computing operates at 10^{12} , petascale at 10^{15} , exascale at 10^{18} , zettascale at 10^{21} , and yottascale at 10^{24} . These scales represent exponential leaps, each multiplying the previous by a factor of 1,000, and serve as benchmarks for advancing computational capability in (HPC). Milestones in achieving these scales highlight the progression of HPC hardware. The terascale threshold was crossed in the 1990s, with Intel's ASCI Red supercomputer becoming the first to sustain 1 teraFLOPS in December 1996. Petascale performance was realized in 2008 by IBM's Roadrunner system, which achieved 1.026 petaFLOPS on the LINPACK benchmark. Exascale computing emerged in 2022, when Oak Ridge National Laboratory's Frontier supercomputer delivered 1.102 exaFLOPS, marking the first system to surpass this level on the TOP500 list. These scale jumps, each representing a 1,000-fold increase in computational power, demand fundamental shifts in and efficiency to overcome escalating challenges in power, concurrency, and reliability. For instance, the transition from terascale to petascale involved widespread of multi-core processors and cluster-based designs to handle growing parallelism, while the move to exascale required CPU-GPU architectures, enhanced interconnects, and a data-locality-centric approach to minimize energy costs and data movement—shifting from computation-focused paradigms to resilient, power-aware systems. Such innovations ensure that performance gains do not outpace practical constraints like energy budgets, which could otherwise balloon to unsustainable levels. Zettascale computing, as a 1,000-fold extension beyond exascale, positions itself within this progression to unlock simulations of scales previously unattainable.

Historical Development

Origins and Early Concepts

The concept of zettascale computing, defined as systems capable of performing 10^21 floating-point operations per second, emerged in the mid-2000s amid discussions on the limitations of then-current petascale supercomputers and the need for vastly greater computational power to address grand scientific challenges. A pivotal early forum was the Frontiers of Extreme Computing 2007 Zettaflops Workshop, held October 21-25, 2007, in , which built on the 1994 Petaflops Workshop and explored pathways to zettaflop-scale performance. Sponsored by organizations including , the workshop convened experts in , algorithms, and applications to assess the feasibility of such systems, emphasizing that traditional scaling under would likely provide only a factor of 100 improvement in efficiency, necessitating breakthroughs in parallelism, interconnects, and beyond-CMOS technologies like reversible logic. Key conceptual drivers identified at the workshop centered on the inability of exascale (10^18 ) systems to fully model complex phenomena in fields such as climate dynamics and . For instance, climate modelers argued that 1 zettaflop would be essential for high-fidelity simulations of scenarios, enabling resolutions that capture regional impacts and long-term feedbacks beyond the scope of petascale machines. Similarly, in —a proxy for —computational demands for simulating individual genomes and drug interactions were projected to require zettaflop-scale resources to achieve practical timelines for tailored treatments. These discussions highlighted zettascale as a theoretical necessity for integrating massive datasets and multi-scale simulations, though participants noted that realizing such systems could be "many decades in the future" without architectural innovations. In the 2010s, international forecasts began projecting timelines for zettascale deployment, driven by national strategic priorities in . A 2018 analysis by researchers at the , published in a special issue on , envisioned the feasibility of building a zettascale machine by 2035, outlining hardware and software challenges including processor architectures, memory hierarchies, and fault-tolerant programming models. This prediction aligned with broader U.S. Department of Energy () visions articulated in 2014 reports on post-exascale computing, which emphasized sustained investments in co-design to extend beyond 10^18 for mission-critical simulations in energy and environmental sciences, implicitly targeting zettascale horizons to overcome exascale limitations in data-intensive workflows. The slowing pace of further motivated these early concepts, as scaling alone could no longer deliver the exponential gains needed for zettascale ambitions.

Recent Milestones

In 2022, the United States achieved a major milestone in high-performance computing with the deployment of the Frontier supercomputer at Oak Ridge National Laboratory, which reached a peak performance of 1.1 exaFLOPS, marking the world's first exascale system and laying the groundwork for subsequent pursuits toward zettascale capabilities. Building on this foundation, 2024 saw significant announcements advancing zettascale ambitions. In September, Oracle revealed plans for the world's first zettascale cloud computing cluster, featuring up to 131,072 NVIDIA Blackwell GPUs and targeted for launch in the first half of 2025 to support large-scale AI workloads. In August, Japan announced the initiation of development for its zeta-class supercomputer, a successor to the Fugaku system, designed to achieve 1 zettaFLOPS—1,000 times the performance of current top machines—with construction planned to start in 2025 and backed by over $761 million in government funding. By 2025, these efforts progressed further. In August, RIKEN launched the FugakuNEXT project in partnership with Fujitsu and NVIDIA, targeting over 600 exaFLOPS in FP8 precision (approaching 0.6 zettaFLOPS) for hybrid AI-HPC workloads by around 2030. In October, Oracle unveiled the OCI Zettascale10 cluster, an expansion of its initial zettascale system, delivering a peak performance of 16 zettaFLOPS powered by up to 800,000 NVIDIA GPUs and optimized for massive AI training and inference tasks. Japan continued development of its zeta-class supercomputer, focusing on AI applications with a targeted peak of 1 zettaFLOPS to enhance national research in climate modeling and drug discovery, with construction set to begin in 2025. In Europe, the Barcelona Zettascale Lab achieved a key hardware milestone in July with the receipt of physical prototypes for its CincoRanch TC1 RISC-V chip, advancing indigenous processor development for future zettascale systems and promoting strategic autonomy in HPC.

Enabling Technologies

Hardware Innovations

Advancements in processor architectures are pivotal for achieving zettascale performance, pushing beyond traditional limits through aggressive scaling to 2nm and 1nm nodes. TSMC's 2nm node, entering in the second half of 2025, offers improved density and over prior nodes, with projections of 10-15% performance gains and 25-30% power reduction compared to its 3nm , contributing to cumulative advancements from 7nm baselines. Intel's roadmap targets 1nm by 2027 via its 10A , incorporating gate-all-around s to sustain scaling amid diminishing returns from planar designs. These sub-2nm nodes address constraints by enhancing computational density, though further progress relies on beyond- materials like two-dimensional dichalcogenides (TMDCs), projected for integration in post-2028 integrated circuits to enable atomic-scale s. Memory technologies are also advancing to support the data demands of zettascale systems. High Bandwidth Memory 4 (HBM4), entering production in 2026, provides up to 2x the bandwidth of HBM3E (over 2 TB/s per stack) with improved capacity, integrated via standards like Compute Express Link (CXL) for coherent data sharing across heterogeneous nodes. Photonic integration emerges as a transformative innovation, merging optical and electronic components to overcome electronic interconnect bottlenecks in HPC systems. Photonic-electronic integrated circuits (PICs) provide ultrahigh bandwidth and low-latency data transfer, essential for zettascale architectures where electronic signaling alone cannot scale efficiently. For instance, silicon photonic accelerators achieve latencies as low as 5 ns—two orders of magnitude better than GPU counterparts like NVIDIA's A10—while delivering 8.19 TOPS throughput and up to 4.21 TOPS/W energy efficiency, excluding lasers. Cryogenic cooling further supports hybrid quantum-classical systems by maintaining superconducting qubits at near-absolute zero temperatures, enabling seamless integration with classical processors in future zettascale setups through optimized thermal management that minimizes heat from co-located components. Networking fabrics have evolved to support massive parallelism, with ultra-low-latency protocols like enabling high- interconnects in distributed systems. Oracle's Acceleron platform leverages for , achieving latencies of 2.5 to 9.1 microseconds and up to 3.2 Tb/s aggregate per , facilitating efficient GPU-to-GPU communication in AI-driven HPC. This supports scale-out architectures, such as Oracle's OCI Zettascale10 clusters deploying up to 800,000 B200 GPUs across multiple data centers for 16 zettaFLOPS peak performance, demonstrating viable paths to million- configurations via cloud federation. Efficiency targets for zettascale systems emphasize 10 teraFLOPS per watt within modular designs accommodating 100 MW+ power envelopes, balancing performance gains with sustainable scaling. These metrics, derived from HPC roadmaps, project per-node peaks of 10 petaFLOPS and 1.6 Tb/s inter-node bandwidth to realize 10^21 FLOPS without prohibitive energy demands. Modular architectures, such as those in projected 2035 systems, incorporate disaggregated components for incremental upgrades, though AMD forecasts 500 MW total draw at 2140 GFLOPS/W efficiency under aggressive AI accelerator trends.

Software and Programming

Zettascale computing demands advanced software ecosystems capable of harnessing unprecedented parallelism across heterogeneous architectures, where systems may comprise billions of cores and face frequent hardware failures. Traditional programming models, such as the (MPI), which rely on bulk synchronous communication, struggle with the communication overhead and issues at this scale. Instead, emerging paradigms emphasize asynchronous execution and data-centric approaches to improve and efficiency. For instance, task-based models like those in extensions allow for dynamic scheduling of independent tasks, reducing synchronization barriers and enabling better overlap of computation and communication in distributed environments. A key paradigm shift involves transitioning from MPI-dominated synchronous models to asynchronous, data-centric programming frameworks that prioritize data locality and dependency management over explicit . Data-centric models, such as the Legion programming system, enable developers to express high-level data relationships and privileges, allowing the runtime to automatically manage distribution, partitioning, and coherence across heterogeneous nodes without manual intervention. This approach is particularly suited for zettascale applications, where data movement costs can dominate performance, as demonstrated in exascale prototypes that achieve scalable execution for irregular workloads like graph analytics. Complementing these shifts are fault-tolerant algorithms designed to handle node failures at rates approaching one every few minutes in million-node systems. Techniques like algorithm-based fault tolerance (ABFT) embed redundancy—such as checksums in operations—to detect and correct errors without full restarts, incurring minimal overhead (e.g., 2-3% in large-scale linear algebra solvers). Hierarchical checkpointing further mitigates rollback costs by coordinating checkpoints in processor groups, optimizing recovery time for systems with (MTBF) dropping to seconds per node. Optimizing legacy scientific codes for zettascale requires substantial refactoring to exploit heterogeneous architectures, often involving the of compute-intensive kernels on GPUs while preserving CPU . Millions of lines in established simulations, such as those for or climate modeling, must be ported using directives-based approaches like OpenACC, which enable incremental GPU offloading with minimal code changes—achieving up to 10x speedups in codes without full rewrites. Performance-portable libraries like Kokkos facilitate this by abstracting hardware specifics through policies for execution spaces (e.g., for GPUs), allowing a single codebase to target diverse accelerators while delivering near-peak bandwidth (e.g., 692 GB/s on V100 GPUs for multi-dimensional kernels). These optimizations address the need to retrofit decades-old or C++ applications, ensuring they scale efficiently on zettascale clusters dominated by GPU nodes. Specialized frameworks extend accelerator APIs for zettascale deployment, bridging vendor-specific tools like and . For example, HIPFort automates the conversion of Fortran code to HIP (Heterogeneous-compute Interface for Portability), enabling legacy simulations to run on GPUs with performance comparable to native implementations, as seen in compressible CFD codes achieving multi-petaflop throughput. In AI-native environments, tools like scale to exascale via distributed strategies such as Horovod for multi-GPU synchronization and NCCL for collective operations, attaining over 40 teraflops per GPU in mixed-precision training on systems like —paving the way for zettascale AI workloads through optimized data pipelines and large-batch convergence algorithms. Load balancing in zettascale systems focuses on decentralized mechanisms to equitably distribute tasks across volatile, heterogeneous components, preventing bottlenecks in irregular applications. Dynamic algorithms, such as those using for event-driven workloads, adaptively migrate tasks based on node utilization, reducing imbalance by up to 50% in distributed exascale prototypes without central coordination. Fully distributed strategies, like diffusive load balancing, further enhance by locally adjusting workloads via neighbor exchanges, proving effective for time-varying computations in large-scale simulations where global is infeasible. These techniques ensure sustained performance amid decentralized failures and varying hardware capabilities.

Key Challenges

Energy and Power Constraints

Zettascale computing systems, aiming for performance levels of 10^21 floating-point operations per second, face immense power demands that far exceed those of current exascale machines. While exascale supercomputers like consume approximately 20-30 megawatts—equivalent to the output of a small power plant—zettascale architectures could require up to 500 megawatts or more, potentially rivaling the energy needs of multiple facilities. For context, a hypothetical zettascale system built with today's technology might demand the equivalent of 21 plants to operate continuously, highlighting the physical limits imposed by heat dissipation and electrical infrastructure. To mitigate these challenges, researchers are pursuing efficiency breakthroughs in cooling and interconnect technologies. Liquid immersion cooling, where servers are submerged in non-conductive dielectric fluids, enables direct heat extraction from components, reducing energy overhead by up to 40% compared to traditional in high-density setups. Photonics-based interconnects further alleviate power constraints by replacing electrical signaling with optical links, which consume significantly less energy for data transfer at zettascale scales—potentially cutting interconnect power by orders of magnitude while minimizing thermal loads. Additionally, integrating sources, such as on-site or arrays, into HPC facilities helps offset grid dependency, with studies showing that power systems can help reduce operational energy costs in large-scale deployments. The environmental ramifications of these power scales are profound, with zettascale systems contributing to the IT sector's alongside broader growth, where are projected to account for 3-7% of global by 2030 if unchecked. Innovations like Oracle's Zettascale10 exemplify progress, leveraging power-efficient optics and liquid-cooled GB200 instances to deliver up to 16 zettaFLOPS while prioritizing lower power per performance unit, thereby curbing emissions growth. Such advancements aim to keep carbon emissions manageable, though full zettascale realization could still equate to millions of tons of CO2 annually without sustained gains. Policy frameworks in the U.S. and are increasingly supporting these efforts through green HPC incentives. The U.S. allocates billions for clean energy technologies, including subsidies for renewable-integrated data centers that host advanced computing. In the , the Net-Zero Industry Act and Green Deal Industrial Plan provide streamlined permitting and funding for low-carbon manufacturing of HPC components, fostering collaboration on sustainable supercomputing infrastructures. These measures not only incentivize efficiency but also align zettascale development with broader goals.

Scalability and Data Management

Zettascale computing systems are projected to encompass millions of compute , necessitating robust node coordination mechanisms to maintain operational amid frequent component failures. In such large-scale clusters, hardware faults, including processor errors and interconnect disruptions, occur with increasing frequency as system size grows, demanding redundancy models like coordinated checkpointing and coordinated message logging to ensure continuity without full system restarts. These techniques involve periodically saving application states to stable storage and logging inter- communications, allowing recovery from partial failures while minimizing in environments with up to a million . Data management in zettascale environments presents formidable challenges due to the enormous volumes and velocities of processed, with (I/O) requirements scaling to 10-100 petabytes per second to support sustained at 10^21 floating-point operations per second. Storage capacities must reach up to 1 zettabyte to accommodate intermediate results and datasets, far exceeding current exascale limits of hundreds of petabytes. Inter-node constraints, projected at around 10 petabits per second for aggregate communication, further exacerbate these issues by creating bottlenecks in transfer across the distributed , requiring optimized placement strategies such as hierarchical caching and in-situ to reduce latency. Reliability becomes critically precarious at zettascale, where the (MTBF) for the entire system could diminish to mere seconds, extrapolated from exascale projections of 30 minutes MTBF in million-node clusters assuming constant component reliability. This rapid failure cadence, driven by the Poisson-like distribution of independent hardware events, necessitates adaptive algorithms that dynamically adjust computation granularity, such as task migration and selective replication, to tolerate errors without compromising overall progress. Software tools for , including multilevel checkpointing frameworks, provide essential support by enabling localized recovery and reducing global synchronization overhead. To mitigate centralization risks and enhance , decentralized approaches such as edge-cloud hybrids are emerging as viable strategies for zettascale systems, distributing computational loads across geographically dispersed nodes while leveraging resources for . These models employ protocols for dynamic workload partitioning and , reducing dependency on single points of failure and improving resilience in heterogeneous environments. By integrating for low-latency tasks with cloud-scale aggregation, such architectures address coordination bottlenecks inherent in fully centralized million-node setups.

Future Prospects

Timelines and Projections

In the short term, Oracle Cloud Infrastructure (OCI) anticipates the availability of its Zettascale10 cluster in the second half of 2026, marking a significant step toward zettascale deployment in cloud environments with up to 16 zettaFLOPS of peak performance. Japan's Ministry of Education, Culture, Sports, Science and Technology (MEXT) plans to begin construction of the nation's first zeta-class in 2025, with operations expected by 2030 to achieve zettascale performance for advanced scientific simulations. Medium-term projections indicate widespread zettascale adoption between 2030 and 2035, as forecasted by industry leaders and international roadmaps. AMD's CEO has projected zettascale supercomputers achievable in this timeframe, contingent on advancements in reaching 2140 GFLOPS per watt. Chinese research outlines the feasibility of zettascale systems by 2035, addressing hardware and software challenges to support national computing goals. originally targeted zettascale computing by 2027 as part of a phased roadmap building on exascale systems like , though no recent updates confirm this timeline. These timelines depend on sustained progress from current exascale deployments. Long-term visions position zettascale as an interim milestone before transitioning to yottascale around 2040, enabling even greater scales for global challenges in and . Global approaches to zettascale vary, with the emphasizing cloud-based deployments for and commercial applications, as seen in Oracle's OCI initiatives, while prioritizes scientific and research-focused systems through collaborations like RIKEN's FugakuNEXT project with and .

System Configurations

Prototypical designs for zettascale computing systems emphasize high-performance nodes, massive interconnectivity, and vast storage capacities to achieve unprecedented computational scales. A notable proposal from the (NUDT) in outlines a zettascale with 10 petaFLOPS of peak per , enabling efficient scaling across millions of such units to reach overall zettascale throughput. This design incorporates inter-node communication of 1.6 terabits per second to minimize in transfers, alongside a total storage capacity of 1 zettabyte for handling exascale datasets in scientific simulations. The system is projected to occupy a footprint of approximately 1000 square meters, balancing density with cooling and power distribution needs. In the commercial sector, Oracle's OCI Zettascale10 represents a cloud-based zettascale configuration tailored for workloads, delivering up to 16 zettaFLOPS of peak performance through the integration of 800,000 GPUs distributed across multiple data centers. This setup leverages networking to ensure low-latency communication between GPUs, supporting multi-gigawatt clusters for large-scale model . The modular of the Zettascale10 allows for incremental , with initial deployments focusing on high-bandwidth GPU-to-GPU interconnects to optimize and efficiency. Japan's planned zeta-class , led by and in partnership with , targets more than 600 exaFLOPS in FP8 precision for AI-specific performance within a broader , utilizing a hybrid architecture that combines CPUs, GPUs, and specialized accelerators for versatile scientific computing. This configuration prioritizes AI-driven discovery, with development starting in 2025 and aiming for operational status by 2030, emphasizing energy-efficient components to sustain peak zettascale operations within approximately 40 MW power limits. Across these proposals, general trends in zettascale system configurations highlight modular, data-centric architectures comprising over 10^6 components, where compute resources dynamically adapt to data locality for reduced movement overhead. Such designs reference efficiency targets around 2 teraFLOPS per watt to manage power demands at scale, as projected in recent industry analyses.

Applications

Scientific Simulations

Zettascale computing is expected to dramatically accelerate simulations in physics, climate science, and by providing the computational power to model complex phenomena at finer resolutions and longer timescales than currently possible with exascale systems. This capability stems from the 1,000-fold increase in performance over exascale, enabling simulations that capture intricate multi-physics interactions and large-scale . High-precision floating-point operations remain essential for maintaining accuracy in these models. In climate modeling, zettascale systems could enable higher-resolution global forecasts and simulations of atmospheric dynamics with greater detail, improving predictions of extreme events like hurricanes or heatwaves. Current exascale efforts, such as those using the Energy Exascale Earth System Model (E3SM), achieve kilometer-scale resolutions for climate simulations. For , zettascale computing would significantly reduce the time required for simulations of core-collapse e and other complex phenomena, facilitating more detailed studies of dynamics, neutrino transport, and element formation. Exascale simulations capture multi-physics processes such as hydrodynamics and nuclear reactions in 3D supernova explosions. In biological modeling, zettascale capabilities would support atomic-level simulations or whole-genome in near real-time, bridging the gap between current microsecond-scale and physiological timescales of seconds to days. Molecular simulations of biological systems, which are more compute-intensive than those in due to long-range and large conformational sampling, currently reach milliseconds on exascale hardware; zettascale would enable more extensive explorations for or cellular processes. Examples include through , where zettascale speeds could simulate protein-ligand interactions for billions of compounds, accelerating the identification of novel therapeutics for diseases like cancer or Alzheimer's.

Artificial Intelligence

Zettascale computing enables the training of trillion-parameter models in by providing unprecedented computational scale and efficiency. Oracle Cloud Infrastructure's (OCI) Zettascale10 , for instance, connects up to 800,000 GPUs to support such workloads, achieving 16 zettaFLOPS while handling massive datasets for large language models. This capability addresses the in model complexity, where training a single trillion-parameter model can require millions of exaFLOPS-days on prior systems, but zettascale architectures reduce this to feasible timelines through optimized parallelism and . Oracle's specifically supports multi-gigawatt power envelopes for these AI tasks, ensuring sustained performance without thermal throttling. For inference, zettascale systems deliver low-latency responses essential for generative applications at global scale, enabling interactions for millions of users. OCI's , leveraging high-throughput RDMA networks, facilitates ultra-low inference on trillion-parameter models, as demonstrated in deployments handling diverse batch and interactive workloads. This is particularly vital for generative tasks like , where sub-second response times prevent user friction in enterprise settings, such as collaborative tools powered by summarization. GPU-heavy architectures, including Blackwell platforms, underpin these inference optimizations by distributing computations across vast clusters. In scientific applications, zettascale computing accelerates -driven discoveries in energy and by integrating with high-fidelity simulations. For , optimizes control and resilience under extreme conditions, as outlined in reports on for commercialization. Similarly, in , advanced HPC enables the exploration of vast chemical spaces for energy-efficient compounds, with surrogates accelerating simulations. Such advancements rely on zettascale's ability to handle large-scale predictive modeling. A prominent 2025 example is 's , which deploys on OCI Zettascale10 clusters to power next-generation AI models. The supercluster, developed in collaboration with , underpins 's initial 4.5-gigawatt capacity for and , with expansions to additional U.S. sites advancing the project's $500 billion commitment. This deployment marks the first operational zettascale AI infrastructure, enabling to scale generative models beyond prior limits.

References

  1. [1]
    Zettascale by 2035? China Thinks So - HPCwire
    Dec 6, 2018 · For the purposes of having a consistent metric, they've defined zettascale as a system capable of 10^21 double-precision 64-bit floating-point ...
  2. [2]
    Moving from exascale to zettascale computing: challenges and ...
    Nov 28, 2018 · In this study, we discuss the challenges of enabling zettascale computing with respect to both hardware and software.
  3. [3]
    AMD CEO: The Next Challenge Is Energy Efficiency - IEEE Spectrum
    Feb 21, 2023 · The energy efficiency of computing has not been keeping pace, which could lead to future supercomputers that require as much as 500 megawatts a decade from now.Missing: consumption | Show results with:consumption
  4. [4]
    RIKEN launches international initiative with Fujitsu and NVIDIA for ...
    Aug 22, 2025 · Building the next-generation “AI-HPC platform” to solve complex social challenges through computational science.
  5. [5]
    Oracle Unveils Next-Generation Oracle Cloud Infrastructure ...
    Oct 14, 2025 · OCI Zettascale10 is a powerful evolution of the first Zettascale cloud computing cluster, which was introduced in September 2024. OCI ...
  6. [6]
    AMD says zettascale supercomputers will need half a gigawatt to ...
    Jun 11, 2025 · AMD predicts zetta-scale supercomputers will require 2140GF/watt of power, or half a gigawatt of power.Missing: definition | Show results with:definition
  7. [7]
    Expert questions the validity of Zettascale and Exascale-class AI ...
    Oct 29, 2024 · A leading expert has raised critical questions about the validity of claims surrounding "Zettascale" and "Exascale-class" AI supercomputers.
  8. [8]
    Forget Zettascale, Trouble is Brewing in Scaling Exascale ... - HPCwire
    Nov 14, 2023 · The system delivered peak performance at 1.1 Exaflops of performance and remained the only Exascale system on the list. “I would say that ...
  9. [9]
    What is Kilo, Mega, Giga, Tera, Peta, Exa, Zetta and All That?
    Nov 28, 2022 · Zetta (1 sextillion) was added to the SI metric prefixes in 1991. Examples of -byte sizes. When the prefixes are added to the term byte, it ...
  10. [10]
    SI prefixes - BIPM
    ... SI units can be written using the SI prefixes listed below: Name, Symbol, Multiplying factor. quetta. Q. 1030. ronna. R. 1027. yotta. Y. 1024. zetta. Z. 1021.
  11. [11]
    DOE Explains...Exascale Computing - Department of Energy
    Exascale computers are digital computers, roughly similar to today's computers and supercomputers but with much more powerful hardware.
  12. [12]
    Sandia's ASCI Red, world's first teraflop supercomputer, is ...
    Jun 29, 2006 · ASCI Red first broke the teraflops barrier in December, 1996 and topped the world-recognized LINPAC top-500 computer speed ratings seven ...
  13. [13]
    Breaking the petaflop barrier - IBM
    In 2008, IBM's Roadrunner supercomputer became the first to hit the mark. A ... Roadrunner broke the petaflop barrier on May 25, 2008, but not without ...
  14. [14]
    ORNL's Frontier First to Break the Exaflop Ceiling | TOP500
    May 30, 2022 · The 59th edition of the TOP500 revealed the Frontier system to be the first true exascale machine with an HPL score of 1.102 Exaflop/s.
  15. [15]
    [PDF] Exascale Computing and Big Data: The Next Frontier - OSTI.GOV
    In addition to resilience, scale also brings new challenges in energy management and thermal dissipation. Today's advanced computing systems consume megawatts ...Missing: zettascale | Show results with:zettascale
  16. [16]
    At the Frontier: DOE Supercomputing Launches the Exascale Era
    Jun 7, 2022 · Frontier is a DOE Office of Science exascale supercomputer that was ranked the fastest in the world on the Top500 list released May 30, 2022.
  17. [17]
    Getting To Zettascale Without Needing Multiple Nuclear Power Plants
    Mar 3, 2023 · Moving from double-precision floating to point to other math formats means more efficient computing, particularly if AI and machine learning can ...Missing: definition | Show results with:definition
  18. [18]
    [PDF] Frontiers of Extreme Computing 2007 Zettaflops Workshop - OSTI.gov
    This talk will explore the evolution of high-end computing architectures from the 100 TF scale we are at today through the petaflops and exaflops regimes. Key ...
  19. [19]
    [PDF] Summary of FEC 2007/Zettaflops Workshop
    Dec 10, 2007 · Some workshop participants analyzed computer applications with high societal impact requiring Exaflops and Zettaflops levels of computation.
  20. [20]
    [PDF] DOE ASCAC Subcommi.ee Report February 10, 2014 - The Netlib
    Feb 10, 2014 · 1 Exascale computing is critical for executing the DOE mission. ASCAC reaffirms its findings from previous reports that leadership in high ...Missing: visions zettascale
  21. [21]
    Oracle unveils Zettascale10 AI supercomputer, claims it will be ...
    Oct 14, 2025 · According to Oracle, this can reach up to 16 zettaflops of peak AI performance (typically either FP8 or FP16). It will be available in the ...
  22. [22]
    Japan to start building 1st 'zeta-class' supercomputer ... - Live Science
    Sep 10, 2024 · Japan to start building 1st 'zeta-class' supercomputer in 2025, 1,000 times more powerful than today's fastest machines · Installation of the ...<|control11|><|separator|>
  23. [23]
    Japan Building 'Zeta-Class' Supercomputer, 1,000 Times ... - NDTV
    Sep 24, 2024 · A supercomputer with a speed of 1 zetaFLOPS could make one sextillion (1 followed by 21 zeros) calculations per second. Today's most powerful ...
  24. [24]
    Strategic Meeting for CincoRanch TC2 RTL Freeze
    Jul 30, 2025 · On July 29, 2025, BZL received the first physical prototypes of the CincoRanch TC1 chip, a crucial milestone on our path towards zettascale ...
  25. [25]
    TSMC's 2nm Node: Will It Power the Next Growth Cycle or Pressure ...
    Oct 30, 2025 · As demand for 2nm chips grows for AI and high-performance computing, TSMC is set to gain in the long run. For now, Taiwan Semiconductor is ...
  26. [26]
    Intel plots 1nm silicon for 2027 but are the wheels coming off its ...
    Feb 28, 2024 · The company just doubled down on what was supposedly an ultra-aggressive existing roadmap by revealing plans for a new 10A or 1nm process in 2027.
  27. [27]
    Sub-1-nm-node beyond-silicon materials and devices
    2D materials, especially 2D TMDCs, are considered one of the most competitive new beyond-silicon materials and will be applied to post-Moore ICs in 2028.<|separator|>
  28. [28]
    [2403.14806] Photonic-Electronic Integrated Circuits for High ... - arXiv
    Mar 21, 2024 · Among various emerging technologies, integrated photonics stands out as a promising solution for next-generation high-performance computing, ...
  29. [29]
    An integrated large-scale photonic accelerator with ultralow latency
    Apr 9, 2025 · As an analogue computing approach, photonic computing has the potential advantage of achieving high bandwidth, high energy efficiency and low ...
  30. [30]
    [PDF] Thermal Optimization of Hybrid Cryogenic Computing Systems
    Hybrid cryogenic computing systems are an emerging tech- nology motivated primarily by cryogenic HPC and quantum computing networks. The operating temperature ...
  31. [31]
    High-Performance Networking for AI Infrastructure at Scale
    May 2, 2025 · Custom-designed RDMA over converged ethernet protocol (RoCE v2) · 2.5 to 9.1 microseconds of latency for cluster networking · Up to 3,200 Gb/sec ...Missing: Acceleron | Show results with:Acceleron
  32. [32]
    The New North Star for High Performance Computing | Ayar Labs
    Jun 6, 2023 · ... metrics for a zettascale computer: Total power consumption: 100 MW; Power efficiency: 10 teraflops/watt; Peak performance per node: 10 petaflops ...
  33. [33]
    Parallel Paradigms in Modern HPC: A Comparative Analysis of MPI ...
    Jun 18, 2025 · This paper presents a comprehensive comparison of three dominant parallel programming models in High Performance Computing (HPC)
  34. [34]
    Legion Programming System
    Legion is a data-centric parallel programming system for writing portable high performance programs targeted at distributed heterogeneous architectures.
  35. [35]
  36. [36]
    [PDF] Fault tolerance techniques for high-performance computing
    This report provides an overview of fault tolerance techniques for High Performance Computing (HPC). We present scheduling algorithms to cope with faults on ...
  37. [37]
    NASA@SC19: Teaching Your Legacy Code New (GPU) Tricks
    Mar 25, 2020 · By using OpenACC, we can rapidly add GPU acceleration with minimal impact on the original code base, and maintain full portability—the code ...
  38. [38]
    [PDF] Kokkos 3: Programming Model Extensions for the Exascale Era
    Apr 1, 2022 · Kokkos 3 is a programming model for diverse high-performance architectures, with new features like hierarchical parallelism and arbitrary-sized ...
  39. [39]
    High-speed turbulent flows towards the exascale: STREAmS-2 ...
    STREAmS-2, a compressible CFD code now runs using both NVIDIA and AMD GPUs. CUDA Fortran code can be automatically converted to HIPFort.
  40. [40]
    Scaling the Summit of Exascale Deep Learning - The TensorFlow Blog
    Feb 11, 2019 · TensorFlow, in turn, leveraged routines implemented in C and C++ for high performance, providing precision and freeing developers to move fast.Dl Hardware: Gpus And Mixed... · Open Challenges · Data Management At ScaleMissing: zettascale | Show results with:zettascale
  41. [41]
    A reinforcement learning-based mechanism for managing dynamic ...
    Dec 1, 2024 · In this paper, we propose a load balancing mechanism named ExaQ to solve the challenge of events with D&I nature, that are not considered in the initial ...
  42. [42]
    [PDF] Optimizing Distributed Load Balancing for Workloads with Time ...
    This paper explores fully distributed load balancing strategies that have shown great promise for exascale-level computing but are chal- lenging to ...
  43. [43]
    Computing Testbeds | PNNL
    Computing testbeds offer experimental platforms for testing advanced technologies, driving innovation in computing performance and scalability.
  44. [44]
    [PDF] Characterizing the Opportunity for Low-Carbon and Low-Cost High ...
    growing power consumption poses significant challenges in terms of operational costs, infras- ... A zettascale computer today would need 21 nuclear power plants.
  45. [45]
    ZettaScaler 3.0 (Liquid Immersion Cooling and Air Cooling System)
    HPC (High Performance Computing) system that is equipped with the PEZY-SC3 processor. It has a liquid-immersion cooling system and an air-cooling system.Missing: advanced photonics zettascale
  46. [46]
    Challenges of exascale to zettascale computing and technology ...
    This means that a machine's peak performance, the theoretical highest processing power it can reach, will be different from its real-world, sustainable ...<|control11|><|separator|>
  47. [47]
    TSMC's Silicon Photonics Architecture: Why Couplers and Optical ...
    Jun 1, 2025 · Silicon photonics has emerged as a key solution for next-generation optical interconnects, offering high speed, low power consumption, and seamless integration ...
  48. [48]
    Chapter 2: High Performance Computing and Data Centers
    Dec 3, 2021 · The promise of photonics and spot-cooling challenges and solutions are discussed in detail in the Photonics chapter and the crosscutting ...<|separator|>
  49. [49]
    A review on the decarbonization of high-performance computing ...
    1. Renewable energy sources. The integration of RES, such as wind or photovoltaic systems, in the power supply of computing centers is an obvious measure to ...
  50. [50]
    How to Advance Sustainable High Performance Computing - TierPoint
    May 30, 2024 · Exchanging hardware for more energy-efficient equipment can be as simple as buying something newer. Renewable and Alternative Energy Integration.Heat Reuse And Waste Heat... · Workload Optimization And... · Can Artificial Intelligence...
  51. [51]
    Intel Editorial: Accelerated Innovations for Sustainable, Open HPC
    May 31, 2022 · It has been estimated that by 2030, between 3% and 7%1 of global energy production will be consumed by data centers, with computing ...
  52. [52]
    Oracle debuts AI infrastructure advances, multicloud billing flexibility
    Oct 16, 2025 · “Customers can build, train, and deploy their largest AI models into production using less power per unit of performance and achieving higher ...
  53. [53]
    [PDF] Toward Sustainable HPC: Carbon Footprint Estimation and ... - arXiv
    In this work 1, we present a comprehensive analysis of the carbon footprint of high- performance computing (HPC) systems, considering the carbon footprint ...Missing: zettascale | Show results with:zettascale
  54. [54]
    The US-EU race for green subsidies can help fight climate change
    Feb 14, 2023 · Hailed as a major step toward addressing climate change, the Inflation Reduction Act (IRA) of 2022 provides $369 billion in green subsidies ...
  55. [55]
    EU finalizes green tech bill, responding to US effort - Politico.eu
    Feb 6, 2024 · The Net-Zero Industry Act won't match the U.S. subsidy splurge, but it will offer incentives to local green tech manufacturers.
  56. [56]
    The Green Deal Industrial Plan - European Commission
    The Green Deal Industrial Plan enhances the competitiveness of Europe's net-zero industry and is accelerating the transition to climate neutrality. It does so ...
  57. [57]
    [PDF] Addressing failures in exascale computing - Marc Snir
    In addition, disk failure rates are often higher than the nominal MTBF would indicate, with a. 2%–4% yearly failure rate common (Schroeder and. Gibson, 2007); ...
  58. [58]
    Moving from exascale to zettascale computing - SpringerLink
    Nov 28, 2018 · Abstract: High-performance computing (HPC) is essential for both traditional and emerging scientific fields,.
  59. [59]
    [PDF] Challenges in High-Performance Computing - The Distant Reader
    Aug 1, 2023 · Peak performance per node 10 Pflops/node. Bandwidth between nodes. 1.6 Tb/s. I/O bandwidth. 10-100 PB/s. Storage capacity. 1 ZB. Floor space.
  60. [60]
  61. [61]
    Intel and AMD Path to Zettaflop Supercomputers | NextBigFuture.com
    Feb 23, 2023 · In 2021, Intel talked reaching Zettascale in three phases. AMD CEO talked about getting to a Zettaflop in 2030-2035. AMD CEO indicated that anMissing: projections | Show results with:projections
  62. [62]
    Intel Aims For Zettaflops By 2027, Pushes Aurora Above 2 Exaflops
    Oct 27, 2021 · ... Intel's zettascale goals as a final thought. These turn out to be more important than what Gelsinger said about Aurora having its processing ...
  63. [63]
    Supercomputing Is Heading Toward an Existential Crisis | TOP500
    Dec 11, 2018 · Where the zettascale concept falls apart is the expected use of these machines for applications other than 64-bit numerical computing. In fact, ...
  64. [64]
    Oracle's big bet for AI: Zettascale10 - Network World
    Oct 15, 2025 · Oracle's Zettascale10 supercomputer connects 800,000 Nvidia GPUs for massive-context training and inference.
  65. [65]
    Japan to begin developing ZetaFLOPS-scale supercomputer in 2025
    Aug 27, 2024 · This new supercomputer aims to achieve an unprecedented 50 ExaFLOPS of AI performance with Zetta-scale peak performance in mind to use AI for scientific ...
  66. [66]
    Will 1000 ExaFlop Supercomputers Come from Brute Force Scaling ...
    Feb 10, 2019 · The NUDT researchers described other metrics for a zettascale system. Power consumption of 100 MW. A power efficiency of 10 teraflops/watt
  67. [67]
  68. [68]
    Exascale Computing and Data Handling: Challenges and ...
    This paper describes key technical and budgetary challenges, identifies gaps and ways to address them, and makes a number of recommendations.Missing: zettascale | Show results with:zettascale
  69. [69]
    Next-Generation 3D Core-Collapse Supernova
    Hours 1,850,000 Node-Hours. Year 2022. Domain Physics ... and resources are aligning, and erect a standard model for core-collapse supernova explosions in the emerging era of the exascale.
  70. [70]
    A closer look at "training" a trillion-parameter model on Frontier
    Jan 13, 2024 · “A rough estimate [11] tells us that training a Trillion parameter model on 1-30 Trillion tokens will require [...] 6 - 180 Million exa-flops ( ...
  71. [71]
    Oracle Unveils Next-Generation Oracle Cloud Infrastructure ...
    Oct 14, 2025 · OCI Zettascale10 is a powerful evolution of the first Zettascale cloud computing cluster, which was introduced in September 2024.
  72. [72]
  73. [73]
    Oracle Offers First Zettascale Cloud Computing Cluster
    Sep 11, 2024 · OCI Superclusters with H100 GPUs can scale up to 16,384 GPUs with up to 65 ExaFLOPS of performance and 13Pb/s of aggregated network throughput.Missing: inter- node Pb/
  74. [74]
    Oracle Debuts Zettascale Supercomputers Powered by NVIDIA ...
    Sep 18, 2024 · Oracle has unveiled a new category of supercomputers that will provide unprecedented compute power for advanced large-scale model training.
  75. [75]
    [PDF] fusion-ai-report.pdf - Clean Air Task Force
    By applying these standards, the project aims to enhance the availability and utility of fusion data for ML applications. The project, led by the Massachusetts ...
  76. [76]
    MATERIALS – Multiscale Accelerated Research for Innovative ...
    AI can be effectively used to drive discovery of energy materials. Key Capabilities Driving Success. AI-Driven Materials Discovery: Expertise in integrating ...Missing: fusion zettascale
  77. [77]
    Stargate advances with 4.5 GW partnership with Oracle - OpenAI
    Jul 22, 2025 · Oracle and OpenAI have entered an agreement to develop 4.5 gigawatts of additional Stargate data center capacity in the US.
  78. [78]
    OpenAI, Oracle and Vantage to build Stargate data center ... - Reuters
    Oct 22, 2025 · OpenAI, Oracle and Vantage to build Stargate data center site in Wisconsin. By Reuters. October 22, 20251:42 PM PDTUpdated October 22, 2025.