High-performance computing (HPC) involves using multiple interconnected computers combined with parallel processing approaches to solve problems that are too large or complex for a single computer. Generally, HPC systems, such as supercomputers and clusters, use high-bandwidth, low-latency network interconnects to enable fast and efficient communication across processes running on the HPC system. In geographic information science (GIScience), growth in data size and analytical complexity drive more demand for computing power and the need for HPC. Researchers and practitioners use HPC to process more geospatial data, run finer-grained simulations, and explore problems at spatial and temporal scales that were previously impossible. The key challenge in using HPC is coordinating and synchronizing dozens, hundreds, or thousands of processors simultaneously.
Shook, E. (2026). High-Performance Computing and GIS. The Geographic Information Science & Technology Body of Knowledge (Issue 1, 2026 Edition), John P. Wilson (ed). DOI: 10.22224/gistbok/2026.1.2
High-performance computing (HPC) involves using multiple interconnected computers, combined with parallel processing, to solve problems that are too large or complex for a single computer. Achieving suitable performance with HPC systems demands technical skills and a deep understanding of the problems you are trying to solve, because these problems must be decomposed into multiple sub-problems to be solved simultaneously by dozens, hundreds, or thousands of processes, while maintaining synchronous communication across all processes. This approach is called parallel programming (see GIS & Parallel Programming, Shook 2019). A hallmark of geospatial problems that require HPC is that they involve highly synchronous processing, meaning processes must communicate frequently, which is common in complex analyses and models with significant spatial and/or temporal dependencies. This is distinct from High-Throughput Computing (HTC), which allows tasks to be mainly processed independently of one another. For an in-depth comparison between HPC and HTC, see High-Throughput Computing (Carbajales-Dale et al., 2025). HPC is also distinct from Cloud Computing, which provides on-demand computing that may or may not be high-performance, with low-latency/high-bandwidth interconnects, large memory, or accelerators. A prime example of cloud computing is Amazon Web Services (AWS).
HPC systems are not a single computer but rather a collection of computers, commonly referred to as a
Clusters have two common types of nodes, another name for computers. A
Job management systems are responsible for allowing multiple users to submit jobs that request compute resources and scheduling jobs across the available resources as efficiently as possible. Slurm is a common job management system used in many clusters and supercomputers. Common features of job management systems include a job submission, which allows users to request resources; a job queue, which allows jobs to wait until resources are available; a job scheduler, which prioritizes and sequences job execution and allocates them across resources; and a job monitoring system, to ensure jobs are functioning properly.
Interconnects connect nodes together using network cables. Infiniband and Ethernet are two common networking technologies used in HPC systems. Interconnects commonly feature low-latency, meaning that signals are sent between nodes quickly, and high-bandwidth, meaning that many signals can be sent between nodes. Specialized interconnects can be a major cost in clusters and supercomputers to improve overall performance and allow the cluster of computers to act as a single processing system. Interconnects not only allow nodes to communicate for coordination, but also allow for data to be transferred quickly and efficiently between nodes. HPC systems oftentimes use different network topologies to achieve improved performance, such as fat tree and torus.
Processing power is central to HPC systems. There are different types of processors used in HPC. Below, three types of processors are summarized. HPC systems can comprise different types of processors. If an HPC system supports a single type of computing unit, such as only CPUs, it has a
Central Processing Units (
Graphics Processing Units (
Accelerators are a broader class of specialized processing units designed to accelerate domain-specific components of a problem. A common characteristic of accelerators is they diverge from the seminal von Neumann architecture model which is used by CPUs. This traditional architecture model features control units and logic units that are separated from memory units. In contrast, accelerators integrate control, logic, and memory in unique ways to maximize performance for specific problems. However, a common tradeoff to these more exotic architecture models is they are more difficult to program (Shi, 2014).
4. Challenges and Opportunities
Three key challenges exist for the GIS&T community to use HPC. Firstly, is the modality of development and use of computational resources. Many GIScientists use graphical user interfaces (GUIs) to conduct geospatial analytics and modeling. For example, commercial GIS tools such as Esri ArcPro and open-source tools such as QGIS are GUIs. However, for many HPC resources, the primary mode of interaction is through command-line interfaces. Secondly, almost all clusters and supercomputers use Linux or Unix-based operating systems (OSs). Many GIScientists are familiar with Windows-based operating systems. Switching OSs can break existing code and may require learning new software and libraries. Thirdly, parallel programming is generally required to effectively use HPC and learning parallel programming can be considered a challenge (Shook et al., 2016). For example, high-resolution environmental data such as temperature and precipitation at the country or global scale can be terabytes in size, yet analyzing it is a typical GIS task. Conducting potentially multi-terabyte-scale analyses on a desktop computer can be impractical or impossible due to storage and memory limitations. To overcome these limitations, GIS professionals can use data parallelism and functional parallelism to decompose the problem into multiple sub-problems that can be solved simultaneously. For more information on parallel programming, see GIS & Parallel Programming (Shook, 2019).
Three key opportunities exist. Firstly, cyberinfrastructure (CI) can lower barriers to accessing HPC. In the United States, ACCESS is a National Science Foundation (NSF) supported cyberinfrastructure that provides researchers and students free access to multiple HPC systems across the US. Another example of CI in the US is the I-GUIDE project (I-GUIDE Project, n.d.). Secondly, HPC removes computational barriers for GIS&T, enabling scientists and professionals to tackle larger, more complex geospatial problems worldwide (Armstrong, 2000). Thirdly, advances in software systems are lowering barriers. One example is Open OnDemand (https://www.openondemand.org/), which allows for web-based access to traditional GUI software running on HPC hardware. Open OnDemand and similar systems can run traditional GIS software such as QGIS, reducing a key barrier in needing to learn and use command-line interfaces.