👉 Book Web Page: https://torres.ai/supercomputing-for-ai
Supercomputing for Artificial Intelligence is a practical, systems-oriented guide to training modern AI models at scale.
This book is not about writing code faster.
It is about understanding what happens when that code runs — on GPUs, across nodes, under real resource constraints.
In an era where AI tools can generate entire training pipelines in minutes, the real engineering challenge has shifted: performance, scalability, efficiency, and informed trade-offs.
HPC for AI is about judgment, not recipes.
Supercomputing for Artificial Intelligence provides a rigorous yet hands-on introduction to High Performance Computing as it applies to modern AI workloads.
The focus is explicitly on training, not inference.
Readers are guided from foundational supercomputing concepts to the efficient and scalable training of deep learning models on real supercomputing platforms.
The book integrates:
- computer architecture and modern GPU systems
- parallel and distributed execution models
- deep learning frameworks (TensorFlow and PyTorch)
- performance analysis and scalability metrics
- reproducible experiments on production-grade infrastructures
Rather than presenting isolated techniques, the book is structured as a learning path whose technical culmination is the ability to reason about and execute large-scale AI training workloads.
AI-assisted coding tools are changing how software is written. They are not changing the fundamental physics of computation.
Generating code is becoming trivial. Understanding bottlenecks, overheads, scaling limits, and cost–performance trade-offs is not. This book addresses that gap.
It is written for readers who want to:
- understand why performance behaves the way it does
- measure instead of guess
- scale only when it makes sense
- avoid mistaking “more GPUs” for “better systems”
In short: to develop engineering judgment in AI systems.
The primary technical focus of the book is the training of AI models on high-performance computing systems. Later chapters guide the reader through:
- efficient single-node training
- data parallelism and distributed training
- scalability analysis and diminishing returns
- end-to-end workflows for modern deep learning models
This includes preparation for training contemporary Large Language Models on distributed GPU-based supercomputers.
Earlier chapters cover:
- supercomputing fundamentals
- system architecture
- software environments and schedulers
- classical parallel programming models
These sections can also be read independently as a rigorous introduction to High Performance Computing.
Topics such as inference optimization, deployment, and edge execution are introduced only where needed for system-level context.
This book is currently used as core reference material in master’s-level courses on supercomputing and artificial intelligence, including the
HPC for AI course (MEI master, FIB–UPC).
All examples and experiments are designed to run on real supercomputing platforms, not simplified toy environments. Companion code and fully reproducible experiments are provided.
The second edition of Supercomputing for Artificial Intelligence will be published openly in HTML format at the end of January 2026, once the current publishing agreement concludes.
This open edition will make the full content freely accessible to students, researchers, and practitioners worldwide.
The book is currently available in digital and print editions via Amazon
Please use the following reference format when citing this book in academic or teaching materials:
Torres, J. (2025). Supercomputing for Artificial Intelligence: Foundations, Architectures, and Scaling Deep Learning. WATCH THIS SPACE Book Series – Barcelona. Amazon KDP. ISBN: 979-831932835-9.
@book{HPC4AI2025,
author = {Jordi Torres},
title = {Supercomputing for Artificial Intelligence: Foundations, Architectures, and Scaling Deep Learning},
year = {2025},
publisher = {Amazon KDP},
series = {WATCH THIS SPACE Book Series – Barcelona},
isbn = {979-831932835-9},
}