Publications
2024
- Stochastic and Low-Scaling Techniques: General DiscussionAli Alavi, Marcus Allen, Kemal Atalar, Timothy C. Berkelbach, George H. Booth, Hugh G. A. Burton, Garnet Kin-Lic Chan, Luca Craciunescu, Don Danilov, Werner Dobrautz, Francesco A. Evangelista, Maria-Andreea Filip, Emmanuel Giner, Gabriel Greene-Diniz, Andreas Grüneis, Yang Guo, Gaurav Harsha, Basil Ibrahim, Venkat Kapil, Daniel Kats, Peter J. Knowles, Marie-Bernadette Lepetit, Ke Liao, Giovanni Li Manni, Pierre-François Loos, Erika Magnusson, Antoine Marie, Carlos Mejuto-Zaera, Frank Neese, Arman Nejad, Verena A. Neufeld, Katarzyna Pernal, Felix Plasser, Andrew W. Prentice, Visagan Ravindran, Umatur Rehman, Markus Reiher, Christian Schilling, Gustavo Scuseria, James Shee, Benjamin X. Shi, Kalman Szenes, David P. Tew, Alex J. W. Thom, Zikuan Wang, Dominika Zgid, Andrew Zhu, and Martijn A. ZwijnenburgFaraday Discussion on Correlated Electronic Structure, London, UK, 2024
- Striking the right balance of encoding electron correlation in the Hamiltonian and the wavefunction ansatzKalman Szenes, Maximilian Mörchen, Paul Fischill, and Markus ReiherFaraday Discussion on Correlated Electronic Structure, London, UK, 2024
Multi-configurational electronic structure theory delivers the most versatile approximations to many-electron wavefunctions, flexible enough to deal with all sorts of transformations, ranging from electronic excitations, to open-shell molecules and chemical reactions. Multi-configurational models are therefore essential to establish universally applicable, predictive ab initio methods for chemistry. Here, we present a discussion of explicit correlation approaches which address the nagging problem of dealing with static and dynamic electron correlation in multi-configurational active-space approaches. We review the latest developments and then point to their key obstacles. Our discussion is supported by new data obtained with tensor network methods. We argue in favor of simple electron-only correlator expressions that may allow one to define transcorrelated models in which the correlator does not bear a dependence on molecular structure.
- Domain-Specific Implementation of High-Order Discontinuous Galerkin Methods in Spherical GeometryKalman Szenes, Niccolò Discacciati, Luca Bonaventura, and William SawyerComputer Physics Communications, 2024
In recent years, domain-specific languages (DSLs) have achieved significant success in large-scale efforts to reimplement existing meteorological models in a performance portable manner. The dynamical cores of these models are based on finite difference and finite volume schemes, and existing DSLs are generally limited to supporting only these numerical methods. In the meantime, there have been numerous attempts to use high-order Discontinuous Galerkin (DG) methods for atmospheric dynamics, which are currently largely unsupported in main-stream DSLs. In order to link these developments, we present two domain-specific languages which extend the existing GridTools (GT) ecosystem to high-order DG discretization. The first is a C++-based DSL called G4GT, which, despite being no longer supported, gave us the impetus to implement extensions to the subsequent Python-based production DSL called GT4Py to support the operations needed for DG solvers. As a proof of concept, the shallow water equations in spherical geometry are implemented in both DSLs, thus providing a blueprint for the application of domain-specific languages to the development of global atmospheric models. We believe this is the first GPU-capable DSL implementation of DG in spherical geometry. The results demonstrate that a DSL designed for finite difference/volume methods can be successfully extended to implement a DG solver, while preserving the performance-portability of the DSL.
2023
- SC’23High-Performance and Programmable Attentional Graph Neural Networks with Global Tensor FormulationsMaciej Besta, Pawel Renc, Robert Gerstenberger, Paolo Sylos Labini, Alexandros Ziogas, Tiancheng Chen, Lukas Gianinazzi, Florian Scheidl, Kalman Szenes, Armon Carigiet, Patrick Iff, Grzegorz Kwasniewski, Raghavendra Kanakagiri, Chio Ge, Sammy Jaeger, Jarosław Wąs, Flavio Vella, and Torsten HoeflerSupercomputing, Denver, CO, USA, 2023
Graph attention models (A-GNNs), a type of Graph Neural Networks (GNNs), have been shown to be more powerful than simpler convolutional GNNs (C-GNNs). However, A-GNNs are more complex to program and difficult to scale. To address this, we develop a novel mathematical formulation, based on tensors that group all the feature vectors, targeting both training and inference of A-GNNs. The formulation enables straightforward adoption of communication-minimizing routines, it fosters optimizations such as vectorization, and it enables seamless integration with established linear algebra DSLs or libraries such as GraphBLAS. Our implementation uses a data redistribution scheme explicitly developed for sparse-dense tensor operations used heavily in GNNs, and fusing optimizations that further minimize memory usage and communication cost. We ensure theoretical asymptotic reductions in communicated data compared to the established message-passing GNN paradigm. Finally, we provide excellent scalability and speedups of even 4–5x over modern libraries such as Deep Graph Library.