Advancements in Scalable Computing and Tensor Processing

The field of high-performance computing is witnessing significant advancements in scalable computing and tensor processing. Researchers are exploring innovative approaches to improve the strong scaling of molecular dynamics simulations, enabling faster time-to-solution on heterogeneous supercomputers. Furthermore, there is a growing focus on developing distributed frameworks for multimodal image registration, allowing for the analysis of large-scale datasets. The development of performance-portable compiler frameworks is also gaining traction, enabling the integration of codes from different frameworks and facilitating extensibility to new architectures. Additionally, libraries for interactive GPU-accelerated large tensor processing and visualization are being developed, enabling researchers to efficiently process and visualize large datasets. Noteworthy papers in this area include: LAPIS, a performance-portable compiler framework that addresses the challenges of portability, performance, and productivity. Palace, a library for interactive GPU-accelerated large tensor processing and visualization that enables the development of out-of-core pipelines on workstation hardware.

Sources

Redesigning GROMACS Halo Exchange: Improving Strong Scaling with GPU-initiated NVSHMEM

A Scalable Distributed Framework for Multimodal GigaVoxel Image Registration

LAPIS: A Performance Portable, High Productivity Compiler Framework

Palace: A Library for Interactive GPU-Accelerated Large Tensor Processing and Visualization

Improving Runtime Performance of Tensor Computations using Rust From Python

Programming RISC-V accelerators via Fortran

Built with on top of