News

I am working on a project that requires high-performance batched matrix multiplication (i.e., 3D tensor multiplication) where all inputs are stored in the FP8 data type. However, I noticed that ...
We investigate a novel approach to approximate tensor-network contraction via the exact, matrix-free decomposition of full tensor-networks. We study this method as a means to eliminate the propagat ...
We investigate the efficient combination of the canonical polyadic decomposition (CPD) and tensor hyper-contraction (THC) approaches. We first present a novel low-cost CPD solver that leverages a ...
According to Google DeepMind, AlphaEvolve has successfully discovered multiple new algorithms for matrix multiplication, surpassing the previous AlphaTensor model in efficiency and performance (source ...
Discover how nvmath-python leverages NVIDIA CUDA-X math libraries for high-performance matrix operations, optimizing deep learning tasks with epilog fusion, as detailed by Szymon Karpiński.
Matrix multiplication (MatMul) is a fundamental operation in most neural networks, primarily because GPUs are highly optimized for these computations. Despite its critical role in deep learning, ...
Engheta and colleagues have now set their sights on vector–matrix multiplication, which is a vital operation for the artificial neural networks used in some artificial intelligence systems. The team ...
Matrix multiplication advancement could lead to faster, more efficient AI models At the heart of AI, matrix math has just seen its biggest boost "in more than a decade.” ...