Accelerating Machine Learning with Innovative Architectures and Hardware

The field of machine learning is moving towards more efficient and scalable architectures, with a focus on accelerating model training and improving energy efficiency. Recent developments have explored novel adaptive batch size algorithms, neural network expansion methods, and hardware-software co-designs to achieve state-of-the-art performance. Notably, innovative approaches have been proposed to prevent neuron inactivity during network expansion and to leverage physical hardware properties to improve energy efficiency. Furthermore, the integration of optical and photonic technologies has enabled the development of high-speed and low-latency computing systems.

Noteworthy papers include: DIVEBATCH, which proposes a gradient-diversity aware batch size adaptation algorithm to accelerate model training. Otters, which introduces an energy-efficient spiking transformer via optical time-to-first-spike encoding, achieving state-of-the-art accuracy and improved energy efficiency. High Clockrate Free-space Optical In-Memory Computing, which presents a fanout spatial time-of-flight optical neural network for fast and low-latency computing.

Sources

DIVEBATCH: Accelerating Model Training Through Gradient-Diversity Aware Batch Size Adaptation

Shared-Weights Extender and Gradient Voting for Neural Network Expansion

Otters: An Energy-Efficient SpikingTransformer via Optical Time-to-First-Spike Encoding

High Clockrate Free-space Optical In-Memory Computing

Built with on top of