Home | Research | Groups | Suvrit Sra

Research Group Suvrit Sra


Link to website at TUM

Suvrit Sra

Prof. Dr.

Principal Investigator

Resource Aware Machine Learning

Suvrit Sra

is Professor for Resource Aware Machine Learning at TU Munich.

He specializes in robust, reliable, and resource-efficient machine learning methods. His research focuses, in particular, on solving optimization problems for machine learning with multiple parameters. For example, these complex optimization problems are used in autonomous driving so that a car can reliably distinguish a sign from a person.

Team members @MCML

PhD Students

Link to website

Pouria Fatemi

Resource Aware Machine Learning

Link to website

Xuhui Zhang

Resource Aware Machine Learning

Recent News @MCML

Publications @MCML

2025


[1]
A. Bergmeister, M. K. Lal, S. Jegelka and S. Sra.
A projection-based framework for gradient-free and parallel learning.
Preprint (Jun. 2025). arXiv
Abstract

We present a feasibility-seeking approach to neural network training. This mathematical optimization framework is distinct from conventional gradient-based loss minimization and uses projection operators and iterative projection algorithms. We reformulate training as a large-scale feasibility problem: finding network parameters and states that satisfy local constraints derived from its elementary operations. Training then involves projecting onto these constraints, a local operation that can be parallelized across the network. We introduce PJAX, a JAX-based software framework that enables this paradigm. PJAX composes projection operators for elementary operations, automatically deriving the solution operators for the feasibility problems (akin to autodiff for derivatives). It inherently supports GPU/TPU acceleration, provides a familiar NumPy-like API, and is extensible. We train diverse architectures (MLPs, CNNs, RNNs) on standard benchmarks using PJAX, demonstrating its functionality and generality. Our results show that this approach is as a compelling alternative to gradient-based training, with clear advantages in parallelism and the ability to handle non-differentiable operations.

MCML Authors
Link to website

Andreas Bergmeister

Foundations of Deep Neural Networks

Link to Profile Stefanie Jegelka

Stefanie Jegelka

Prof. Dr.

Foundations of Deep Neural Networks

Link to Profile Suvrit Sra

Suvrit Sra

Prof. Dr.

Resource Aware Machine Learning