Research Direction: Linear Operators in Modern ML

A research-facing overview of how basic linear maps reappear inside transformers, graph models, operator learning, and representation pipelines.
Modified

April 26, 2026

Keywords

research direction, matrices, operators, transformers, graph neural networks

1 Direction Summary

Modern ML models are not purely linear, but they keep relying on linear operators as structural building blocks.

The stable backbone is:

  • learned projection matrices
  • compositions of maps across layers
  • operator viewpoints on graph propagation and feature transport

The frontier lies in how these operators are parameterized, constrained, or interpreted inside much larger nonlinear systems.

2 Core Math

  • linear maps and matrix representations
  • composition and change of coordinates
  • spectra of learned or graph-based operators
  • structured matrices and parameter sharing

3 Representative Problems

  • how should learned feature maps be parameterized?
  • which operator structures improve efficiency or interpretability?
  • how do graph or sequence operators reshape representation spaces over many layers?
  • when should a model be read as repeated operator application rather than only as a generic neural network?

4 Representative Venues

  • NeurIPS
  • ICML
  • ICLR
  • JMLR
  • Numerical Algorithms

5 Starter Reading Trail

  1. Matrices and Linear Maps
  2. Learned Linear Projections in Transformers
  3. Linear Operators in Modern AI Systems
  4. Deep learning, transformers and graph neural networks: a linear algebra perspective

6 Open Questions

  • which operator structures matter most for generalization versus efficiency?
  • how should we compare dense learned maps with sparse, low-rank, or graph-structured alternatives?
  • when can operator-level interpretation survive the surrounding nonlinear architecture?

7 What To Learn Next

8 Sources and Further Reading

Back to top