Directions
research directions, machine learning theory, graph learning, diffusion, optimization
1 Why This Page
Research directions are where the site should stop feeling like a curriculum and start feeling like a gateway.
But a good directions page should not just say what is fashionable. It should say:
- what stable math the direction stands on
- what the active questions are
- what breaks if you skip the prerequisites
- what a realistic first reading trail looks like
That is the design rule here:
stable backbone first, active frontier second
2 Directions At A Glance
Type: top-level research mapSetting: readers who want to move from topic mastery into current research areasMain claim: research directions make the most sense when organized by the theorem families behind themWhy it matters: this is the layer that turns “I studied the topic” into “I know what to read next”
3 How To Read This Page
For each direction below, read in this order:
Stable backboneWhat is active nowWhere it appearsStarter trail
If the stable backbone already feels weak, follow the trail backward rather than pushing deeper into frontier papers.
4 Direction 1: High-Dimensional Probability And Random Matrices
4.1 Stable backbone
- concentration inequalities
- random vectors and random matrices
- norm control and geometric probability
- empirical process intuition
4.2 What is active now
This direction stays central because many modern questions in data science and ML become high-dimensional before they become algorithmic.
Typical active questions:
- how do random matrix effects shape generalization and interpolation?
- how do concentration tools scale to dependent, structured, or heavy-tailed settings?
- which probabilistic tools remain reliable in modern overparameterized regimes?
4.3 Where it appears
- high-dimensional statistics
- covariance and regression theory
- sketching and randomized linear algebra
- generalization analysis
- compressed sensing and inverse problems
4.4 Starter trail
5 Direction 2: Modern Learning Theory
5.1 Stable backbone
- empirical risk and population risk
- bias-variance and validation
- stability, complexity, and uniform convergence
- optimization-generalization interaction
5.2 What is active now
The frontier is no longer only classical VC-style bounds.
Current emphasis includes:
- overparameterization
- implicit regularization
- distribution shift
- self-supervised and unsupervised settings
- theory for large nonlinear models
5.3 Where it appears
- deep learning theory
- representation learning
- calibration and uncertainty
- model selection and validation
- robustness and shift analysis
5.4 Starter trail
6 Direction 3: Graph Learning Beyond Simple Message Passing
6.1 Stable backbone
- spectral methods
- graph operators and propagation
- embedding geometry
- message passing as local operator application
6.2 What is active now
Graph learning is no longer just about applying a standard GNN layer repeatedly.
Active questions include:
- heterophily and when homophily assumptions fail
- oversquashing and long-range dependence
- rewiring and graph augmentation
- spectral invariances and structure-aware architectures
6.3 Where it appears
- graph neural networks
- molecule and protein learning
- reasoning over relational data
- network science
- structured scientific systems
6.4 Starter trail
7 Direction 4: Generative Modeling Through Score, Flow, And Transport
7.1 Stable backbone
- multivariable calculus
- probability over continuous spaces
- vector fields and local dynamics
- denoising and score estimation
7.2 What is active now
This direction is one of the clearest examples of stable math powering a fast frontier.
Current emphasis includes:
- score-based generation
- reverse-time SDE views
- flow matching
- transport-based formulations
- faster sampling and better trajectory design
7.3 Where it appears
- image, video, and multimodal generation
- scientific generative modeling
- inverse problems
- uncertainty-aware simulation
7.4 Starter trail
8 Direction 5: Optimization Inside Learning And Inference Pipelines
8.1 Stable backbone
- convex sets and convex functions
- duality and certificates
- first-order methods
- constrained optimization
8.2 What is active now
Optimization is increasingly used not just as a solver layer, but as part of the model itself.
Current themes include:
- differentiable optimization layers
- optimization as an inference primitive
- structured prediction and constrained learning
- solver-aware architectures
8.3 Where it appears
- control and robotics
- inverse problems
- constrained ML pipelines
- differentiable programming
- scientific machine learning
8.4 Starter trail
9 Direction 6: Representation Geometry And In-Context Structure
9.1 Stable backbone
- vectors and linear maps
- similarity geometry
- operator viewpoints on attention
- local linearization and feature maps
9.2 What is active now
This direction asks what large models are really doing in representation space.
Typical active questions:
- how should embeddings be interpreted geometrically?
- when do linear probes reveal useful structure versus artifacts?
- why does in-context learning sometimes look like local linear regression or adaptation?
9.3 Where it appears
- transformers
- foundation models
- multimodal learning
- interpretability and diagnostics
- representation analysis
9.4 Starter trail
10 Which Direction Should You Pick?
Use this quick rule.
- choose
high-dimensional probabilityif random fluctuation and matrix behavior keep appearing in what you read - choose
modern learning theoryif you want clearer answers to why models generalize or fail - choose
graph learningif your objects and relations are naturally networked - choose
score / flow / transportif you care about modern generative modeling - choose
optimization in the loopif constrained computation or solver structure is central - choose
representation geometryif your questions keep returning to embeddings, attention, or model internals
If you are unsure, start with the direction whose prerequisites feel least painful. Momentum matters more than perfect taste.
11 What Has Changed Recently
The directions above are not equally active for the same reasons.
Right now, some of the strongest movement is happening where stable math collides with large modern models:
- generalization and implicit bias under scale
- graph learning beyond homophily assumptions
- diffusion, score, and flow viewpoints for generation
- optimization layers and solver-structured models
- representation geometry and in-context behavior
That is why this page is grouped by mathematical direction, not by model brand.
12 What To Learn Next
- Surveys, if you want structured literature entry points after choosing a direction
- Venues, if you want to understand where work in this direction usually gets published
- Paper Lab, if you want reading workflows instead of direction maps
- Applications > Machine Learning, if you want problem-first bridges
13 Sources And Further Reading
- CS229T Course Description -
Paper bridge- concise official map of active machine learning theory themes such as overparameterization, distribution shift, and self-supervision. Checked2026-04-25. - CS224W 2024 -
Paper bridge- current official course hub showing active graph-learning themes, including GNNs and broader graph reasoning. Checked2026-04-25. - EE364a: Convex Optimization I -
Second pass- current official convex optimization entry point, useful for research directions built on certificates, duality, and solver structure. Checked2026-04-25. - High-Dimensional Probability -
Paper bridge- strong current gateway into the probability tools that keep reappearing in modern theory work. Checked2026-04-25. - The Trained Transformer as a Linear Unit -
Paper bridge- representative current paper for the direction connecting in-context behavior and linearized perspectives. Checked2026-04-25. - Understanding Heterophily for Graph Neural Networks -
Paper bridge- representative current paper for the direction beyond simple homophily-based message passing. Checked2026-04-25.