SIAM Journal on Mathematics of Data Science

Papers
(The TQCC of SIAM Journal on Mathematics of Data Science is 6. The table below lists those papers that are above that threshold based on CrossRef citation counts [max. 250 papers]. The publications cover those that have been published in the past four years, i.e., from 2022-01-01 to 2026-01-01.)
ArticleCitations
Adaptivity of Stochastic Gradient Methods for Nonconvex Optimization43
A Simple and Optimal Algorithm for Strict Circular Seriation31
Taming Neural Networks with TUSLA: Nonconvex Learning via Adaptive Stochastic Gradient Langevin Algorithms30
Spectral Barron Space for Deep Neural Network Approximation28
On the Inconsistency of Kernel Ridgeless Regression in Fixed Dimensions25
A Note on the Regularity of Images Generated by Convolutional Neural Networks23
Randomized Nyström Approximation of Non-negative Self-Adjoint Operators23
New Equivalences between Interpolation and SVMs: Kernels and Structured Features22
Poisson Reweighted Laplacian Uncertainty Sampling for Graph-Based Active Learning18
Block Majorization Minimization with Extrapolation and Application to \({\beta }\)-NMF17
Resolving the Mixing Time of the Langevin Algorithm to Its Stationary Distribution for Log-Concave Sampling17
Learning Functions Varying along a Central Subspace14
Efficient Algorithms for Regularized Nonnegative Scale-Invariant Low-Rank Approximation Models14
Wassmap: Wasserstein Isometric Mapping for Image Manifold Learning13
Deep Block Proximal Linearized Minimization Algorithm for Nonconvex Inverse Problems13
Online Machine Teaching under Learner Uncertainty: Gradient Descent Learners of a Quadratic Loss13
Nonlinear Tomographic Reconstruction via Nonsmooth Optimization13
Quantitative Approximation Results for Complex-Valued Neural Networks13
Nonbacktracking Spectral Clustering of Nonuniform Hypergraphs12
CA-PCA: Manifold Dimension Estimation, Adapted for Curvature11
Safe Rules for the Identification of Zeros in the Solutions of the SLOPE Problem11
Nonlinear Meta-learning Can Guarantee Faster Rates11
Persistent Laplacians: Properties, Algorithms and Implications11
Function-Space Optimality of Neural Architectures with Multivariate Nonlinearities10
Convergence of a Piggyback-Style Method for the Differentiation of Solutions of Standard Saddle-Point Problems10
Stochastic Variance-Reduced Majorization-Minimization Algorithms10
Scalable Tensor Methods for Nonuniform Hypergraphs10
The GenCol Algorithm for High-Dimensional Optimal Transport: General Formulation and Application to Barycenters and Wasserstein Splines9
Asymptotics of the Sketched Pseudoinverse9
The Sample Complexity of Sparse Multireference Alignment and Single-Particle Cryo-Electron Microscopy9
Inverse Evolution Layers: Physics-Informed Regularizers for Image Segmentation9
Covariance Alignment: From Maximum Likelihood Estimation to Gromov–Wasserstein9
A Variational Formulation of Accelerated Optimization on Riemannian Manifolds9
Group-Invariant Tensor Train Networks for Supervised Learning9
Bi-Invariant Dissimilarity Measures for Sample Distributions in Lie Groups8
Finite-Time Analysis of Natural Actor-Critic for POMDPs8
The Geometric Median and Applications to Robust Mean Estimation8
On Neural Network Approximation of Ideal Adversarial Attack and Convergence of Adversarial Training8
Random Multitype Spanning Forests for Synchronization on Sparse Graphs8
Convergence of Gradient Descent for Recurrent Neural Networks: A Nonasymptotic Analysis7
Supervised Gromov–Wasserstein Optimal Transport with Metric-Preserving Constraints7
ABBA Neural Networks: Coping with Positivity, Expressivity, and Robustness7
Optimal Dorfman Group Testing for Symmetric Distributions7
Benefit of Interpolation in Nearest Neighbor Algorithms7
LASSO Reloaded: A Variational Analysis Perspective with Applications to Compressed Sensing7
Post-training Quantization for Neural Networks with Provable Guarantees7
A Nonlinear Matrix Decomposition for Mining the Zeros of Sparse Data7
Numerical Considerations and a new implementation for invariant coordinate selection7
Computing Wasserstein Barycenters via Operator Splitting: The Method of Averaged Marginals7
Efficient Identification of Butterfly Sparse Matrix Factorizations7
Operator Shifting for General Noisy Matrix Systems6
Adaptive Joint Distribution Learning6
Complete and Continuous Invariants of 1-Periodic Sequences in Polynomial Time6
Robust Classification Under $\ell_0$ Attack for the Gaussian Mixture Model6
Phase Retrieval with Semialgebraic and ReLU Neural Network Priors6
Randomized Wasserstein Barycenter Computation: Resampling with Statistical Guarantees6
Memory Capacity of Two Layer Neural Networks with Smooth Activations6
0.037112951278687