Search

Your search keyword '"Telgarsky, Matus"' showing total 142 results

Search Constraints

Start Over You searched for: Author "Telgarsky, Matus" Remove constraint Author: "Telgarsky, Matus"
142 results on '"Telgarsky, Matus"'

Search Results

1. Benefits of Early Stopping in Gradient Descent for Overparameterized Logistic Regression

2. One-layer transformers fail to solve the induction heads task

3. Spectrum Extraction and Clipping for Implicitly Linear Layers

4. Large Stepsize Gradient Descent for Logistic Loss: Non-Monotonicity of the Loss Improves Optimization Efficiency

5. Transformers, parallel computation, and logarithmic depth

6. On Achieving Optimal Adversarial Test Error

7. Representational Strengths and Limitations of Transformers

8. Feature selection with gradient descent on two-layer networks in low-rotation regimes

9. Convex Analysis at Infinity: An Introduction to Astral Space

10. Stochastic linear optimization never overfits with quadratically-bounded losses on general data

11. Actor-critic is implicitly biased towards high entropy optimal policies

12. Fast Margin Maximization via Dual Acceleration

13. Early-stopped neural networks are consistent

14. Generalization bounds via distillation

15. Gradient descent follows the regularization path for general losses

16. Directional convergence and alignment in deep learning

17. Neural tangent kernels, transportation mappings, and universal approximation

18. Polylogarithmic width suffices for gradient descent to achieve arbitrarily small test error with shallow ReLU networks

19. Approximation power of random neural networks

20. Characterizing the implicit bias via a primal-dual analysis

21. A gradual, semi-discrete approach to generative network training via explicit Wasserstein minimization

22. Size-Noise Tradeoffs in Generative Networks

23. Gradient descent aligns the layers of deep linear networks

24. Risk and parameter convergence of logistic regression

25. Social welfare and profit maximization from revealed preferences

26. Spectrally-normalized margin bounds for neural networks

27. Neural networks and rational functions

28. Non-convex learning via Stochastic Gradient Langevin Dynamics: a nonasymptotic analysis

29. Greedy bi-criteria approximations for $k$-medians and $k$-means

30. Benefits of depth in neural networks

31. Rate of Price Discovery in Iterative Combinatorial Auctions

32. Representation Benefits of Deep Feedforward Networks

33. Convex Risk Minimization and Conditional Probability Estimation

34. Scalable Nonlinear Learning with Adaptive Polynomial Expansions

35. Moment-based Uniform Deviation Bounds for $k$-means and Friends

36. Boosting with the Logistic Loss is Consistent

37. Margins, Shrinkage, and Boosting

38. Dirichlet draws are sparse with high probability

39. Tensor decompositions for learning latent variable models

40. Agglomerative Bregman Clustering

41. Statistical Consistency of Finite-dimensional Unregularized Linear Classification

42. Blackwell Approachability and Minimax Theory

43. A Primal-Dual Convergence Analysis of Boosting

44. Central Binomial Tail Bounds

45. Social Welfare and Profit Maximization from Revealed Preferences

46. Tensor Decompositions for Learning Latent Variable Models

47. Tensor Decompositions for Learning Latent Variable Models (A Survey for ALT)

49. Duality and Data Dependence in Boosting /

Catalog

Books, media, physical & digital resources