Search

Your search keyword '"Riekert, Adrian"' showing total 27 results

Search Constraints

Start Over You searched for: Author "Riekert, Adrian" Remove constraint Author: "Riekert, Adrian"
27 results on '"Riekert, Adrian"'

Search Results

1. An Overview on Machine Learning Methods for Partial Differential Equations: from Physics Informed Neural Networks to Deep Operator Learning

2. Learning rate adaptive stochastic gradient descent optimization methods: numerical simulations for deep learning methods for partial differential equations and convergence analyses

4. Non-convergence to global minimizers for Adam and stochastic gradient descent optimization and constructions of local minimizers in the training of artificial neural networks

6. Deep neural network approximation of composite functions without the curse of dimensionality

7. Algorithmically Designed Artificial Neural Networks (ADANNs): Higher order deep operator learning for parametric partial differential equations

8. Convergence to good non-optimal critical points in the training of neural networks: Gradient descent optimization with one random initialization overcomes all bad non-global local minima with high probability

9. Normalized gradient flow optimization in the training of ReLU artificial neural networks

10. On the existence of infinitely many realization functions of non-global local minima in the training of artificial neural networks with ReLU activation

11. On the existence of global minima and convergence analyses for gradient descent methods in the training of deep neural networks

12. Convergence proof for stochastic gradient descent in the training of deep neural networks with ReLU activation for constant target functions

13. Existence, uniqueness, and convergence rates for gradient flows in the training of artificial neural networks with ReLU activation

14. A proof of convergence for the gradient descent optimization method with random initializations in the training of neural networks with ReLU activation for piecewise linear target functions

15. Convergence analysis for gradient flows in the training of artificial neural networks with ReLU activation

16. A proof of convergence for stochastic gradient descent in the training of artificial neural networks with ReLU activation for constant target functions

17. A proof of convergence for gradient descent in the training of artificial neural networks for constant target functions

18. Convergence Rates for Empirical Measures of Markov Chains in Dual and Wasserstein Distances

19. Strong overall error analysis for the training of artificial neural networks via random initializations

27. A proof of convergence for the gradient descent optimization method with random initializations in the training of neural networks with ReLU activation for piecewise linear target functions.

Catalog

Books, media, physical & digital resources