65 results on '"Martha White"'
Search Results
2. Demystifying the Recency Heuristic in Temporal-Difference Learning.
3. Investigating the Interplay of Prioritized Replay and Generalization.
4. Position: Benchmarking is Limited in Reinforcement Learning Research.
5. Averaging n-step Returns Reduces Variance in Reinforcement Learning.
6. Measuring and Mitigating Interference in Reinforcement Learning.
7. Asymptotically Unbiased Off-Policy Policy Evaluation when Reusing Old Data in Nonstationary Environments.
8. Trajectory-Aware Eligibility Traces for Off-Policy Reinforcement Learning.
9. Exploiting Action Impact Regularity and Exogenous State Variables for Offline Reinforcement Learning (Abstract Reprint).
10. An Alternate Policy Gradient Estimator for Softmax Policies.
11. Understanding and mitigating the limitations of prioritized experience replay.
12. A Temporal-Difference Approach to Policy Gradient Estimation.
13. Greedy Actor-Critic: A New Conditional Cross-Entropy Method for Policy Improvement.
14. The In-Sample Softmax for Offline Reinforcement Learning.
15. General Munchausen Reinforcement Learning with Tsallis Kullback-Leibler Divergence.
16. Structural Credit Assignment in Neural Networks using Reinforcement Learning.
17. Continual Auxiliary Task Learning.
18. From Language to Language-ish: How Brain-Like is an LSTM's Representation of Atypical Language Stimuli?
19. Maximizing Information Gain in Partially Observable Environments via Prediction Rewards.
20. Gradient Temporal-Difference Learning with Regularized Corrections.
21. Optimizing for the Future in Non-Stationary MDPs.
22. Selective Dyna-Style Planning Under Limited Model Capacity.
23. Resonance in Weight Space: Covariate Shift Can Drive Divergence of SGD with Momentum.
24. Meta-Learning Representations for Continual Learning.
25. Importance Resampling for Off-policy Prediction.
26. Learning Macroscopic Brain Connectomes via Group-Sparse Factorization.
27. Planning with Expectation Models.
28. Hill Climbing on Value Estimates for Search-control in Dyna.
29. The Utility of Sparse Representations for Control in Reinforcement Learning.
30. Meta-Descent for Online, Continual Prediction.
31. Fuzzy Tiling Activations: A Simple Approach to Learning Sparse Representations Online.
32. Training Recurrent Neural Networks Online by Learning Explicit State Variables.
33. Maxmin Q-learning: Controlling the Estimation Bias of Q-learning.
34. Towards Safe Policy Improvement for Non-Stationary MDPs.
35. An implicit function learning approach for parametric modal regression.
36. An Off-policy Policy Gradient Theorem Using Emphatic Weightings.
37. Context-dependent upper-confidence bounds for directed exploration.
38. Supervised autoencoders: Improving generalization performance with unsupervised regularizers.
39. Organizing Experience: a Deeper Look at Replay Mechanisms for Sample-Based Planning in Continuous State Domains.
40. High-confidence error estimates for learned value functions.
41. Comparing Direct and Indirect Temporal-Difference Methods for Estimating the Variance of the Return.
42. Improving Regression Performance with Distributional Losses.
43. Reinforcement Learning with Function-Valued Action Spaces for Partial Differential Equation Control.
44. Multi-view Matrix Factorization for Linear Dynamical System Estimation.
45. Learning Sparse Representations in Reinforcement Learning with Sparse Coding.
46. Unifying Task Specification in Reinforcement Learning.
47. Adapting Kernel Representations Online Using Submodular Maximization.
48. Accelerated Gradient Temporal Difference Learning.
49. Recovering True Classifier Performance in Positive-Unlabeled Learning.
50. Two-Timescale Networks for Nonlinear Value Function Approximation.
Catalog
Books, media, physical & digital resources
Discovery Service for Jio Institute Digital Library
For full access to our library's resources, please sign in.