Two Timescale Stochastic Approximation with Controlled Markov noise and Off-policy temporal difference learning
March 31, 2015 ยท Declared Dead ยท ๐ Mathematics of Operations Research
"No code URL or promise found in abstract"
Evidence collected by the PWNC Scanner
Authors
Prasenjit Karmakar, Shalabh Bhatnagar
arXiv ID
1503.09105
Category
math.DS
Cross-listed
cs.AI,
stat.ML
Citations
28
Venue
Mathematics of Operations Research
Last Checked
1 month ago
Abstract
We present for the first time an asymptotic convergence analysis of two time-scale stochastic approximation driven by `controlled' Markov noise. In particular, both the faster and slower recursions have non-additive controlled Markov noise components in addition to martingale difference noise. We analyze the asymptotic behavior of our framework by relating it to limiting differential inclusions in both time-scales that are defined in terms of the ergodic occupation measures associated with the controlled Markov processes. Finally, we present a solution to the off-policy convergence problem for temporal difference learning with linear function approximation, using our results.
Community Contributions
Found the code? Know the venue? Think something is wrong? Let us know!
๐ Similar Papers
In the same crypt โ math.DS
R.I.P.
๐ป
Ghosted
R.I.P.
๐ป
Ghosted
Linearly-Recurrent Autoencoder Networks for Learning Dynamics
R.I.P.
๐ป
Ghosted
Gradient Descent Only Converges to Minimizers: Non-Isolated Critical Points and Invariant Regions
R.I.P.
๐ป
Ghosted
Eigendecompositions of Transfer Operators in Reproducing Kernel Hilbert Spaces
R.I.P.
๐ป
Ghosted
From rate distortion theory to metric mean dimension: variational principle
R.I.P.
๐ป
Ghosted
Double variational principle for mean dimension
Died the same way โ ๐ป Ghosted
R.I.P.
๐ป
Ghosted
Language Models are Few-Shot Learners
R.I.P.
๐ป
Ghosted
PyTorch: An Imperative Style, High-Performance Deep Learning Library
R.I.P.
๐ป
Ghosted
XGBoost: A Scalable Tree Boosting System
R.I.P.
๐ป
Ghosted