Obserwuj
Nan Jiang
Nan Jiang
Assistant Professor of Computer Science, UIUC
Zweryfikowany adres z illinois.edu - Strona główna
Tytuł
Cytowane przez
Cytowane przez
Rok
Doubly Robust Off-policy Value Evaluation for Reinforcement Learning
N Jiang, L Li
Proceedings of the 33rd International Conference on Machine Learning (ICML-16), 2015
7142015
Contextual Decision Processes with Low Bellman Rank are PAC-Learnable
N Jiang, A Krishnamurthy, A Agarwal, J Langford, RE Schapire
Proceedings of the 34th International Conference on Machine Learning (ICML-17), 2016
3742016
Information-Theoretic Considerations in Batch Reinforcement Learning
J Chen, N Jiang
Proceedings of the 36th International Conference on Machine Learning (ICML …, 2019
2862019
Provably efficient RL with Rich Observations via Latent State Decoding
SS Du, A Krishnamurthy, N Jiang, A Agarwal, M Dudík, J Langford
Proceedings of the 36th International Conference on Machine Learning (ICML …, 2019
2052019
Model-based RL in Contextual Decision Processes: PAC bounds and Exponential Improvements over Model-free Approaches
W Sun, N Jiang, A Krishnamurthy, A Agarwal, J Langford
Conference on Learning Theory, 2019
2002019
Hierarchical Imitation and Reinforcement Learning
HM Le, N Jiang, A Agarwal, M Dudík, Y Yue, H Daumé III
Proceedings of the 35th International Conference on Machine Learning (ICML-18), 2018
1782018
Reinforcement Learning: Theory and Algorithms
A Agarwal, N Jiang, SM Kakade
1632019
Bellman-consistent pessimism for offline reinforcement learning
T Xie, CA Cheng, N Jiang, P Mineiro, A Agarwal
Advances in neural information processing systems 34, 6683-6694, 2021
1512021
Minimax Weight and Q-Function Learning for Off-Policy Evaluation
M Uehara, J Huang, N Jiang
arXiv preprint arXiv:1910.12809, 2019
1472019
The Dependence of Effective Planning Horizon on Model Accuracy
N Jiang, A Kulesza, S Singh, R Lewis
Proceedings of the 2015 International Conference on Autonomous Agents and …, 2015
1382015
Sample complexity of reinforcement learning using linearly combined model ensembles
A Modi, N Jiang, A Tewari, S Singh
International Conference on Artificial Intelligence and Statistics, 2010-2020, 2020
1232020
Empirical study of off-policy policy evaluation for reinforcement learning
C Voloshin, HM Le, N Jiang, Y Yue
arXiv preprint arXiv:1911.06854, 2019
1022019
Policy finetuning: Bridging sample-efficient offline and online reinforcement learning
T Xie, N Jiang, H Wang, C Xiong, Y Bai
Advances in neural information processing systems 34, 27395-27407, 2021
972021
On Oracle-Efficient PAC Reinforcement Learning with Rich Observations
C Dann, N Jiang, A Krishnamurthy, A Agarwal, J Langford, RE Schapire
Advances in Neural Information Processing Systems, 2018, 2018
972018
Batch value-function approximation with only realizability
T Xie, N Jiang
International Conference on Machine Learning, 11404-11413, 2021
852021
Provably efficient q-learning with low switching cost
Y Bai, T Xie, N Jiang, YX Wang
Advances in Neural Information Processing Systems, 8004-8013, 2019
812019
Abstraction Selection in Model-based Reinforcement Learning
N Jiang, A Kulesza, S Singh
Proceedings of the 32nd International Conference on Machine Learning (ICML …, 2015
772015
Q* approximation schemes for batch reinforcement learning: A theoretical comparison
T Xie, N Jiang
Conference on Uncertainty in Artificial Intelligence, 550-559, 2020
682020
Repeated Inverse Reinforcement Learning
K Amin, N Jiang, S Singh
Advances in Neural Information Processing Systems, 2017, 2017
682017
Model-free Representation Learning and Exploration in Low-rank MDPs
A Modi, J Chen, A Krishnamurthy, N Jiang, A Agarwal
arXiv preprint arXiv:2102.07035, 2021
642021
Nie można teraz wykonać tej operacji. Spróbuj ponownie później.
Prace 1–20