Obserwuj
Amir-massoud Farahmand
Amir-massoud Farahmand
Polytechnique Montreal, Mila, University of Toronto
Zweryfikowany adres z cs.toronto.edu - Strona główna
Tytuł
Cytowane przez
Cytowane przez
Rok
Error propagation for approximate policy and value iteration
A Farahmand, C Szepesvári, R Munos
Advances in Neural Information Processing Systems (NeurIPS), 568-576, 2010
2822010
Regularized Policy Iteration
A Farahmand, M Ghavamzadeh, S Mannor, C Szepesvári
Advances in Neural Information Processing Systems 21 (NeurIPS 2008), 441-448, 2009
1682009
Manifold-adaptive dimension estimation
A Farahmand, C Szepesvári, JY Audibert
Proceedings of the 24th International Conference on Machine Learning (ICML …, 2007
1462007
Learning from Limited Demonstrations
B Kim, A Farahmand, J Pineau, D Precup
Advances in Neural Information Processing Systems (NeurIPS), 2859-2867, 2013
1452013
Value-aware loss function for model-based reinforcement learning
A Farahmand, A Barreto, D Nikovski
Artificial Intelligence and Statistics (AISTATS), 1486-1494, 2017
1372017
Regularized policy iteration with nonparametric function spaces
A Farahmand, M Ghavamzadeh, C Szepesvári, S Mannor
Journal of Machine Learning Research (JMLR) 17 (1), 4809-4874, 2016
129*2016
Regularized fitted Q-iteration for planning in continuous-space Markovian decision problems
A Farahmand, M Ghavamzadeh, C Szepesvári, S Mannor
American Control Conference (ACC), 725-730, 2009
97*2009
Robust jacobian estimation for uncalibrated visual servoing
A Shademan, A Farahmand, M Jägersand
IEEE International Conference on Robotics and Automation (ICRA), 5564-5569, 2010
872010
Model Selection in Reinforcement Learning
AM Farahmand, C Szepesvári
Machine learning 85 (3), 299-332, 2011
802011
Iterative Value-Aware Model Learning
A Farahmand
Advances in Neural Information Processing Systems (NeurIPS), 9072-9083, 2018
732018
Action-Gap Phenomenon in Reinforcement Learning
AM Farahmand
Neural Information Processing Systems (NeurIPS), 2011
682011
Deep reinforcement learning for partial differential equation control
A Farahmand, S Nabi, DN Nikovski
American Control Conference (ACC), 3120-3127, 2017
532017
Global visual-motor estimation for uncalibrated visual servoing
A Farahmand, A Shademan, M Jagersand
IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS …, 2007
53*2007
Regularization in Reinforcement Learning
AM Farahmand
Department of Computing Science, University of Alberta, 2011
462011
Attentional network for visual object detection
K Hara, MY Liu, O Tuzel, A Farahmand
arXiv preprint arXiv:1702.01478, 2017
412017
Model-based and model-free reinforcement learning for visual servoing
A Farahmand, A Shademan, M Jagersand, C Szepesvári
IEEE International Conference on Robotics and Automation (ICRA), 2917-2924, 2009
41*2009
Policy-aware model learning for policy gradient methods
R Abachi, M Ghavamzadeh, A Farahmand
arXiv:2003.00030, 2020
392020
Value Gradient weighted Model-Based Reinforcement Learning
CA Voelcker, V Liao, A Garg, A Farahmand
International Conference on Learning Representations (ICLR), 2022
332022
Approximate MaxEnt Inverse Optimal Control and its Application for Mental Simulation of Human Interactions
DA Huang, AM Farahmand, KM Kitani, JA Bagnell
AAAI Conference on Artificial Intelligence (AAAI), 2015
322015
Improving Skin Condition Classification with a Visual Symptom Checker Trained using Reinforcement Learning
M Akrout, A Farahmand, T Jarmain, L Abid
International Conference on Medical Image Computing and Computer Assisted …, 2019
312019
Nie można teraz wykonać tej operacji. Spróbuj ponownie później.
Prace 1–20