Reinforcement learning: A survey LP Kaelbling, ML Littman, AW Moore Journal of artificial intelligence research 4, 237-285, 1996 | 12011 | 1996 |
Planning and acting in partially observable stochastic domains LP Kaelbling, ML Littman, AR Cassandra Artificial intelligence 101 (1-2), 99-134, 1998 | 5806 | 1998 |
Learning policies for partially observable environments: Scaling up ML Littman, AR Cassandra, LP Kaelbling Machine Learning Proceedings 1995, 362-370, 1995 | 1032 | 1995 |
Acting optimally in partially observable stochastic domains AR Cassandra, LP Kaelbling, ML Littman Aaai 94, 1023-1028, 1994 | 1025 | 1994 |
Learning in embedded systems LP Kaelbling MIT press, 1993 | 964 | 1993 |
Hierarchical task and motion planning in the now LP Kaelbling, T Lozano-Pérez 2011 IEEE International Conference on Robotics and Automation, 1470-1477, 2011 | 821 | 2011 |
Acting under uncertainty: Discrete Bayesian models for mobile-robot navigation AR Cassandra, LP Kaelbling, JA Kurien Proceedings of IEEE/RSJ International Conference on Intelligent Robots and …, 1996 | 770 | 1996 |
On the complexity of solving Markov decision problems ML Littman, TL Dean, LP Kaelbling arXiv preprint arXiv:1302.4971, 2013 | 750 | 2013 |
To transfer or not to transfer MT Rosenstein, Z Marx, LP Kaelbling, TG Dietterich NIPS 2005 workshop on transfer learning 898 (3), 4, 2005 | 743 | 2005 |
Generalization in deep learning K Kawaguchi, LP Kaelbling, Y Bengio arXiv preprint arXiv:1710.05468 1 (8), 2017 | 600 | 2017 |
Effective reinforcement learning for mobile robots WD Smart, LP Kaelbling Proceedings 2002 IEEE International Conference on Robotics and Automation …, 2002 | 572 | 2002 |
An architecture for intelligent reactive systems LP Kaelbling Reasoning about actions and plans, 395-410, 1987 | 525 | 1987 |
The synthesis of digital machines with provable epistemic properties SJ Rosenschein, LP Kaelbling Theoretical aspects of reasoning about knowledge, 83-98, 1986 | 503 | 1986 |
Integrated task and motion planning CR Garrett, R Chitnis, R Holladay, B Kim, T Silver, LP Kaelbling, ... Annual review of control, robotics, and autonomous systems 4 (1), 265-293, 2021 | 483 | 2021 |
Learning to achieve goals LP Kaelbling IJCAI 2, 1094-8, 1993 | 472 | 1993 |
Integrated task and motion planning in belief space LP Kaelbling, T Lozano-Pérez The International Journal of Robotics Research 32 (9-10), 1194-1227, 2013 | 467 | 2013 |
Belief space planning assuming maximum likelihood observations. R Platt Jr, R Tedrake, LP Kaelbling, T Lozano-Perez Robotics: Science and systems 2, 2010 | 433 | 2010 |
Input Generalization in Delayed Reinforcement Learning: An Algorithm and Performance Comparisons. D Chapman, LP Kaelbling Ijcai 91, 726-731, 1991 | 423 | 1991 |
Hierarchical solution of Markov decision processes using macro-actions M Hauskrecht, N Meuleau, LP Kaelbling, TL Dean, C Boutilier arXiv preprint arXiv:1301.7381, 2013 | 409 | 2013 |
Learning to cooperate via policy search L Peshkin, KE Kim, N Meuleau, LP Kaelbling arXiv preprint cs/0105032, 2001 | 392 | 2001 |