Apprenticeship learning via inverse reinforcement learning, Twenty-first international conference on Machine learning , ICML '04, 2004. ,
DOI : 10.1145/1015330.1015430
URL : http://citeseerx.ist.psu.edu/viewdoc/summary?doi=10.1.1.3.6759
Concurrent probabilistic temporal planning with policy-gradients, International Conference on Automated Planning and Scheduling, 2007. ,
Preferencebased policy learning, Joint European Conference on Machine Learning and Knowledge Discovery in Databases, 2011. ,
URL : https://hal.archives-ouvertes.fr/inria-00625001
Planning for decentralized control of multiple robots under uncertainty, 2015 IEEE International Conference on Robotics and Automation (ICRA), 2015. ,
DOI : 10.1109/ICRA.2015.7139350
Planning with macro-actions in decentralized pomdps, International Conference on Autonomous Agents and Multiagent Systems, 2014. ,
A survey of robot learning from demonstration, Robotics and Autonomous Systems, vol.57, issue.5, pp.469-483, 2009. ,
DOI : 10.1016/j.robot.2008.10.024
Planning with concurrency under resources and time uncertainty, European Conference on Artificial Intelligence, 2010. ,
Top-down induction of firstorder logical decision trees, Artificial intelligence, vol.101, issue.1, 1998. ,
DOI : 10.1016/s0004-3702(98)00034-4
URL : http://doi.org/10.1016/s0004-3702(98)00034-4
Classification and regression trees, 1984. ,
and others, 2012. A survey of monte carlo tree search methods, Transactions on Computational Intelligence and AI in Games ,
The factored policy-gradient planner, Artificial Intelligence, vol.173, issue.5-6, pp.5-6, 2009. ,
DOI : 10.1016/j.artint.2008.11.008
URL : https://hal.archives-ouvertes.fr/inria-00330031
User simulation in dialogue systems using inverse reinforcement learning, Interspeech, 2011. ,
URL : https://hal.archives-ouvertes.fr/hal-00652446
Interactive policy learning through confidence-based autonomy, Journal of Artificial Intelligence Research, vol.34, issue.1, 2009. ,
Logical and relational learning, 2008. ,
DOI : 10.1007/978-3-540-68856-3
Relational reinforcement learning, Machine learning, vol.43, issue.12, 2001. ,
DOI : 10.1007/BFb0027307
Hybrid Controllers for Path Planning: A Temporal Logic Approach, Proceedings of the 44th IEEE Conference on Decision and Control, 2005. ,
DOI : 10.1109/CDC.2005.1582935
URL : http://citeseerx.ist.psu.edu/viewdoc/summary?doi=10.1.1.117.4410
Greedy function approximation: a gradient boosting machine, Annals of statistics, vol.29, issue.5, 2001. ,
Extremely randomized trees, Machine Learning, vol.63, issue.1, 2006. ,
DOI : 10.1007/s10994-006-6226-1
URL : https://hal.archives-ouvertes.fr/hal-00341932
Decision-making authority, team efficiency and human worker satisfaction in mixed human?robot teams, Autonomous Robots, vol.39, issue.3, 2015. ,
DOI : 10.15607/rss.2014.x.046
Dogged Learning for Robots, Proceedings 2007 IEEE International Conference on Robotics and Automation, 2007. ,
DOI : 10.1109/ROBOT.2007.363692
URL : http://citeseerx.ist.psu.edu/viewdoc/summary?doi=10.1.1.179.3071
Reducing the Time Complexity of the Derandomized Evolution Strategy with Covariance Matrix Adaptation (CMA-ES), Evolutionary Computation, vol.11, issue.1, 2003. ,
DOI : 10.1162/106365601750190398
Learning Generalizable Control Programs, IEEE Transactions on Autonomous Mental Development, vol.3, issue.3, 2011. ,
DOI : 10.1109/TAMD.2010.2103311
Learning preferences for manipulation tasks from online coactive feedback, The International Journal of Robotics Research, vol.34, issue.10, p.34, 2015. ,
DOI : 10.1177/0278364915581193
Learning trajectory preferences for manipulators via iterative improvement, Conference on Neural Information Processing Systems, 2013. ,
Trial-based heuristic tree search for finite horizon MDPs, International Conference on Automated Planning and Scheduling, 2013. ,
Bellman goes relational, Twenty-first international conference on Machine learning , ICML '04, 2004. ,
DOI : 10.1145/1015330.1015401
URL : http://citeseerx.ist.psu.edu/viewdoc/summary?doi=10.1.1.1.8073
Learning action strategies for planning domains, Artificial Intelligence, vol.113, issue.1-2, 1999. ,
DOI : 10.1016/S0004-3702(99)00060-0
URL : http://doi.org/10.1016/s0004-3702(99)00060-0
Inverse reinforcement learning through structured classification, Conference on Neural Information Processing Systems, 2012. ,
URL : https://hal.archives-ouvertes.fr/hal-00778624
A Cascaded Supervised Learning Approach to Inverse Reinforcement Learning, Joint European Conference on Machine Learning and Knowledge Discovery in Databases, 2013. ,
DOI : 10.1007/978-3-642-40988-2_1
URL : https://hal.archives-ouvertes.fr/hal-00869804
Combining manual feedback with subsequent mdp reward signals for reinforcement learning, International Conference on Autonomous Agents and Multiagent Systems, 2010. ,
Training a Robot via Human Feedback: A Case Study, International Conference on Social Robotics, 2013. ,
DOI : 10.1007/978-3-319-02675-6_46
URL : http://citeseerx.ist.psu.edu/viewdoc/summary?doi=10.1.1.378.2844
Learning motor primitives for robotics, 2009 IEEE International Conference on Robotics and Automation, 2009. ,
DOI : 10.1109/ROBOT.2009.5152577
URL : http://citeseerx.ist.psu.edu/viewdoc/summary?doi=10.1.1.165.8287
Bandit Based Monte-Carlo Planning, European Conference on Machine Learning, 2006. ,
DOI : 10.1007/11871842_29
URL : http://citeseerx.ist.psu.edu/viewdoc/summary?doi=10.1.1.102.1296
Anticipatory Planning for Human-Robot Teams, International Symposium on Experimental Robotics, 2016. ,
DOI : 10.1109/ICRA.2013.6631293
Planning with noisy probabilistic relational rules, Journal of Artificial Intelligence Research, vol.39, issue.1, 2010. ,
Exploration in relational domains for model-based reinforcement learning, The Journal of Machine Learning Research, vol.13, issue.1, 2012. ,
Personalization in HRI, Proceedings of the seventh annual ACM/IEEE international conference on Human-Robot Interaction, HRI '12 ,
DOI : 10.1145/2157689.2157804
A Computational Model of Social-Learning Mechanisms, Adaptive Behavior, vol.17, issue.6, p.17, 2009. ,
DOI : 10.1177/1059712309342757
Active Learning for Reward Estimation in Inverse Reinforcement Learning, Joint European Conference on Machine Learning and Knowledge Discovery in Databases, 2009. ,
DOI : 10.1007/978-3-642-04174-7_3
Adaptive movement sequences and predictive decisions based on hierarchical dynamical systems, 2012 IEEE/RSJ International Conference on Intelligent Robots and Systems, 2012. ,
DOI : 10.1109/IROS.2012.6385651
URL : http://citeseerx.ist.psu.edu/viewdoc/summary?doi=10.1.1.347.6270
Query learning strategies using boosting and bagging, International Conference on Machine Learning, 1998. ,
Robot self-initiative and personalization by learning through repeated interactions, Proceedings of the 6th international conference on Human-robot interaction, HRI '11, 2011. ,
DOI : 10.1145/1957656.1957814
URL : https://hal.archives-ouvertes.fr/hal-00636164
Planning with durative actions in stochastic domains, Journal of Artificial Intelligence Research, p.31, 2008. ,
Adapting Robot Behavior for Human--Robot Interaction, IEEE Transactions on Robotics, vol.24, issue.4, 2008. ,
DOI : 10.1109/TRO.2008.926867
URL : http://citeseerx.ist.psu.edu/viewdoc/summary?doi=10.1.1.475.7765
Impact of Robot Initiative on Human-Robot Collaboration, Proceedings of the Companion of the 2017 ACM/IEEE International Conference on Human-Robot Interaction , HRI '17, 2017. ,
DOI : 10.1145/3029798.3038373
Inverse reinforcement learning in relational domains, International Joint Conference on Artificial Intelligence, 2015. ,
URL : https://hal.archives-ouvertes.fr/hal-01154650
Preference learning on the execution of collaborative human-robot tasks, International Conference on Robotics and Automation, 2017. ,
Imitation learning in relational domains: A functional-gradient boosting approach, International Joint Conference on Artificial Intelligence, 2011. ,
Gradient-based boosting for statistical relational learning: The relational dependency network case, Machine Learning, 2012. ,
DOI : 10.1007/s10994-011-5244-9
Multi-Agent Inverse Reinforcement Learning, 2010 Ninth International Conference on Machine Learning and Applications, 2010. ,
DOI : 10.1109/ICMLA.2010.65
URL : http://citeseerx.ist.psu.edu/viewdoc/summary?doi=10.1.1.224.5667
Apprenticeship learning using inverse reinforcement learning and gradient methods, In Uncertainty in Artificial Intelligence, 2007. ,
Training parsers by inverse reinforcement learning, Machine Learning, vol.285, issue.5, 2009. ,
DOI : 10.1007/s10994-009-5110-1
URL : http://citeseerx.ist.psu.edu/viewdoc/summary?doi=10.1.1.150.3712
Policy invariance under reward transformations: Theory and application to reward shaping, International Conference on Machine Learning, 1999. ,
Algorithms for inverse reinforcement learning, International Conference on Machine Learning, 2000. ,
Incremental Semantically Grounded Learning from Demonstration, Robotics: Science and Systems IX, 2013. ,
DOI : 10.15607/RSS.2013.IX.048
Efficient model learning for human-robot collaborative tasks. arXiv preprint, 2014. ,
DOI : 10.1145/2696454.2696455
Human-robot cross-training: Computational formulation, modeling and evaluation of a human team training strategy, 2013 8th ACM/IEEE International Conference on Human-Robot Interaction (HRI), 2013. ,
DOI : 10.1109/HRI.2013.6483499
Learning from Demonstrations: Is It Worth Estimating a Reward Function?, Joint European Conference on Machine Learning and Knowledge Discovery in Databases, 2013. ,
DOI : 10.1007/978-3-642-40988-2_2
URL : https://hal.archives-ouvertes.fr/hal-00916938
Markov Decision Processes: Discrete Stochastic Dynamic Programming, 1994. ,
DOI : 10.1002/9780470316887
C4. 5: Programs for Machine Learning, 1993. ,
Learning to take concurrent actions, Conference on Neural Information Processing Systems, 2002. ,
Coarticulation, Proceedings of the 22nd international conference on Machine learning , ICML '05, 2005. ,
DOI : 10.1145/1102351.1102442
Learning agents for uncertain environments, Conference on Learning Theory, 1998. ,
DOI : 10.1145/279943.279964
URL : http://citeseerx.ist.psu.edu/viewdoc/summary?doi=10.1.1.152.6795
Relational dynamic influence diagram language (rddl): Language description, 2010. ,
Is imitation learning the route to humanoid robots? Trends in cognitive sciences, 1999. ,
Explanation-based manipulator learning: Acquisition of planning ability through observation, Proceedings. 1985 IEEE International Conference on Robotics and Automation, 1985. ,
DOI : 10.1109/ROBOT.1985.1087311
Bagger: an ebl system that extends and generalizes explanations, AAAI Conference on Artificial Intelligence, 1987. ,
Online structured prediction via coactive learning, 2012. ,
Temporal planning with mutual exclusion reasoning, International Joint Conference on Artificial Intelligence, 1999. ,
Teachable robots: Understanding human teaching behavior to build more effective robot learners, Artificial Intelligence, vol.172, issue.6-7, 2008. ,
DOI : 10.1016/j.artint.2007.09.009
URL : http://doi.org/10.1016/j.artint.2007.09.009
Relational activity processes for modeling concurrent cooperation, 2016 IEEE International Conference on Robotics and Automation (ICRA) ,
DOI : 10.1109/ICRA.2016.7487765
URL : https://hal.archives-ouvertes.fr/hal-01399247
Learning from delayed rewards, 1989. ,
Optimization of Temporal Dynamics for Adaptive Human-Robot Interaction in Assembly Manufacturing, Robotics: Science and Systems VIII, 2012. ,
DOI : 10.15607/RSS.2012.VIII.056
Inductive policy selection for first-order mdps, Uncertainty in Artificial Intelligence, 2002. ,
Policy generation for continuoustime stochastic domains with concurrency, International Conference on Automated Planning and Scheduling, 2004. ,
Learning planning rules in noisy stochastic worlds, AAAI Conference on Artificial Intelligence, 2005. ,
Maximum entropy inverse reinforcement learning, AAAI Conference on Artificial Intelligence, 2008. ,