|
Abel , D., et al.2018. Policy and value transfer in lifelong reinforcement learning. In International Conference on Machine Learning, 20–29. |
|
Ammar , H. B., Eaton , E., Ruvolo , P. & Taylor , M.2014. Online multi-task learning for policy gradient methods. In Proceedings of the 31st International Conference on Machine Learning (ICML-14), 1206–1214. |
|
Ammar , H. B., Eaton , E., Ruvolo , P. & Taylor , M. E.2015. Unsupervised cross-domain transfer in policy gradient reinforcement learning via manifold alignment. In Proceedings of AAAI. |
|
Ammar , H. B., Eaton , E., Taylor , M. E., et al.2014. An automated measure of MDP similarity for transfer in reinforcement learning. In Workshops at the Twenty-Eighth AAAI Conference on Artificial Intelligence. |
|
Ammar , H. B., et al.2012. Reinforcement learning transfer via sparse coding. In Proceedings of the 11th International Conference on Autonomous Agents and Multiagent Systems,. 1. International Foundation for Autonomous Agents and Multiagent Systems, 383–390. |
|
Asadi , M. & Huber , M.2007. Effective control knowledge transfer through learning skill and representation hierarchies. In 20th International Joint Conference on Artificial Intelligence, ICML, 2054–2059. |
|
Asadi , M. & Huber , M.2015. A dynamic hierarchical task transfer in multiple robot explorations. In Proceedings on the International Conference on Artificial Intelligence (ICAI), 8, 22–27. |
|
Barreto , A., et al.2017. Successor features for transfer in reinforcement learning. In Advances in Neural Information Processing Systems, 4055–4065. |
|
Ben-David , S., Blitzer , J., Crammer , K., Kulesza , A., et al.2010. A theory of learning from different domains. Machine Learning79(1–2), 151–175. |
|
Ben-David , S., Blitzer , J., Crammer , K. & Pereira , F.2007. Analysis of representations for domain adaptation. In: Advances in Neural Information Processing Systems, 137–144. |
|
Bocsi , B., Csató , L. & Peters , J.2013. Alignment-based transfer learning for robot models. In The 2013 International Joint Conference on Neural Networks (IJCNN), 1–7. IEEE. |
|
Celiberto L. A., Jr., et al.2011. Using cases as heuristics in reinforcement learning: a transfer learning application. In: IJCAI Proceedings-International Joint Conference on Artificial Intelligence22(1), 1211. |
|
Cheng , Q., Wang , X. & Shen , L.2017. An autonomous inter-task mapping learning method via artificial neural network for transfer learning. In 2017 IEEE International Conference on Robotics and Biomimetics (ROBIO), 768–773. IEEE. |
|
Cheng , Q., Wang , X. & Shen , L.2019. A survey on transfer learning for multiagent reinforcement learning systems. Journal of Artificial Intelligence Research64, 645–703. |
|
Da Silva , F. L. & Reali Costa , A. H.2017. Towards zero-shot autonomous inter-task mapping through object-oriented task description. In: Workshop on Transfer in Reinforcement Learning (TiRL). |
|
Da Silva , F. L. & Reali Costa , A. H.2019. A survey on transfer learning for multiagent reinforcement learning systems. Journal of Artificial Intelligence Research64, 645–703. |
|
Da Silva , F. L., Glatt , R. & Reali Costa , A. H.2017. Simultaneously learning and advising in multiagent reinforcement learning. In Proceedings of the 16th Conference on Autonomous Agents and MultiAgent Systems, 1100–1108. International Foundation for Autonomous Agents and Multiagent Systems. |
|
Dabney , W. & Barto , A. G.2012. Adaptive step-size for online temporal difference learning. In Twenty-Sixth AAAI Conference on Artificial Intelligence. |
|
Fachantidis , A., et al.2011. Transfer learning via multiple inter-task mappings. In European Workshop on Reinforcement Learning, 225–236. Springer. |
|
Fachantidis , A., et al.2015. Transfer learning with probabilistic mapping selection. Adaptive Behavior23(1), 3–19. |
|
Ferns , N., Panangaden , P. & Precup , D.2011. Bisimulation metrics for continuous Markov decision processes. SIAM Journal on Computing40(6), 1662–1714. |
|
Ganin , Y. & Lempitsky , V.2014. Unsupervised domain adaptation by backpropagation. arXiv preprintarXiv:1409.7495. |
|
Ganin , Y. & Lempitsky , V. S.2015. Unsupervised domain adaptation by back-propagation. In ICML. |
|
Ganin , Y., Ustinova , E., et al.2016. Domain-adversarial training of neural networks. The Journal of Machine Learning Research17(1), 2096–2030. |
|
Goodfellow , I., et al.2014. Generative adversarial nets. In Advances in Neural Information Processing Systems, 2672–2680. |
|
Hoffman , J., et al.2017. Simultaneous deep transfer across domains and tasks. In Domain Adaptation in Computer Vision Applications, 173–187. Springer. |
|
Konidaris , G. & Barto , A. G.2009. Skill discovery in continuous reinforcement learning domains using skill chaining. In Advances in Neural Information Processing Systems, 1015–1023. |
|
Konidaris , G., Thomas , P., et al.2011. Value function approximation in reinforcement learning using the Fourier basis. In Proceedings of the Twenty-Fifth Conference on Artificial Intelligence, 380–385. |
|
Konidaris , G., et al.2012. Robot learning from demonstration by constructing skill trees. The International Journal of Robotics Research31(3), 360–375. |
|
Lazaric , A.2012. Transfer in reinforcement learning: a framework and a survey. Reinforcement Learning12, 143–173. |
|
Lazaric , A. & Restelli , M.2011. Transfer from multiple MDPs. In Advances in Neural Information Processing Systems, 1746–1754. |
|
Lazaric , A., Restelli , M. & Bonarini , A.2008. Transfer of samples in batch reinforcement learning. In: Proceedings of the 25th International Conference on Machine Learning – ICML 2008, pp. 544–551. ACM Press. |
|
Liu , M.-Y. & Tuzel , O.2016. Coupled generative adversarial networks. In Advances in Neural Information Processing Systems, 469–477. |
|
Mahadevan , S. & Maggioni , M.2007. Proto-value functions: a Laplacian framework for learning representation and control in Markov decision processes. Journal of Machine Learning Research8, 2169–2231, 16. |
|
Moradi , P., et al.2012. Automatic skill acquisition in reinforcement learning using graph centrality measures. Intelligent Data Analysis16, 113–135. |
|
Puterman , M. L.2014. Markov Decision Processes: Discrete Stochastic Dynamic Programming. John Wiley & Sons. |
|
Shoeleh , F. & Asadpour , M.2017. Graph based skill acquisition and transfer Learning for continuous reinforcement learning domains. Pattern Recognition Letters87, 104–116. |
|
Shoeleh , F. & Asadpour , M.2019. Skill based transfer learning with domain adaptation for continuous reinforcement learning domains. Applied Intelligence, 1–17. |
|
Spector , B. & Belongie , S.2018. Sample-effcient reinforcement learning through transfer and architectural priors. arXiv preprintarXiv:1801.02268. |
|
Sutton , R. S. S., Precup , D. & Singh , S.1999. Between MDPs and semi-MDPs: a framework for temporal abstraction in reinforcement learning. Artificial Intelligence112(1–2), 181–211. |
|
Taylor , M. E. & Stone , P.2009. Transfer learning for reinforcement learning domains: a survey. Journal of Machine Learning Research10, 1633–1685. |
|
Taylor , M. E. & Stone , P.2011. An introduction to intertask transfer for reinforcement learning. AI Magazine32(1), 15. |
|
Taylor , M. E., Stone , P. & Liu , Y.2007. Transfer learning via inter-task mappings for temporal difference learning. Journal of Machine Learning Research8, 2125–2167. |
|
Tzeng , E., et al.2017. Adversarial discriminative domain adaptation. Computer Vision and Pattern Recognition (CVPR)1(2), 4. |