Show simple item record

dc.contributor.authorAkbulut, M. T.
dc.contributor.authorÖztop, Erhan
dc.contributor.authorXue, H.
dc.contributor.authorTekden, A. E.
dc.contributor.authorŞeker, M. Y.
dc.contributor.authorUğur, E.
dc.date.accessioned2024-03-06T04:58:52Z
dc.date.available2024-03-06T04:58:52Z
dc.date.issued2020
dc.identifier.issn2640-3498en_US
dc.identifier.urihttp://hdl.handle.net/10679/9264
dc.identifier.urihttps://proceedings.mlr.press/v155/akbulut21a.html
dc.description.abstractTo equip robots with dexterous skills, an effective approach is to first transfer the desired skill via Learning from Demonstration (LfD), then let the robot improve it by self-exploration via Reinforcement Learning (RL). In this paper, we propose a novel LfD+RL framework, namely Adaptive Conditional Neural Movement Primitives (ACNMP), that allows efficient policy improvement in novel environments and effective skill transfer between different agents. This is achieved through exploiting the latent representation learned by the underlying Conditional Neural Process (CNP) model, and simultaneous training of the model with supervised learning (SL) for acquiring the demonstrated trajectories and via RL for new trajectory discovery. Through simulation experiments, we show that (i) ACNMP enables the system to extrapolate to situations where pure LfD fails; (ii) Simultaneous training of the system through SL and RL preserves the shape of demonstrations while adapting to novel situations due to the shared representations used by both learners; (iii) ACNMP enables order-of-magnitude sample-efficient RL in extrapolation of reaching tasks compared to the existing approaches; (iv) ACNMPs can be used to implement skill transfer between robots having different morphology, with competitive learning speeds and importantly with less number of assumptions compared to the state-of-the-art approaches. Finally, we show the real-world suitability of ACNMPs through real robot experiments that involve obstacle avoidance, pick and place and pouring actions.en_US
dc.description.sponsorshipHorizon 2020 Framework Programme ; Core Research for Evolutional Science and Technology ; Osaka University ; TÜBİTAK
dc.language.isoengen_US
dc.publisherML Research Pressen_US
dc.relationinfo:eu-repo/grantAgreement/EC/H2020/731761
dc.relation.ispartofProceedings of Machine Learning Research
dc.rightsopenAccess
dc.titleACNMP: skill transfer and task extrapolation through learning from demonstration and reinforcement learning via representation sharingen_US
dc.typeConference paperen_US
dc.publicationstatusPublisheden_US
dc.contributor.departmentÖzyeğin University
dc.contributor.authorID(ORCID 0000-0002-3051-6038 & YÖK ID 45227) Öztop, Erhan
dc.contributor.ozuauthorÖztop, Erhan
dc.identifier.volume155en_US
dc.identifier.startpage1896en_US
dc.identifier.endpage1907en_US
dc.subject.keywordsDeep learningen_US
dc.subject.keywordsLearning from demonstrationen_US
dc.subject.keywordsReinforcement learningen_US
dc.subject.keywordsRepresentation learningen_US
dc.identifier.scopusSCOPUS:2-s2.0-85175852693
dc.relation.publicationcategoryConference Paper - International - Institutional Academic Staff


Files in this item

Thumbnail

This item appears in the following Collection(s)

Show simple item record


Share this page