Show simple item record

dc.contributor.authorArslan, Furkan
dc.contributor.authorAydoğan, Reyhan
dc.date.accessioned2023-08-11T06:21:07Z
dc.date.available2023-08-11T06:21:07Z
dc.date.issued2022
dc.identifier.issn1300-0632en_US
dc.identifier.urihttp://hdl.handle.net/10679/8624
dc.identifier.urihttps://journals.tubitak.gov.tr/elektrik/vol30/iss5/2/
dc.description.abstractDesigning an effective and intelligent bidding strategy is one of the most compelling research challenges in automated negotiation, where software agents negotiate with each other to find a mutual agreement when there is a conflict of interests. Instead of designing a hand-crafted decision-making module, this work proposes a novel bidding strategy adopting an actor-critic reinforcement learning approach, which learns what to offer in a bilateral negotiation. An entropy reinforcement learning framework called Soft Actor-Critic (SAC) is applied to the bidding problem, and a self-play approach is employed to train the model. Our model learns to produce the target utility of the coming offer based on previous offer exchanges and remaining time. Furthermore, an imitation learning approach called behavior cloning is adopted to speed up the learning process. Also, a novel reward function is introduced that does take not only the agent’s own utility but also the opponent’s utility at the end of the negotiation. The developed agent is empirically evaluated. Thus, a large number of negotiation sessions are run against a variety of opponents selected in different domains varying in size and opposition. The agent’s performance is compared with its opponents and the performance of the baseline agents negotiating with the same opponents. The empirical results show that our agent successfully negotiates against challenging opponents in different negotiation scenarios without requiring any former information about the opponent or domain in advance. Furthermore, it achieves better results than the baseline agents regarding the received utility at the end of the successful negotiations.en_US
dc.description.sponsorshipScientific and Research Council of Turkey ; TÜBİTAK
dc.language.isoengen_US
dc.publisherTÜBİTAKen_US
dc.relationinfo:turkey/grantAgreement/TUBITAK/118E197
dc.relation.ispartofTurkish Journal of Electrical Engineering and Computer Sciences
dc.rightsAttribution 4.0 International
dc.rightsopenAccess
dc.rights.urihttps://creativecommons.org/licenses/by/4.0/
dc.titleActor-critic reinforcement learning for bidding in bilateral negotiationen_US
dc.typeArticleen_US
dc.description.versionPublisher versionen_US
dc.peerreviewedyesen_US
dc.publicationstatusPublisheden_US
dc.contributor.departmentÖzyeğin University
dc.contributor.authorID(ORCID 0000-0002-5260-9999 & YÖK ID 145578) Aydoğan, Reyhan
dc.contributor.ozuauthorAydoğan, Reyhan
dc.identifier.volume30en_US
dc.identifier.issue5en_US
dc.identifier.startpage1695en_US
dc.identifier.endpage1714en_US
dc.identifier.wosWOS:000904725600002
dc.identifier.doi10.55730/1300-0632.3899en_US
dc.subject.keywordsAutomated bilateral negotiationen_US
dc.subject.keywordsBidding strategyen_US
dc.subject.keywordsDeep reinforcement learningen_US
dc.subject.keywordsEntropy reinforcement learningen_US
dc.subject.keywordsImitation learningen_US
dc.subject.keywordsMulti-agent systemsen_US
dc.identifier.scopusSCOPUS:2-s2.0-85139306597
dc.contributor.ozugradstudentArslan, Furkan
dc.relation.publicationcategoryArticle - International Refereed Journal - Institutional Academic Staff and Graduate Student


Files in this item

Thumbnail

This item appears in the following Collection(s)

Show simple item record

Attribution 4.0 International
Except where otherwise noted, this item's license is described as Attribution 4.0 International

Share this page