Publication:
Actor-critic reinforcement learning for bidding in bilateral negotiation

dc.contributor.authorArslan, Furkan
dc.contributor.authorAydoğan, Reyhan
dc.contributor.departmentComputer Science
dc.contributor.ozuauthorAYDOĞAN, Reyhan
dc.contributor.ozugradstudentArslan, Furkan
dc.date.accessioned2023-08-11T06:21:07Z
dc.date.available2023-08-11T06:21:07Z
dc.date.issued2022
dc.description.abstractDesigning an effective and intelligent bidding strategy is one of the most compelling research challenges in automated negotiation, where software agents negotiate with each other to find a mutual agreement when there is a conflict of interests. Instead of designing a hand-crafted decision-making module, this work proposes a novel bidding strategy adopting an actor-critic reinforcement learning approach, which learns what to offer in a bilateral negotiation. An entropy reinforcement learning framework called Soft Actor-Critic (SAC) is applied to the bidding problem, and a self-play approach is employed to train the model. Our model learns to produce the target utility of the coming offer based on previous offer exchanges and remaining time. Furthermore, an imitation learning approach called behavior cloning is adopted to speed up the learning process. Also, a novel reward function is introduced that does take not only the agent’s own utility but also the opponent’s utility at the end of the negotiation. The developed agent is empirically evaluated. Thus, a large number of negotiation sessions are run against a variety of opponents selected in different domains varying in size and opposition. The agent’s performance is compared with its opponents and the performance of the baseline agents negotiating with the same opponents. The empirical results show that our agent successfully negotiates against challenging opponents in different negotiation scenarios without requiring any former information about the opponent or domain in advance. Furthermore, it achieves better results than the baseline agents regarding the received utility at the end of the successful negotiations.en_US
dc.description.sponsorshipScientific and Research Council of Turkey ; TÜBİTAK
dc.description.versionPublisher versionen_US
dc.identifier.doi10.55730/1300-0632.3899en_US
dc.identifier.endpage1714en_US
dc.identifier.issn1300-0632en_US
dc.identifier.issue5en_US
dc.identifier.scopus2-s2.0-85139306597
dc.identifier.startpage1695en_US
dc.identifier.urihttp://hdl.handle.net/10679/8624
dc.identifier.urihttps://doi.org/10.55730/1300-0632.3899
dc.identifier.volume30en_US
dc.identifier.wos000904725600002
dc.language.isoengen_US
dc.peerreviewedyesen_US
dc.publicationstatusPublisheden_US
dc.publisherTÜBİTAKen_US
dc.relationinfo:eu-repo/grantAgreement/TUBITAK/1001 - Araştırma/118E197
dc.relation.ispartofTurkish Journal of Electrical Engineering and Computer Sciences
dc.relation.publicationcategoryInternational Refereed Journal
dc.rightsAttribution 4.0 International
dc.rightsopenAccess
dc.rights.urihttps://creativecommons.org/licenses/by/4.0/
dc.subject.keywordsAutomated bilateral negotiationen_US
dc.subject.keywordsBidding strategyen_US
dc.subject.keywordsDeep reinforcement learningen_US
dc.subject.keywordsEntropy reinforcement learningen_US
dc.subject.keywordsImitation learningen_US
dc.subject.keywordsMulti-agent systemsen_US
dc.titleActor-critic reinforcement learning for bidding in bilateral negotiationen_US
dc.typearticleen_US
dspace.entity.typePublication
relation.isOrgUnitOfPublication85662e71-2a61-492a-b407-df4d38ab90d7
relation.isOrgUnitOfPublication.latestForDiscovery85662e71-2a61-492a-b407-df4d38ab90d7

Files

Original bundle

Now showing 1 - 1 of 1
Loading...
Thumbnail Image
Name:
Actor-critic reinforcement learning for bidding in bilateral negotiation.pdf
Size:
918.13 KB
Format:
Adobe Portable Document Format
Description:

License bundle

Now showing 1 - 1 of 1
Placeholder
Name:
license.txt
Size:
1.45 KB
Format:
Item-specific license agreed upon to submission
Description:

Collections