z-logo
open-access-imgOpen Access
An autonomous agent for negotiation with multiple communication channels using parametrized deep Q-network
Author(s) -
Siqi Chen,
Ran Su
Publication year - 2022
Publication title -
mathematical biosciences and engineering
Language(s) - English
Resource type - Journals
SCImago Journal Rank - 0.451
H-Index - 45
eISSN - 1551-0018
pISSN - 1547-1063
DOI - 10.3934/mbe.2022371
Subject(s) - negotiation , computer science , bidding , reinforcement learning , artificial intelligence , action (physics) , markov decision process , nasa deep space network , space (punctuation) , human–computer interaction , machine learning , markov process , engineering , mathematics , statistics , physics , quantum mechanics , political science , law , spacecraft , aerospace engineering , operating system , marketing , business
Agent-based negotiation aims at automating the negotiation process on behalf of humans to save time and effort. While successful, the current research considers communication between negotiation agents through offer exchange. In addition to the simple manner, many real-world settings tend to involve linguistic channels with which negotiators can express intentions, ask questions, and discuss plans. The information bandwidth of traditional negotiation is therefore restricted and grounded in the action space. Against this background, a negotiation agent called MCAN (multiple channel automated negotiation) is described that models the negotiation with multiple communication channels problem as a Markov decision problem with a hybrid action space. The agent employs a novel deep reinforcement learning technique to generate an efficient strategy, which can interact with different opponents, i.e., other negotiation agents or human players. Specifically, the agent leverages parametrized deep Q-networks (P-DQNs) that provides solutions for a hybrid discrete-continuous action space, thereby learning a comprehensive negotiation strategy that integrates linguistic communication skills and bidding strategies. The extensive experimental results show that the MCAN agent outperforms other agents as well as human players in terms of averaged utility. A high human perception evaluation is also reported based on a user study. Moreover, a comparative experiment shows how the P-DQNs algorithm promotes the performance of the MCAN agent.

The content you want is available to Zendy users.

Already have an account? Click here to sign in.
Having issues? You can contact us here