Optimizing Automated Negotiation: Integrating Opponent Modeling with Reinforcement Learning for Strategy Enhancement
Abstract
:1. Introduction
- (1)
- Compared to traditional methods, which often rely on manual processes during the pre-negotiation phase, this study introduces an approach that enhances the autonomy of opponent selection and classification through social network analysis. This approach ensures full autonomy throughout the entire process, from opponent selection to strategy execution, eliminating the need for external intervention or human involvement. As a result, it significantly enhances the agent’s autonomy.
- (2)
- Unlike existing research on automated negotiation, which primarily focuses on the formal negotiation stage—especially on modeling the behavior of the negotiating opponent, and which often faces the “cold start” problem—our study introduces opponent modeling during the pre-negotiation phase. This early modeling serves as the foundation for adjusting agent preferences in subsequent learning, thereby significantly reducing the likelihood of the cold start problem.
- (3)
- We innovatively analyze the relationship between negotiating parties in the pre-negotiation phase and its impact on the negotiation strategies. Our methodology links the modeling of opponents during pre-negotiation to the strategy adjustments made in the formal negotiation phase. This association enables agents to exhibit a higher level of personalization and intelligence during the negotiation interaction process.
2. Literature Review
2.1. Opponent Modeling Before Negotiation
2.2. Strategy Optimization in Automated Negotiations
3. The Proposed Model
3.1. Model Framework
- (1)
- Evaluation and decision: Each negotiating agent first assesses the negotiation problem to identify its goals and constraints. This includes determining critical parameters such as negotiation time constraints and termination conditions.
- (2)
- Opponent selection: Agents use the Adamic–Adar algorithm to select negotiation opponents from the relationship network. This algorithm prioritizes potential opponents with stronger connections to the agent and aims to increase the likelihood of a successful negotiation outcome.
- (3)
- Opponent classification: Once the best-matched opponent is selected, the agent employs a logistic regression model to categorize it. This classification clarifies the nature of the relationship between the parties and provides a foundation for developing more targeted negotiation strategies.
- (4)
- Parameter adjustment: Based on the selection and classification outcomes, the agent modifies its preferences and learning parameters to refine its negotiation strategy. This adjustment aims to enhance the negotiation strategy’s adaptability and flexibility.
- (5)
- Proposal update: The agent updates its proposal using the Q-learning algorithm, allowing dynamic adjustments based on negotiation progress and opponent responses. This learning-based approach enables continuous optimization of the agent’s decisions through the interaction process.
- (1)
- The negotiation is limited to bilateral communication.
- (2)
- Agents exchange numerical proposals exclusively, focusing on quantitative attributes such as price or delivery terms.
- (3)
- Negotiations involve multi-attribute proposals, including attributes such as price, quality, and delivery terms.
- (4)
- For either agent, an agreement is reached for the benefit attribute if the opponent’s proposed value is greater than or equal to the agent’s proposed value. For cost-based attributes, an agreement is reached if the opponent’s proposed value is less than or equal to the agent’s proposed value.
- (5)
- T indicates the preset maximum number of rounds, and if no agreement is reached after T rounds, the negotiation is considered a failure.
3.2. Negotiation Opponent Modeling
3.2.1. Opponent Selection
- (1)
- Define the search space for negotiation opponents, which should encompass all potential opponent nodes.
- (2)
- For each opponent node, calculate its association strength with the negotiating agent by the Adamic–Adar algorithm.
- (3)
- Rank the calculated association values in descending order and select the node with the top-1 association value as the best-matched negotiation opponent.
Algorithm 1. Opponent selection based on the Adamic–Adar algorithm |
Input: Agent initiating the negotiation and a social network graph , where denotes nodes (agents) and denotes edges (social ties). Output: The best-matched opponent for negotiation, . Steps: 1. Define the search space 2. for each node : 3. Calculate the association level 4. Create a list containing tuples of each opponent and their association level: 5. Sort in descending order based on : 6. Select the best-matched opponent : 7. Return |
3.2.2. Opponent Classification
- (1)
- Node features
- (2)
- Structure features
- (3)
- Relationship classification prediction based on logistic regression
Algorithm 2. Opponent classification based on features and logistic regression |
Input: Agent initiating the negotiation; the opponent whose relationship with needs to be classified; A social network graph , where denotes nodes (agents), denotes edges (social ties), and denotes the relationship label. Output: The predicted relationship type , where indicates a positive relationship and indicates a negative relationship. Steps: 1. is divided into and 2. Extract node features : 3. Extract structural features : 4. Combine feature vectors : 5. Train the logistic regression model on a training dataset : Classifier (Logistics Regression) ← Train () 6. Compute the probability of a positive relationship: 7. Determine the predicted relationship type : 8. Return |
3.3. Proposal Updating Based on Opponent Modeling
3.3.1. Basic Proposal Updating Based on Q-Learning
3.3.2. Strategy Enhancement Based on Opponent Modeling
4. Experiments
4.1. Experimental Settings
4.2. Experiment Description
4.3. Baseline Models
4.4. Evaluation Metrics
5. Results
5.1. Model Performance
- (1)
- The “full feature baseline group” exhibited the best performance, achieving a relatively high recall, albeit with slightly lower precision. This suggests that the model prioritizes the identification of as many positive class samples as possible, resulting in a higher recall, albeit at the expense of introducing some incorrect positive class predictions, which accounts for the slightly lower precision. The F1 score, calculated at 0.9723, provides a balanced measure between these two metrics.
- (2)
- The removal of node features led to a slight decline in model performance, with accuracy and F1 score decreasing by 0.007 and 0.0037, respectively. This indicates that node features are crucial for identifying relationships between nodes, particularly in capturing potential interactions. Nevertheless, the recall remained high, suggesting that the model continues to identify the majority of positive class samples even in the absence of node features.
- (3)
- Removing structural features caused a more significant decline in performance, with accuracy dropping by 0.0266 and the F1 score by 0.0141. This highlights the importance of structural features in capturing structural relationships between nodes and identifying potential interaction patterns. The removal of structural features resulted in a decreased ability of the model to capture complex relationships between nodes, thereby reducing its classification effectiveness.
5.2. Ablation Study
- (1)
- The negotiation success rate achieved 100% under all experimental conditions. This indicates that whether using the original model (OM), the model with Component 1 added (OM+C1), the model with Component 2 added (OM+C2), or the final proposed model (PM), negotiations were consistently successful. This observation suggests that the original model demonstrates high stability and reliability in terms of negotiation success rate.
- (2)
- Regarding negotiation rounds, the original model (OM) had 12.350 negotiation rounds; with the addition of only Component 1, negotiation rounds significantly decreased to 10.575; with the addition of only Component 2, the number of negotiating rounds slightly decreased to 11.206; and the final proposed model (PM), which incorporates both Component 1 and Component 2, had a negotiation round of 11.758. This observation suggests that the combination of the two components may increase the complexity of the model, leading to a further increase in negotiation rounds. However, the PM still showed an improvement in negotiation rounds compared to the OM.
- (3)
- In terms of joint utility and utility difference, these two indicators showed a gradual improvement as the components were gradually added. As for joint utility, with the gradual addition of components, joint utility showed a steady, increasing trend. As for the utility difference, it can be observed that, after only adding Component 1, the utility difference showed a significant decrease, and after only adding Component 2, the utility difference showed a slight decrease. Overall, those indicate that the optimized final proposed model achieves higher joint utility in negotiations and reduces the utility difference among participants.
5.3. Comparative Analysis
- (1)
- (2)
- Negotiation rounds: although the model in this paper maintains a comparably low number of negotiation rounds with a high success rate, it slightly trails behind the model proposed by Qie et al. [57].
- (3)
- Joint utility: The model proposed in this paper achieves the highest joint utility score of 1.054 among the five models. This indicates its superiority in optimizing the interests of both negotiating parties.
- (4)
- Utility difference: regarding utility difference, the model presented in this paper outperforms others with a value of 0.122, indicating its ability to effectively mitigate imbalances in interest distribution between negotiating parties.
6. Conclusions
Author Contributions
Funding
Data Availability Statement
Conflicts of Interest
References
- Kenney, M.; Zysman, J. The platform economy: Restructuring the space of capitalist accumulation. Camb. J. Reg. Econ. Soc. 2020, 13, 55–76. [Google Scholar] [CrossRef]
- Jiang, Y.Q.; Wu, M.X.; Li, X.H. Cooperation mechanism, product quality, and consumer segment in group buying. Electron. Commer. Res. Appl. 2024, 64, 101357. [Google Scholar] [CrossRef]
- Yadav, R.; Bhatnagar, S. Channel collaboration in e-commerce: A study on channel relationship from the perspective of vendors selling on online platforms, the e-retailers. In Transforming Organizations Through Flexible Systems Management; Springer: Singapore, 2020; pp. 223–241. [Google Scholar]
- Liang, C.-C.; Liang, W.-Y.; Tseng, T.-L. Evaluation of intelligent agents in consumer-to-business e-Commerce. Comput. Stand. Interfaces 2019, 65, 122–131. [Google Scholar] [CrossRef]
- Gao, T.-G.; Huang, M.; Wang, Q.; Wang, X.-W. Dynamic organization model of automated negotiation for 3PL providers selection. Inf. Sci. 2020, 531, 139–158. [Google Scholar] [CrossRef]
- Moghadam, F.S.; Zarandi, M.F. Mitigating bullwhip effect in an agent-based supply chain through a fuzzy reverse ultimatum game negotiation module. Appl. Soft Comput. 2022, 116, 108278. [Google Scholar] [CrossRef]
- Bagga, P.; Paoletti, N.; Alrayes, B.; Stathis, K. ANEGMA: An automated negotiation model for e-markets. Auton. Agents Multi-Agent Syst. 2021, 35, 27. [Google Scholar] [CrossRef]
- Yang, C.H.; Sun, J.Y. Research on Negotiation of Manufacturing Enterprise Supply Chain Based on Multi-agent. J. Internet Technol. 2019, 20, 389–398. [Google Scholar]
- Tesfay, T.; Haouzi, H.B.E.; Demesure, G.; Pannequin, R.; Thomas, A. Multi-agent systems negotiation to deal with dynamic scheduling in disturbed industrial context. J. Intell. Manuf. 2019, 31, 1367–1382. [Google Scholar]
- Pan, Y.H.; Zhang, H.Y.; Zeng, Y.F.; Ma, B.Y.; Tang, J.; Ming, Z. Diversifying agent’s behaviors in interactive decision models. Int. J. Intell. Syst. 2022, 37, 12035–12056. [Google Scholar] [CrossRef]
- Keskin, M.O.; Buzcu, B.; Aydogan, R. Conflict-based negotiation strategy for human-agent negotiation. Appl. Intell. 2023, 53, 29741–29757. [Google Scholar] [CrossRef]
- Cao, M.K.; Hu, Q.; Kiang, M.Y.; Hong, H. A Portfolio Strategy Design for Human-Computer Negotiations in e-Retail. Int. J. Electron. Commer. 2020, 24, 305–337. [Google Scholar] [CrossRef]
- Imran, K.; Zhang, J.; Pal, A.; Khattak, A.; Ullah, K.; Baig, S.M. Bilateral negotiations for electricity market by adaptive agent-tracking strategy. Electr. Power Syst. Res. 2020, 186, 106390. [Google Scholar] [CrossRef]
- Hao, J.; Song, S.; Leung, H.-f.; Ming, Z. An efficient and robust negotiating strategy in bilateral negotiations over multiple items. Eng. Appl. Artif. Intell. 2014, 34, 45–57. [Google Scholar] [CrossRef]
- Eshragh, F.; Shahbazi, M.; Far, B. Real-time opponent learning in automated negotiation using recursive Bayesian filtering. Expert Syst. Appl. 2019, 128, 28–53. [Google Scholar] [CrossRef]
- Wu, J.; Sun, Y.; Li, Y.; Qie, X. A Q-learning approach to generating behavior of emotional persuasion with adaptive time belief in decision-making of agent-based negotiation. Inf. Sci. 2023, 642, 119158. [Google Scholar] [CrossRef]
- Wu, J.; Chen, H.; Li, Y.; Liu, Y. A Behavioral Assessment Model for Emotional Persuasion Driven by Agent-Based Decision-Making. Expert Syst. Appl. 2022, 204, 117556. [Google Scholar] [CrossRef]
- Yao, J.; Storme, M. Trust Building via Negotiation: Immediate versus Lingering Effects of General Trust and Negotiator Satisfaction. Group Decis. Negot. 2021, 30, 507–528. [Google Scholar] [CrossRef]
- Kimiagari, S.; Keivanpour, S.; Jolai, F.; Moazami, M. Application of fuzzy group analytic hierarchy process in partner selection of international joint venture projects. Sci. Iran. 2016, 23, 2959–2976. [Google Scholar] [CrossRef]
- Panzarasa, P.; Jennings, N.R. Social influence, negotiation and cognition. Simul. Model. Pract. Theory 2002, 10, 417–453. [Google Scholar] [CrossRef]
- Zhou, H.Z.; Zhan, J.Y.; Ma, W.J. A negotiation protocol with recommendation for multilateral negotiation in trust networks. Expert Syst. Appl. 2024, 246, 123185. [Google Scholar] [CrossRef]
- Baarslag, T.; Hendrikx, M.J.C.; Hindriks, K.V.; Jonker, C.M. Learning about the opponent in automated bilateral negotiation: A comprehensive survey of opponent modeling techniques. Auton. Agents Multi-Agent Syst. 2016, 30, 849–898. [Google Scholar] [CrossRef]
- Munroe, S.; Luck, M. Motivation-based selection of negotiation opponents. In Engineering Societies in the Agents World V; Lecture Notes in Artificial Intelligence; Gleizes, M.P., Omicini, A., Zambonelli, F., Eds.; Springer: Berlin/Heidelberg, Germany, 2005; Volume 3451, pp. 119–138. [Google Scholar]
- Hou, Y.Q.; Sun, M.Y.; Zhu, W.X.; Zeng, Y.F.; Piao, H.Y.; Chen, X.F.; Zhang, Q. Behavior Reasoning for Opponent Agents in Multi-Agent Learning Systems. IEEE Trans. Emerg. Top. Comput. Intell. 2022, 6, 1125–1136. [Google Scholar] [CrossRef]
- Liu, C.J.; Cong, J.M.; Zhao, T.H.; Zhu, E. Improving Agent Decision Payoffs via a New Framework of Opponent Modeling. Mathematics 2023, 11, 3062. [Google Scholar] [CrossRef]
- Talman, S.; Hadad, M.; Gal, Y.a.; Kraus, S. Adapting to agents’ personalities in negotiation. In Proceedings of the Fourth International Joint Conference on Autonomous Agents and Multiagent Systems, Auckland, New Zealand, 25–29 July 2005; pp. 383–389. [Google Scholar]
- Schröter, K.; Urbig, D. C-IPS: Specifying decision interdependencies in negotiations. In Multiagent System Technologies, Proceedings; Lecture Notes in Computer Science; Lindemann, G., Denzinger, J., Timm, I.J., Unland, R., Eds.; Springer: Berlin/Heidelberg, Germany, 2004; Volume 3187, pp. 114–125. [Google Scholar]
- Cheng, K.; Guo, X.; Cui, X.; Shan, F. Dynamical modeling, analysis, and control of information diffusion over social networks: A deep learning-based recommendation algorithm in social network. Discret. Dyn. Nat. Soc. 2020, 2020, 3273451. [Google Scholar] [CrossRef]
- Beauprez, E.; Caron, A.C.; Morge, M.; Routier, J.C. A Multi-Agent Negotiation Strategy for Reducing the Flowtime. In Proceedings of the 13th International Conference on Agents and Artificial Intelligence (ICAART), Virtual, 4–6 February 2021; pp. 58–68. [Google Scholar]
- Alkasasbeh, A.H.; Badr, E.; Attiya, H.; Shabana, H.M. Radio Number for Friendship Communication Networks. Mathematics 2023, 11, 4232. [Google Scholar] [CrossRef]
- Gong, Y.H.; Chen, L.; Ma, T.H. A Comprehensive Trust Model Based on Social Relationship and Transaction Attributes. Secur. Commun. Netw. 2020, 2020, 8887596. [Google Scholar] [CrossRef]
- Parhizkar, E.; Nikravan, M.H.; Holte, R.C.; Zilles, S. Combining Direct Trust and Indirect Trust in Multi-Agent Systems. In Proceedings of the 29th International Joint Conference on Artificial Intelligence, Yokohama, Japan, 7–15 January 2021; pp. 311–317. [Google Scholar]
- Player, C.; Griffiths, N. Improving trust and reputation assessment with dynamic behaviour. Knowl. Eng. Rev. 2020, 35, e29. [Google Scholar] [CrossRef]
- Hu, L.; Wang, Z. Partner selection of GSC for furniture industry based on method of entropy correction G1-TOPSIS. J. Cent. South Univ. For. Technol. 2018, 38, 129–135. [Google Scholar]
- Kadowaki, K.; Kobayashi, K.; Kitamura, Y. Influence of social relationships on multiagent persuasion. In Proceedings of the 7th International Joint Conference on Autonomous Agents and Multiagent Systems, Estoril, Portugal, 12–16 May 2008; Volume 3, pp. 1221–1224. [Google Scholar]
- Jennings, N.R.; Faratin, P.; Lomuscio, A.R.; Parsons, S.; Sierra, C.; Wooldridge, M. Automated negotiation: Prospects, methods and challenges. Int. J. Group Decis. Negot. 2001, 10, 199–215. [Google Scholar] [CrossRef]
- Kiruthika, U.; Somasundaram, T.S.; Raja, S.K.S. Lifecycle Model of a Negotiation Agent: A Survey of Automated Negotiation Techniques. Group Decis. Negot. 2020, 29, 1239–1262. [Google Scholar] [CrossRef]
- Pandian, J.A.; Thirunavukarasu, R.; Mariappan, L.T. Enhancing lane detection in autonomous vehicles with multi-armed bandit ensemble learning. Sci. Rep. 2025, 15, 3198. [Google Scholar] [CrossRef] [PubMed]
- Hu, J.; Zou, L.; Xu, R. Bilateral Multi-Issue Negotiation Model for a Kind of Complex Environment. Int. Arab J. Inf. Technol. 2018, 15, 396–404. [Google Scholar]
- Sim, K.M.; Guo, Y.Y.; Shi, B.Y. BLGAN: Bayesian Learning and Genetic Algorithm for Supporting Negotiation With Incomplete Information. IEEE Trans. Syst. Man Cybern. Part B-Cybern. 2009, 39, 198–211. [Google Scholar] [CrossRef]
- Mell, J.; Beissinger, M.; Gratch, J. An expert-model and machine learning hybrid approach to predicting human-agent negotiation outcomes in varied data. J. Multimodal User Interfaces 2021, 15, 215–227. [Google Scholar] [CrossRef]
- Chen, L.; Dong, H.; Zhou, Y. A reinforcement learning optimized negotiation method based on mediator agent. Expert Syst. Appl. 2014, 41, 7630–7640. [Google Scholar] [CrossRef]
- Wang, Y.; Liang, X. Application of Reinforcement Learning Methods Combining Graph Neural Networks and Self-Attention Mechanisms in Supply Chain Route Optimization. Sensors 2025, 25, 955. [Google Scholar] [CrossRef]
- Liu, Y.; Yang, T.; Tian, L.; Pei, J. SGD-TripleQNet: An Integrated Deep Reinforcement Learning Model for Vehicle Lane-Change Decision. Mathematics 2025, 13, 235. [Google Scholar] [CrossRef]
- Adamic, L.A.; Adar, E. Friends and neighbors on the web. Soc. Netw. 2003, 25, 211–230. [Google Scholar] [CrossRef]
- Samoylenko, I.; Aleja, D.; Primo, E.; Alfaro-Bittner, K.; Vasilyeva, E.; Kovalenko, K.; Musatov, D.; Raigorodskii, A.M.; Criado, R.; Romance, M.; et al. Why Are There Six Degrees of Separation in a Social Network? Phys. Rev. X 2023, 13, 021032. [Google Scholar] [CrossRef]
- Yang, C.; Liu, T.T.; Liu, L.; Chen, X.H.; Hao, Z.Y. A Personalized Friend Recommendation Method Combining Network Structure Features and Interaction Information. In Proceedings of the 9th International Conference on Swarm Intelligence (ICSI), Shanghai, China, 17–22 July 2018; pp. 267–274. [Google Scholar]
- Milgram, S. The small world problem. Psychol. Today 1967, 2, 60–67. [Google Scholar]
- Heider, F. The Psychology of Interpersonal Relations; Psychology Press: New York, NY, USA, 2013. [Google Scholar]
- Liu, M.; Guo, J.; Chen, J. Link Prediction in Signed Networks Based on Similarity and Structural Balance Theory. Adv. Eng. Sci. 2018, 50, 161–169. [Google Scholar]
- Zhang, B.; Liu, W.Q.; Zhang, Y.; Yang, R.; Li, M.Z. Implicit Negative Link Prediction With a Network Topology Perspective. IEEE Trans. Comput. Soc. Syst. 2023, 10, 3132–3142. [Google Scholar] [CrossRef]
- Tang, L.; Ma, X.; Zhou, Y.; Shi, X.; Ma, J. Social relations, public interventions and land rent deviation: Evidence from Jiangsu Province in China. Land Use Policy 2019, 86, 406–420. [Google Scholar] [CrossRef]
- Son, B.-G.; Kocabasoglu-Hillmer, C.; Roden, S. A dyadic perspective on retailer–supplier relationships through the lens of social capital. Int. J. Prod. Econ. 2016, 178, 120–131. [Google Scholar] [CrossRef]
- Ye, Z.; Zhong, Z. Benefit expectation, social trust and farmer’s choice of farmland transfer-out contract. J. Arid Land Resour. Environ. 2022, 36, 72–80. [Google Scholar] [CrossRef]
- Leskovec, J.; Huttenlocher, D.; Kleinberg, J. Signed networks in social media. In Proceedings of the SIGCHI Conference on Human Factors in Computing Systems, Atlanta, GA, USA, 10–15 April 2010; pp. 1361–1370. [Google Scholar]
- Wu, J.; Wang, W.; Li, Y. Emotional agents enabled bilateral negotiation: Persuasion strategies generated by agents’ affect infusion and preference. Expert Syst. Appl. 2024, 252, 124202. [Google Scholar] [CrossRef]
- Qie, X.T.; Wu, J.H.; Li, Y.; Sun, Y. A stage model for agent-based emotional persuasion with an adaptive target: From a social exchange perspective. Inf. Sci. 2022, 610, 90–113. [Google Scholar] [CrossRef]
Node Feature | Meaning in Network Analysis |
---|---|
Positive in-degree, indicating the quantity of positive relationships pointing to node . | |
Positive out-degree, representing the quantity of positive relationships originating from node . | |
Negative in-degree, indicating the quantity of negative relationships pointing to node . | |
Negative out-degree, representing the quantity of negative relationships originating from node . | |
Total in-degree, denoting the total quantity of relationships originating from node . | |
Total out-degree, denoting the total quantity of relationships pointing to node . |
Metrics | Meanings | Significance |
---|---|---|
SR | Defined as the ratio of agreements reached in n negotiation experiments. | The success rate primarily reflects the model’s ability to facilitate consensus under permitted conditions and parameter settings. |
NR | Reflects the time efficiency of negotiations, calculated as the average number of negotiation rounds across all experiments. | The average number of negotiation rounds indicates the average rounds required to reach a consensus, serving as a crucial metric for measuring negotiation efficiency. |
UD | Represents the degree of difference in utility obtained by the negotiating parties. | Utility difference indicates the disparity in utilities at the conclusion of negotiations, which can reflect the fairness of the negotiation outcome to some extent. |
JU | Represents the sum of the utility gained by both parties upon reaching an agreement. | Under the premise of small utility difference, the higher the common utility, the more beneficial the negotiation result is to achieve a win-win situation. |
Real Case | |||
---|---|---|---|
Positive | Negative | ||
Prediction class | Positive | True Positive () | False Positive () |
Negative | False Negative () | True Negative () |
Experiment Group | Accuracy | Precision | Recall | F1-Score |
---|---|---|---|---|
Full feature baseline group | 0.9487 | 0.9532 | 0.9923 | 0.9723 |
Node feature ablation group | 0.9417 | 0.9481 | 0.9899 | 0.9686 |
Structural feature ablation group | 0.9221 | 0.9333 | 0.9845 | 0.9582 |
Performance Indicators | Wu et al. [56] | Qie et al. [57] | Wu et al. [17] | Cao et al. [12] | Our Method | Friedman Test (p-Value) |
---|---|---|---|---|---|---|
SR | 97.2% | 100% | 100% | 53.2% | 100% | - |
NR | 12.133 | 3.710 | 12.350 | 71.247 | 11.758 | |
JU | 1.051 | 1.041 | 1.043 | 1.534 | 1.054 | |
UD | 0.128 | 0.420 | 0.131 | 0.431 | 0.122 |
Post Hoc Nemenyi Test (p-Value) | |||||
---|---|---|---|---|---|
NR | Wu et al. [56] | Qie et al. [57] | Wu et al. [17] | Cao et al. [12] | Our method |
Wu et al. [56] | 1.00 | ||||
Qie et al. [57] | 1.00 | ||||
Wu et al. [17] | 1.00 | ||||
Cao et al. [12] | 1.00 | ||||
Our method | 1.00 | ||||
JU | Wu et al. [56] | Qie et al. [57] | Wu et al. [17] | Cao et al. [12] | Our method |
Wu et al. [56] | 1.00 | ||||
Qie et al. [57] | 1.00 | ||||
Wu et al. [17] | 1.00 | ||||
Cao et al. [12] | 1.00 | ||||
Our method | 1.00 | ||||
UD | Wu et al. [56] | Qie et al. [57] | Wu et al. [17] | Cao et al. [12] | Our method |
Wu et al. [56] | 1.00 | ||||
Qie et al. [57] | 1.00 | ||||
Wu et al. [17] | 1.00 | ||||
Cao et al. [12] | 1.00 | ||||
Our method | 1.00 |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2025 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Zhang, Y.; Wu, J.; Cao, R. Optimizing Automated Negotiation: Integrating Opponent Modeling with Reinforcement Learning for Strategy Enhancement. Mathematics 2025, 13, 679. https://doi.org/10.3390/math13040679
Zhang Y, Wu J, Cao R. Optimizing Automated Negotiation: Integrating Opponent Modeling with Reinforcement Learning for Strategy Enhancement. Mathematics. 2025; 13(4):679. https://doi.org/10.3390/math13040679
Chicago/Turabian StyleZhang, Ya, Jinghua Wu, and Ruiyang Cao. 2025. "Optimizing Automated Negotiation: Integrating Opponent Modeling with Reinforcement Learning for Strategy Enhancement" Mathematics 13, no. 4: 679. https://doi.org/10.3390/math13040679
APA StyleZhang, Y., Wu, J., & Cao, R. (2025). Optimizing Automated Negotiation: Integrating Opponent Modeling with Reinforcement Learning for Strategy Enhancement. Mathematics, 13(4), 679. https://doi.org/10.3390/math13040679