Version 24 (modified by mark, 13 years ago) ( diff )

--

Literature Survey: Opponent Models

This page provides an overview of the literature survey of Mark Hendrikx. The first Section discusses the researched papers, including a summary
of the relevant details. The second Section provides an overview of all meetings. Finally, the third Section outlines a global planning.

Papers

TitleAgentFSEGA - Time Constrained Reasoning Model for Bilateral Multi-Issue Negotiation
Author(s)L.D. Serban, G.C. Silaghi, and C.M. Litan
Subject(s)
Summary
Relevance
Bibtex


TitleAn Architecture for Negotiating Agents that Learn
Author(s)H.H. Bui, S. Venkatesh, and D. Kieronska
Subject(s)
Summary
Relevance
Bibtex


TitleAnticipating Agent's Negotiation Strategies in an E-marketplace Using Belief Models
Author(s)F. Teuteberg, K. Kurbel
Subject(s)
Summary
Relevance
BibtexLink


TitleAnalysis of Negotiation Dynamics
Author(s)K. Hindriks, C.M. Jonker, D. Tykhonov
Subject(s)
Summary
Relevance
Bibtex


TitleBayesian Learning in Bilateral Multi-issue Negotiation and its Application in MAS-based Electronic Commerce
Author(s)J. Li, Y. Cao
Subject(s)
Summary
Relevance
Bibtex


TitleBenefits of Learning in Negotiation
Author(s)D. Zeng, K. Sycara
Subject(s)Benefits of learning, Bayesian learning, reservation values
Summary Growing interest in e-commerce motivates research in automated negotiation. Building intelligent negotiation agents is still
emerging. In contrast to most negotiation models, sequential decision model allows for learning. Learning can help understand
human behaviour, but can also result in better results for the learning party. Bayesian learning of reservation
values can be used to determine the zone of agreement for an issue based on the domain knowledge and bidding interactions.
Concluding for one-issue, learning positively influences bargaining quality, number of exchanged proposals,
and leads to a better compromise if both learn. Learning works always works better in the proposed case.
Relevance8. Strong example of Bayesian learning
BibtexLink


TitleBayesian Learning in Negotiation
Author(s)D. Zeng, K. Sycara
Subject(s)
Summary
Relevance
Bibtex


TitleBilateral Negotiation with Incomplete and Uncertain Information
Author(s)C. Mudgal, J. Vassileva
Subject(s)
Summary
Relevance
Bibtex


TitleCompromising Strategy based on Estimated Maximum Utility for Automated Negotiating Agents
Author(s)S. Kawaguchi, K. Fujita, T. Ito
Subject(s)
Summary
Relevance
Bibtex


TitleUsing Similarity Criteria to Make Issue Trade-offs in Automated Negotiations
Author(s)P. Faratin, C. Sierra, N.R. Jennings
Subject(s)
Summary
Relevance
Bibtex


TitleFacing the Challenge of Human-Agent Negotiations via Effective General Opponent Modeling
Author(s)J. Li, Y. Cao
Subject(s)
Summary
Relevance
Bibtex


TitleIAMhaggler: A Negotiation Agent for Complex Environments
Author(s)C.R. Williams, V. Robu, E.H. Gerding, and N.R. Jennings
Subject(s)
Summary
Relevance
Bibtex


TitleLearning an Opponent's Preferences to Make Effective Multi-Issue Negotiation Trade-Offs
Author(s)R.M. Coehoorn, N.R. Jennings
Subject(s)KDE Learning, Negotiation model, Concession based strategy
Summary Effective and efficient multi-issue negotiation requires an agent to have some indication of it's opponent's preferences
over the issues in the domain. Kernel Density Estimation (KDE) is used to estimate the weight attached to different issues
by different agents. It is assumed that if the value of an issue increases, that this is positive for one agent, and negative
for the other. No assumptions about relation between time, negotiation history and issue-weight are required, in contrast
to Bayesian learning. The difference between concessive (counter)offers is used to estimate the weights of the issues
(assumption: stronger concessions are made later on in the negotiation). Faratin's hill climbing algorithm augmented with KDE is
used to propose the next bid. KDE proved succesful on the used negotiation model. Future works entails testing the approach
against different opponent strategies and extending the approach to other negotiation models (see assumption in summary).
Relevance9. KDE learning described in detail. Strong related work section
BibtexLink


TitleLearning other agents' preferences in multiagent negotiation
Author(s)H.H. Bui, D. Kieronska, S. Venkatesh
Subject(s)
Summary
Relevance
Bibtex


TitleNegotiation Decision Functions for Autonomous Agent
Author(s)P. Faratin, C. Sierra, N.R. Jennings
Subject(s)
Summary
Relevance
Bibtex


TitleOn-Line Incremental Learning in Bilateral Multi-Issue Negotiation
Author(s)V. Soo, C. Hung
Subject(s)
Summary
Relevance
Bibtex


TitleOpponent Model Estimation in Bilateral Multi-issue Negotiation
Author(s)N. van Galen Last
Subject(s)
Summary
Relevance
Bibtex


TitleOpponent Modelling in Automated Multi-Issue Negotiation Using Bayesian Learning
Author(s)K. Hindriks, D. Tykhonov
Subject(s)
Summary
Relevance
BibtexLink


TitleThe First Automated Negotiating Agents Competition (ANAC 2010)
Author(s)T. Baarslag, K. Hindriks, C. Jonker, S. Kraus, R. Lin
Subject(s)ANAC, overview multiple agents, opponent models, acceptance conditions
SummaryThe ANAC competition models bilateral multi-issue closed negotiations and provides a benchmark for negotiation agents.
Opponent models can also be used to identify the type of strategy of the opponent. Interesting agents for further analysis
are: IAM(crazy)Haggler, FSEGA (profile learning), and Agent Smith. Issues can be predicatable, which means that they
have a logical order, or unpredicatable, such as colors. This paper also includes acceptance conditions.
Relevance5, too global, however interesting citations
BibtexLink


TitleTowards a Quality Assessment Method for Learning Preference Profiles in Negotiation
Author(s)K.V. Hindriks and D. Tykhonov
Subject(s)
Summary
Relevance
Bibtex


TitleYushu: a Heuristic-Based Agent for Automated Negotiating Competition
Author(s)B. An and V. Lesser
Subject(s)
Summary
Relevance
Bibtex


Meetings

Planning

Attachments (3)

Download all attachments as: .zip

Note: See TracWiki for help on using the wiki.