Login | Register

Designing and trusting multi-agent systems for B2B applications


Designing and trusting multi-agent systems for B2B applications

Alam, Rafiul (2008) Designing and trusting multi-agent systems for B2B applications. Masters thesis, Concordia University.

[thumbnail of MR45510.pdf]
Text (application/pdf)
MR45510.pdf - Accepted Version


This thesis includes two main contributions. The first one is designing and implementing B usiness-to-B usiness (B2B ) applications using multi-agent systems and computational argumentation theory. The second one is trust management in such multi-agent systems using agents' credibility. Our first contribution presents a framework for modeling and deploying B2B applications, with autonomous agents exposing the individual components that implement these applications. This framework consists of three levels identified by strategic, application, and resource, with focus here on the first two levels. The strategic level is about the common vision that independent businesses define as part of their decision of partnership. The application level is about the business processes, which are virtually integrated as result of this common vision. Since conflicts are bound to arise among the independent applications/agents, the framework uses a formal model based upon computational argumentation theory through a persuasion protocol to detect and resolve these conflicts. Termination, soundness, and completeness properties of this protocol are presented. Distributed and centralized coordination strategies are also supported in this framework, which is illustrated with an online purchasing case study followed by its implementation in Jadex, a java-based platform for multi-agent systems. An important issue in such open multi-agent systems is how much agents trust each other. Considering the size of these systems, agents that are service providers or customers in a B2B setting cannot avoid interacting with others that are unknown or partially known regarding to some past experience. Due to the fact that agents are self-interested, they may jeopardize the mutual trust by not performing the actions as they are supposed to. To this end, our second contribution is proposing a trust model allowing agents to evaluate the credibility of other peers in the environment. Our multi-factor model applies a number of measurements in trust evaluation of other party's likely behavior. After a period of time, the actual performance of the testimony agent is compared against the information provided by interfering agents. This comparison process leads to both adjusting the credibility of the contributing agents in trust evaluation and improving the system trust evaluation by minimizing the estimation error.

Divisions:Concordia University > Gina Cody School of Engineering and Computer Science > Concordia Institute for Information Systems Engineering
Item Type:Thesis (Masters)
Authors:Alam, Rafiul
Pagination:xi, 97 leaves : ill. ; 29 cm.
Institution:Concordia University
Degree Name:M.A. Sc.
Program:Institute for Information Systems Engineering
Thesis Supervisor(s):Bentahar, J
Identification Number:LE 3 C66Q35M 2008 A43
ID Code:975232
Deposited By: Concordia University Library
Deposited On:22 Jan 2013 15:45
Last Modified:13 Jul 2020 20:07
Related URLs:
All items in Spectrum are protected by copyright, with all rights reserved. The use of items is governed by Spectrum's terms of access.

Repository Staff Only: item control page

Downloads per month over past year

Research related to the current document (at the CORE website)
- Research related to the current document (at the CORE website)
Back to top Back to top