This work formalizes an informant-based structured argumentation approach in a multi-agent setting, where the knowledge base of an agent may include information provided by other agents, and each piece of knowledge comes attached with its informant. In that way, arguments are associated with the set of informants corresponding to the information they are built upon. Our approach proposes an informant-based notion of argument strength, where the strength of an argument is determined by the credibility of its informant agents. Moreover, we consider that the strength of an argument is not absolute, but it is relative to the resolution of the conflicts the argument is involved in. In other words, the strength of an argument may vary from one context to another, as it will be determined by comparison to its attacking arguments (respectively, the arguments it attacks). Finally, we equip agents with the means to express reasons for or against the consideration of any piece of information provided by a given informant agent. Consequently, we allow agents to argue about the arguments’ strength through the construction of arguments that challenge (respectively, defeat) or are in favour of their informant agents.
Argumentation has proved to be a powerful paradigm for conceptualizing commonsense reasoning and a promising research area within the field of Artificial Intelligence [7,11,14,40]. One of the main issues an argumentation system has to address is the selection of the acceptable arguments, whose conclusions can be considered to be justified. For this purpose, it is necessary to account for the conflicts between the arguments in the system, and how those conflicts are to be resolved. At this point, a crucial notion comes into play: the notion of argument strength. In general, given an attack from an argument to an argument , the attack will only succeed if is stronger than . However, there is no unique way of establishing the strength of the arguments, as evidenced by the wide variety of approaches existing in the literature of argumentation. For instance, in [3,5] the arguments’ strength is established by a general preference relation. On the other hand, [13] defines the strength of an argument through a formula yielding a numerical value, based on the intuition that the strength of an argument depends on the strength of its attackers. Finally, for instance, approaches to value-based argumentation [10] consider that the strength of an argument depends on the social values it advances, and determining whether the attack of one argument on another succeeds depends on the comparative strength of the values advanced by the arguments concerned. In contrast to these, in this work we will adopt a symbolic informant-based approach to argument strength in structured argumentation, where the strength of an argument is established in terms of the strength or credibility of its informant agents. Moreover, the strength of arguments in our approach will not be absolute and, furthermore, could be challenged.
In this paper we assume a multi-agent setting where agents share domain knowledge with one another. Hence, each deliberative agent may obtain pieces of information from other informant agents, which can have different levels of credibility. In order to reach conclusions to establish their beliefs, agents will build arguments based on the information in their knowledge bases, which include the pieces of information received from a set of their informant agents. In this context, we introduce an informant-based argumentative approach where the credibility of the information sources (i.e., the informant agents) will be used for determining the arguments’ strength. On the one hand, as it is usual in argumentation systems, arguments could be challenged due to the information they use. On the other hand, since the pieces of information used for building the arguments are associated with their informant agents, our approach is such that arguments can also be challenged on their credibility, in other words, on their strength.
The knowledge representation and reasoning capabilities of our approach will take their basis from DeLP [25]. Specifically, our proposal greatly extends and further develops the approach introduced in [21] by providing a formal account of the arguments’ strength based on the credibility of their informant sources. In particular, we extend DeLP’s representational language by including backing and detracting rules, respectively enabling to express reasons for and against the consideration of the information provided by some informant agents. Then, these new types of rules will be used for building backing and detracting arguments, allowing to argue about the arguments’ strength. As a result, in a situation where some informants are challenged whereas others are not, the existence of multiple informants for a given piece of information in an agent’s knowledge base may strengthen its position (hence, of the arguments using them).
For example, let us consider an agent that has to decide whether to buy a used smartphone or not; this agent receives information from agents , , and . Let us also assume that agent regards as less credible than , and as more credible than . Now, suppose informs that “the phone is a good option”, whereas receives information from expressing that “the phone is not a good option”. With this information, agent will be able to build arguments for buying the phone and for not buying it, and these two arguments will attack each other. Then, taking the credibility of the informants into account, the former argument will prevail over the latter and agent will arrive to the conclusion that “the phone is a good option” based on information provided by the more credible agent. However, it could be the case that there exist reasons for or against the consideration of information received from some informant. For instance, suppose agent provides information saying that is an expert on electronic devices. In addition, suppose informs that should not be trusted when it comes to recommending an item it owns. This new information could be encoded through a backing and a detracting rule, respectively, leading to the construction of their homonym arguments. As a result, the backing and detracting arguments would be in conflict and, taking the credibility of their informants into account, the detracting argument would prevail over the backing argument (because is more credible than ). In particular, the approach proposed in this paper will be able to handle situations like these, where the strength of the argument expressing that the phone was a good option is being challenged by the detracting argument.
To summarize, the contribution of this paper is two-fold. On the one hand, we formalize an informant-based structured argumentation approach where the strength of an argument is determined by the strength (i.e., the credibility) of its informant agents. Furthermore, as will become clear later, the strength of an argument in our approach is not absolute, but it is relative to the resolution of the conflicts the argument is involved in. In other words, the strength of an argument will vary from one context to another, as it will be determined by comparison to its attacking arguments (respectively, the arguments it attacks). As a result, it could be the case that some informant providing information for building an argument is relevant for establishing the argument’s strength in a given context, but not in others. On the other hand, through the incorporation of backing and detracting rules (and their homonym arguments) we provide the means for reasoning about the arguments’ strength. In particular, the former will allow us to express reasons for the consideration of any piece of information provided by a given informant agent; analogously, the latter enable to express reasons against the consideration of an informant and thus, will be challenging the strength of arguments making use of information provided by that informant.
The rest of this paper is organized as follows. In Section 2 we will introduce the elements that will be used to represent the agents’ knowledge. Section 3 shows how an agent can build different kinds of arguments and identify different kinds of attacks between them. Then, in Section 4, we will introduce the ways in which the conflicts between arguments are resolved, leading to defeats. Moreover, taking those defeats into account, in Section 5 we formalize the agents’ reasoning mechanism enabling to determine their warranted beliefs. Finally, Sections 6 and 7 discuss relevant related work, draw some conclusions and comment on future lines of work.
Knowledge representation
In this section we will introduce our proposal for agents’ knowledge representation that will be defined as an informant-based DeLP program. We assume that agents may share tentative information in the form of defeasible rules, and that each agent can obtain information from other informant agents that have different degrees of credibility. Also, backing and detracting rules will be proposed in our formalization, allowing to express reasons for and against the consideration of the informant agents providing knowledge used for building the arguments.
The assumption of the existence of a total credibility order over informants is not quite realistic in many multi-agent application domains, and a similar observation applies to the existence of a global order shared by all agents. With this observation in mind, the approach to be introduced below will consider that every agent has its own partial order defined over the set of informant agents, representing the credibility it assigns to each informant. Each agent has a knowledge base where every piece of information is attached with an agent identifier representing that the corresponding agent is the source of that piece of information. In addition, agents could communicate with their peers for obtaining new information or for sharing their beliefs. Clearly, as agents may disagree with one another, the beliefs an agent has may be in conflict with another agent’s knowledge. Then, when sharing conflicting information, the credibility order among the informant agents can be used to decide which information prevails. Next, we briefly introduce the notion of credibility order, which will be used throughout the rest of the paper.
We assume a finite set of identifiers for naming informant agents, shared by all agents. Agent identifiers will be denoted with an uppercase typewriter letter “” that can have letters and natural numbers as subscripts (i.e., ), and each identifier is unequivocally associated with a single agent.
Each agent will have its own credibility order, represented by an irreflexive, asymmetric and transitive binary relation over , denoted (i.e., is a strict partial order over ), where stands for the agent identifier this order belongs to; for instance, is the credibility order of the agent identified as . In this paper we will assume that the credibility order relates agents that are sources of information about the same topic; multi-topic or multi-context credibility orders will be considered in future work. The notation means that agent deems agent as less credible than agent (equivalently, under ’s perspective, is more credible than ). Then, the notation is used to express that, for agent , is not less credible than . Furthermore, if and , then agent regards and to be incomparable, noted as . The following example illustrates a credibility order over agents, which will be used as part of the running example throughout the rest of the paper.
Consider the set of informants . Suppose agent has the following credibility order . Figure 1 shows a graphical representation of the credibility order . Following this notation, an edge from a node to a node represents that . For instance, the edge from to in Fig. 1 denotes the credibility relation .
Graphical representation of the credibility order from Example 1.
It should be noted that agents may change their assessment of one another; as a result, this change would impact their credibility orders, resulting in an update. The dynamic nature of credibility orders is outside the scope of this paper. However, for instance, the formalism proposed in [42], which provides a mechanism for handling the dynamics of a credibility order, can be a complement to our proposal.
As stated before, agents can receive information from different sources that are not equally credible. Thus, the information they provide, which may be contradictory, will be considered to be tentative. In this setting we need a mechanism for dealing with uncertain and conflicting information, able to ultimately determine the beliefs an agent should commit to. In particular, when deciding between contradictory conclusions, the reasoning mechanism will rely on the credibility order the agent has. In our approach, knowledge representation and reasoning is inspired on Defeasible Logic Programming (DeLP) [25]. Our interest in this particular computational tool is that its language provides the declarative capability of representing weak information in the form of defeasible rules and presumptions, and its defeasible argumentation inference mechanism allows warranting conclusions in the presence of contradictory information. Below we will introduce the elements that will be used for representing an agent’s knowledge, some of which are taken from DeLP’s representational language.
Let be a set of ground atoms. A ground literal L (or literal, for short) is an atom A or a negated atom , where and the symbol “∼” represents strong negation.
(Defeasible rule).
A defeasible rule is an ordered pair, denoted “”, where the first element is a ground literal and the second element is a finite set of ground literals. A defeasible rule with head and body () will also be written as .
A defeasible rule “” expresses that “reasons to believe in the antecedent Body give reasons to believe in the consequent Head”. For instance, the defeasible rule “” represents that “reasons to believe that it is snowing provide reasons to believe that the roads are closed”, whereas “” represents that “reasons to believe that snowplows are working provide reasons to believe that the roads are not closed”. In particular, a defeasible rule with an empty body is noted as “” and is called a presumption [37]. Hence, for instance, the presumption “” expresses that “there are defeasible reasons to believe that snowplows are working”. We associate defeasible rules (including presumptions) with their informant agents; for this, we introduce the notion of defeasible domain object.
(Defeasible domain object).
Let be a finite set of agent identifiers. A defeasible domain object is a tuple , where and R is a defeasible rule.
In addition to defeasible rules, we introduce backing rules and detracting rules to express reasons for and against the consideration of informants, respectively. Formally:
(Backing rule).
Let be a finite set of agent identifiers. A backing rule is an ordered pair, denoted “”, where is the identifier of an informant agent and is a finite and non-empty set of literals. A backing rule with head and body () will also be written as .
(Detracting rule).
Let be a finite set of agent identifiers. A detracting rule is an ordered pair, denoted “”, where is the identifier of an informant agent and is a finite and non-empty set of literals. A detracting rule with head and body () will also be written as .
Syntactically, the only difference between backing and detracting rules is the use of ⊕ and ⊘; however, these two types of rules are semantically opposite. A backing rule “” expresses that “the antecedentgives reasons for considering information provided by the informant”. On the contrary, a detracting rule “” expresses that “the antecedentgives reasons against considering information provided by the informant”. Also, note that backing and detracting rules do not have an associated agent (i.e., their source is not identified), meaning that they correspond to the agent’s own knowledge. Nevertheless, since these rules have a set of literals in their body, it is also possible to argue about the reasons for or against the consideration of the informants in their heads. When convenient, we will refer to backing and detracting rules as informant rules.
The entire knowledge of an agent, which will be used to make inferences and construct arguments, is composed of a set of defeasible domain objects and a set of informant rules. This knowledge base will be called informant-based DeLP program.
(Informant-based DeLP program).
An informant-based DeLP program (IBDP for short) is a pair , where Δ is a finite set of defeasible domain objects and Σ is a finite set of informant rules.
The following example introduces an IBDP that will serve as a running example, to illustrate the different notions proposed in this paper.
Let us consider the IBDP of an agent . Note that encodes agent ’s own information as well as information obtained from the following informant agents: , , , , , , .
The program has a set of defeasible domain objects in that include defeasible rules and presumptions obtained from different informants. It is worth mentioning that two different defeasible domain objects can have the same informant: for instance, and . Also, two different defeasible domain objects can have the same defeasible rule with different informant agents: for instance, and . On the other hand, the set includes backing and detracting rules. For instance, the detracting rule expresses that “g and q give reasons against the consideration of any piece of information provided by ”, and the backing rule expresses that “p gives reasons for considering the information provided by ”.
As evidenced in Example 2, an IBDP can have two or more defeasible domain objects with the same defeasible rule R but different informant agents. This does not mean that the agent’s knowledge base is redundant; rather, it encodes the fact that the same piece of information was received from different sources. This feature can be considered as an advantage of our approach since, as mentioned before, the credibility order of an agent may change (even dynamically); hence, at any moment, the more credible informant of R could be considered. Furthermore, given the existence of detracting rules, a specific informant of R could be challenged, whereas others are not. Again, a situation like this is illustrated by Example 2, where there exists a detracting rule for the informant of “”, but not for the informant . As a result, the existence of multiple informants may strengthen the position of a defeasible rule by allowing for different defeasible domain objects within the agent’s knowledge base (therefore, of the arguments using them). Finally, it is important to remark that the existence of backing and detracting rules in an IBDP is not mandatory.
In our approach, an agent will be specified in terms of four components: its own agent identifier, an informant-based DeLP program used to store its knowledge, a credibility order among informants, and an informant-based comparison criterion. The first three elements have been introduced above. On the other hand, the informant-based comparison criterion will make use of the credibility order to establish strict preferences among sets of informants. Briefly, sets of informant agents will be compared as part of the warranting process with the aim of deciding between conflicting arguments, to ultimately determine the accepted arguments and the justified conclusions of the agent. Consequently, since the notion of argument and the way in which they become in conflict (attack) have not been formalized yet, we will postpone the characterization of informant-based comparison criteria until Section 4; that is, we will formally introduce that notion after providing the formal context in which such criteria will be applied. An agent is defined as follows:
(Agent).
Let be a finite set of agent identifiers. An agent is a tuple where , is an informant-based DeLP program, is a credibility order over , and is an informant-based comparison criterion over sets of informants.
Consider the IBDP from Example 2 and the credibility order presented in Example 1. Then, agent can be specified as . The informant-based comparison criterion will be introduced in Section 4.
Given the characterization of an agent, it can be the case that two different agents and have the same IBDP and the same informant-based comparison criterion (thus, and ) but a different credibility order. In such a case, even though the two agents share the same knowledge, since the comparison criterion is defined in terms of the credibility order, the conflicts arising from the consideration of inconsistent information may be resolved differently for the two agents. Hence, in such a case, the agents’ inferences (warranted conclusions or beliefs) may differ. On the other hand, even though two agents and cannot have the same agent identifier, they can share every other component (i.e., it can be the case that , and ); clearly, in such a case, the two agents will draw the same conclusions.
In the following section we will introduce the three kinds of arguments that can be built using the knowledge represented in an IBDP. After that, the different kinds of conflicts that may arise between those arguments (referred to as attacks) will be identified. Then, in Section 4, we will introduce the notion of valid informant-based comparison criterion, used for determining the successful attacks.
Arguments and attacks
A central piece of this formalism that will allow agents to handle contradictory domain information is the notion of argument. Intuitively, an argument is a structure whose conclusion is obtained from a set of premises, informed by some agents, through the use of a reasoning mechanism. In particular, the claims of the arguments will be the tentative beliefs of the agents. When analyzing an argument for a particular belief, an agent can find other arguments, referred to as counter-arguments, that are in conflict with it. Specifically, a conflict may arise because the counter-argument contradicts some information (i.e., a premise, an intermediate conclusion or the final claim) in the argument or because it challenges one of its informants. In this situation, it is necessary to have a mechanism for comparing the conflicting arguments to decide which one prevails. This analysis leads to a dialectical process seeking to validate the arguments in conflict. The arguments that survive all possible attacks from their counter-arguments will be said to warrant their conclusions or claims, and these will be the agent’s beliefs.
Next, we will show how an agent can build different types of arguments using the defeasible domain objects and the informant rules stored in its informant-based DeLP program. As a preliminary notion, before formally defining the arguments, we introduce the concept of defeasible derivation.
(Defeasible derivation).
Let be an IBDP and L a literal. A defeasible derivation of L from , denoted , consists of a finite sequence of literals, where each literal () is in the sequence because there exists a defeasible domain object in such that: ; or , and every () is an element of the sequence appearing before ().
A derivation for a literal L is called “defeasible” because, as we will show next, there may exist information in contradiction with L or any other literal appearing in the sequence, and under certain conditions this could prevent the acceptance of L as a warranted belief. It should be noted that rules from different informants can be combined to derive a literal. Also, note that the set contains all the defeasible domain objects available for obtaining the derivation. Finally, it is important to observe that from the same IBDP it is possible to obtain several, distinct, defeasible derivations for a given literal. Furthermore, as the following example shows, from a given IBDP it is possible to obtain defeasible derivations for complementary literals (w.r.t. the strong negation “∼”).
Let us consider the IBDP from Example 2. From that program it is possible to obtain the following defeasible derivation for the literal “a”: the sequence ‘b, g, a’ that makes use of the defeasible domain objects , and . Also, allows to derive the literal “” using the defeasible domain objects , and to obtain the sequence ‘c, d, ’. Thus, from it is possible to obtain defeasible derivations for complementary literals.
As shown in Example 2, an IBDP may contain defeasible domain objects such that their heads correspond to complementary literals. This is reasonable because different informants may provide pieces of knowledge providing reasons for or against a given conclusion. Moreover, it could be the case that the same informant, under different conditions, gives reasons for or against a literal. Consequently, as illustrated in Example 4, defeasible derivations for complementary literals can be obtained from the same IBDP. In order to be able to identify coherent sets of elements within an IBDP, the notion of a contradictory set of defeasible domain objects of an IBDP is defined next.
(Contradictory set).
Let be an IBDP and . We say that the set is contradictory if and only if there exist two complementary literals L and such that and .
To illustrate this notion, let be the IBDP from Example 2. Then, the set is contradictory whereas the set is not. As a result, for instance, every superset of will also be contradictory.
Observe that backing and detracting rules are not used for obtaining defeasible derivations. As will be shown next, they will only be used to build arguments that, respectively, support or challenge the consideration of informant agents. Moreover, as will be formalized in Section 4, these new types of argument will allow to argue about the arguments’ strength. The usual definition of argument is then extended to consider backing and detracting rules, when required. As a result, we will distinguish among three different types of arguments: the first type regards arguments that conclude literals, whereas the other two deal with arguments for or against informant agents, respectively.
(Claim argument).
Let be an IBDP and L a literal. is a claim argument for the literal L, built from , if the following conditions hold:
,
,
is non-contradictory, and
is minimal: there is no satisfying conditions 2 and 3.
(Backing argument).
Let be a finite set of agent identifiers, an IBDP and . is a backing argument for the informant , built from , if , with , and the following conditions hold:
,
,
,
is non-contradictory, and
is minimal: there is no satisfying conditions 2, 3 and 4.
(Detracting argument).
Let be a finite set of agent identifiers, an IBDP and . is a detracting argument for the informant , built from , if , with , and the following conditions hold:
,
,
,
is non-contradictory, and
is minimal: there is no satisfying conditions 2, 3 and 4.
Briefly, as specified by the last clause in Definitions 9–11, all arguments share the characteristic of having a minimal and non-contradictory set of rules that allows to defeasibly derive a conclusion or the conditions for or against the consideration of an informant agent. In particular, this minimality requirement aligns with the definition of argument structures in [25], and has the aim of avoiding irrelevant information in the arguments (consequently, minimizing the points of attack). Also, the first clause in Definitions 10 and 11 is meant to ensure that the backing and detracting rules (as well as the corresponding defeasible domain objects) used for building the arguments actually belong to the IBDP from which the arguments are built. Another feature shared by backing and detracting arguments is that, when obtaining the required derivations, they cannot make use of defeasible domain objects provided by the informant they support or challenge, respectively. This constraint is meant to maintain a kind of coherence within the arguments that is not captured by the requirement of them being non-contradictory, as discussed below.
On the one hand, the third clause of Definition 10 has the aim of preventing the construction of backing arguments for a given informant, which are based on information provided by that same informant.1
Such arguments can be seen as an instance of the fallacy known as begging the question, which refers to its own assertion to prove the assertion.
On the other hand, for detracting arguments, this constraint might not seem as intuitive as it is for backing arguments. This is because, in some cases, one might take into account pieces of information provided by agent in order to distrust that agent.2
We thank one of the reviewers for pointing out this example.
For instance, suppose the existence of a medical application where knowledge is expressed through an IBDP. Now, consider the existence of a detracting rule , expressing reasons against the consideration of informant if he is not a physician. Then, if agent says he is not a physician (for instance, by providing a defeasible domain object ), one could think of using this piece of knowledge for building, together with the previous detracting rule, a detracting argument for agent .3
Recall that in this work we are assuming that the knowledge encoded through an IBDP belongs to the same topic. In this example all the information being modeled, in particular, the one claiming that is not a physician, pertains to the medical domain.
Nonetheless, it should be noted that such a detracting argument for would not only be against other arguments built using the information provided by but would also be challenging itself, since its reasoning is based on information stated by . Hence, the third clause in Definition 11 is meant to avoid situations like these, where detracting arguments would end up being self-attacking (see Definition 14 below) and therefore, inconsistent.
It is worth mentioning that backing and detracting arguments (also, their homonym rules) resemble the backing and undercutting arguments (respectively, rules) of [19]. However, their intended meanings differ: whereas backing and undercutting arguments in [19] respectively provide reasons for and against the use of specific defeasible rules (with no notion of information source), the aim of backing and detracting arguments here is to allow arguing about the arguments’ strength by providing reasons for and against the consideration of any piece of information provided by some informant agents. Also, differently from [19], in our approach there is no explicit notion of support between arguments, as it is usually done in the literature of bipolar argumentation (see [20] for an overview).
Finally note that, since informant rules are not used for obtaining defeasible derivations, claim arguments will not include informant rules. In contrast, a backing (respectively, detracting) argument will only include one backing (respectively, detracting) rule. Nevertheless, despite their common features, the three argument types are mutually exclusive. When convenient, we will abstract from an argument’s type, referring to it just as an argument. Then, given an argument built from an IBDP, we can define the notion of sub-argument as follows.
(Sub-argument).
Let be an IBDP and , two arguments built from . We say that is a sub-argument of iff .
It should be noted that every proper sub-argument of a backing or a detracting argument is a claim argument. In general, if is a proper sub-argument of (i.e., ) then is a claim argument.
Given the agent’s specification introduced in Example 3, agent will be able to build the following arguments from , among others:
, with
, with
, with
, with
, with
, with
, with
Note that , and are claim arguments. Also, is a backing argument, whereas , and are detracting arguments. In addition, observe that is a proper sub-argument of .
The above example illustrates that from a given IBDP it is possible to build arguments that are in conflict with one another, such as the claim arguments and . These conflicts become evident since the arguments’ conclusions correspond to complementary literals. On the other hand, the existence of backing and detracting arguments may lead to new kinds of conflict, that will allow to argue about the arguments’ strength in the resolution of other conflicts. The following definitions formalize the different kinds of conflict that may occur between arguments built from an IBDP, from hereon referred to as attacks.
The first kind of attack, called conclusion attack (or c-attack) captures the usual conflict in an argumentation system, where the claim of one argument contradicts a conclusion (premise, intermediate or final claim) of another.
(Conclusion attack).
Let be an IBDP, a claim argument built from , and any argument built from . We say that c-attacks at the literal L iff there exists a claim sub-argument of such that and L are complementary literals with respect to “∼”. We refer to as the disagreement sub-argument in the attack.
The second kind of attack we consider, called strength attack (or s-attack) aims at capturing the intuition that a detracting argument for a given informant is challenging any argument that makes use of a defeasible domain object provided by that informant agent. Then, since the strength of an argument is determined in terms of the strength of its informant agents, according to the adopted informant-based comparison criterion, the detracting argument is somehow attacking the other argument’s strength.
(Strength attack).
Let be a finite set of agent identifiers, , an IBDP, a detracting argument built from , and any argument built from . We say that s-attacks at the informant iff there exists a defeasible domain object .
The next kind of attack we consider, referred to as strength-defense attack (or sd-attack), corresponds to the situation where a backing argument for a given informant exists, and the backing argument attacks a detracting argument for that informant. In such a situation, we consider that the backing argument is defending the strength of the argument being attacked by the detracting argument.
(Strength-defense attack).
Let be a finite set of agent identifiers, , an IBDP, a backing argument built from , and a detracting argument built from . We say that sd-attacks.
Given the existence of a backing argument for an informant , who might originate a strength-defense attack, any detracting argument for can be considered as providing a counter-defense for the strength of the argument for whom was being challenged. This kind of attack, called strength-counter-defense attack (or scd-attack), is defined as follows.
(Strength-counter-defense attack).
Let be a finite set of agent identifiers, , an IBDP, a detracting argument built from , and a backing argument built from . We say that scd-attacks.
Consequently, whenever an sd-attack exists, an scd-attack will also exist and vice-versa. This is because, as explained above, the latter is meant to provide a counter-defense to the defense provided by a backing argument. More generally, whenever a backing argument and a detracting argument for the same informant exist, a 2-cycle of attacks will exist between such arguments (respectively, an sd-attack from the backing argument to the detracting argument, and an scd-attack from the detracting argument to the backing argument).
Finally, if an argument attacks (either c-attacks, s-attacks, sd-attacks or scd-attacks) another argument , we will say that is a counter-argument of .
Given the arguments listed in Example 5, some attacks are identified next: argument c-attacks arguments and at the literal “a”; the disagreement sub-argument in both cases is . Similarly, c-attacks at the literal “”, where the disagreement sub-argument is . In addition, the detracting arguments and s-attack because . Then, the detracting argument s-attacks because . On the other hand, for instance, the backing argument sd-attacks the detracting arguments and , and both and scd-attack argument .
These arguments and attacks are illustrated in Fig. 2. In particular, arguments are depicted with triangles, and attacks are represented with dashed arrows between the arguments. Furthermore, the circles beside the symbol of the defeasible rule “” within the arguments indicate the informant associated with each defeasible domain object in those arguments.
As the preceding example shows, conclusion attacks are symmetric in a sense. That is, if c-attacks , then the disagreement sub-argument of is such that it c-attacks . In particular, if c-attacks on its final conclusion , then c-attacks ; these situations were illustrated by arguments , and in Example 6. In contrast, strength attacks need not be symmetric. This is due to the fact that the attacking argument is a detracting argument, and no particular attacked sub-argument is identified. However, this does not prevent detracting arguments from attacking each other. For instance, it could be the case that s-attacks at the informant , and s-attacks at the informant ; what would occur in such a case is that there exist two defeasible domain objects and such that and . On the other hand, as mentioned before, whenever a backing argument sd-attacks a detracting argument , scd-attacks ; moreover, any other detracting argument for the same informant will also scd-attack (consequently, will be sd-attacked by ). That is, the existence of a backing argument and a detracting argument for the same informant leads to a two-way conflict between those arguments (the former being an sd-attack and the latter an scd-attack); again, these situations were illustrated by arguments , and in Example 6.
Given that an agent may build multiple arguments, which in turn may have several counter-arguments, in order to determine the agent’s beliefs we need to determine the undefeated arguments. To establish whether an argument is undefeated, it is necessary to explicitly account for all its counter-arguments. Let be the counter-arguments of . If any counter-argument is (according to the informant-based comparison criterion) better than or unrelated to , then is a candidate for defeating . However, if argument is better than , then this counter-argument will not be taken in consideration as a defeater for . Therefore, in order to determine the defeaters of an argument, we will make use of the informant-based comparison criterion. Then, once all the successful attackers (i.e., the defeaters) of an argument are identified, we will be able to determine its acceptance status and establish whether both the argument and its conclusion are warranted or not. These issues will be addressed in the following section.
Defeats
Agents build arguments from their knowledge bases with the aim of establishing their beliefs. However, as discussed in Section 3, the existence of conflicting information within an agent’s knowledge base leads to the existence of attacks between those arguments. Furthermore, since attacks could succeed or fail, it is necessary to have a comparison criterion to determine whether the attacking argument in a conflict prevails, in which case it becomes a defeater. In general, an attack will be considered to be effective when the set of informants from the attacked argument is not better than that of the attacking argument, with respect to an informant-based comparison criterion.
When comparing sets of informants corresponding to arguments built from an agent’s IBDP, the comparison criterion “≺” will be the one provided in the agent’s specification. In other words, the informant-based comparison criterion is modular. However, in order to be considered as valid, the criterion has to satisfy some constraints. The notion of a valid informant-based comparison criterion is formalized below. Following the usual convention, means that the set of informants is strictly better than (or preferred to) the set , and means that the set is not better than (or not preferred to) the set .
(Valid informant-based comparison criterion).
Let be a finite set of agent identifiers and the specification of an agent . We say that is a valid informant-based comparison criterion iff it holds that: for every , if , then there exist and such that (based on ); for every , (irreflexive); and ( for every , if , then (asymmetric).
Following the preceding definition, a valid informant-based comparison criterion should be based on the agent’s credibility order, and it should be irreflexive and asymmetric. We do not intend these conditions to be the only ones that can be satisfied by an informant-based comparison criterion. In contrast, we aim at establishing the minimum requirements surrounding such criteria. Specifically, the first condition is meant to link an informant-based comparison criterion with the credibility order it is based on, so that it actually makes use of the information provided by that order and does not contradict it when considering the corner case of singleton sets. That is, a valid informant-based comparison criterion is such that, when comparing unitary sets of informants and , if then ; in other words, if a unitary set of informants is better than another unitary set of informants, then it holds that the informant in the latter set is less credible than the one in the former set (according to the credibility order). On the other hand, by requiring the valid criteria to be irreflexive and asymmetric (second and third conditions), we are establishing the characteristic of them being strict.4
As will be shown later in this section, this is a key aspect in the resolution of attacks into defeats.
Finally, as stated above, a valid informant-based comparison criterion might impose additional constraints as long as it complies with the conditions established in Definition 17.
Next, we will introduce a valid informant-based comparison criterion that will be used in our examples for determining the successful attacks between the arguments built from an IBDP. This criterion, called single informant credibility criterion, is an adapted version of the single rule criterion from [43], modified to account for sets of informant agents. Intuitively, it prefers a set of informants over a set of informants if there is at least one informant in that is more credible than an informant in , and no informant in is more credible than an informant in . Formally:
(Single informant credibility criterion).
Let be a finite set of agent identifiers, a credibility order over and . We say that is preferred to , denoted , iff it holds that: there exist and such that ; and there is no and no such that .
Letbe a finite set of agent identifiers andthe specification of an agent, whereis the single informant credibility criterion. It holds thatis a valid informant-based comparison criterion.
We have to show that for everyit holds that:
If, then there existandsuch that(based on). This follows directly from the first clause in Definition
18
.
(irreflexive). Suppose by contradiction that. Then, it would be the case that there existsuch that, contradicting the second clause of Definition
18
.
If, then(asymmetric). Suppose by contradiction that. Then, by the second clause of Definition
18
it would be the case that there is noand nosuch that. However, this would violate the first clause of Definition
18
, contradicting the hypothesis that.
□
It is worth remarking that, as specified by Definition 17, a valid informant-based comparison criterion is not required to be transitive. As discussed before, this does not imply that any valid criteria cannot be transitive; thus, transitive as well as non-transitive criteria can be considered as long as they meet the requirements imposed by Definition 17. Then, by not imposing such a constraint on valid criteria, we allow for a wider family of criteria to be considered. Among others, this allows us to consider the single informant credibility criterion which, as introduced in Definition 18, is not transitive. To illustrate the fact that this criterion does not satisfy transitivity, let us consider the following example.5
We thank one of the reviewers for suggesting this example.
Consider the set of agents and a credibility order such that and . By applying the single informant credibility criterion , it holds that and . However, since and are incomparable w.r.t. , it does not hold that and thus, the criterion is not transitive. Notwithstanding this, the non-transitivity of the single informant credibility criterion does not undermine its usefulness. In fact, this criterion is analogous to the rules priorities criterion from the standard literature of structured argumentation (see e.g., [25, Def. 3.7]). On the other hand, as an example of a transitive and valid informant-based comparison criterion, we can consider the max-max lifting criterion introduced in [9] when taking a credibility order as the basis of the ≻ relation (see Section 6 for a detailed discussion).
We will now turn to establish the conditions under which the attacks introduced in Section 3 succeed and become defeats by making use of a valid informant-based comparison criterion. As mentioned before, in general, an attack will succeed if the set of informants from the attacking argument is not worse than that of the attacked argument, with respect to an informant-based comparison criterion. Hence, for each kind of attack, we will establish the sets of informants to be accounted for by the comparison criterion. For this, we need to formally characterize the set of informants associated with an argument.
(Argument informants set).
Let be a finite set of agent identifiers and an argument built from an IBDP. The informants set of is .
Having established a mechanism for identifying the set of informants associated with an argument, we now turn to formalize the different kinds of defeat that may occur between a pair of arguments built from an IBDP. In the following, whenever we want to refer to a generic argument (without caring for its conclusion or the informant it supports or challenges), for convenience we will sometimes write instead of for claim arguments (respectively, instead of for backing and detracting arguments); this is because an argument can be unequivocally identified by its associated set of rules. Then, for instance, and will sometimes be written as .
(Conclusion defeat).
Let be a finite set of agent identifiers, the specification of an agent , and , two arguments built from such that c-attacks at the literal L. We say that c-defeats (equivalently, is a c-defeater of ) iff the disagreement sub-argument of is such that . In particular, if , we say that is a proper c-defeater of ; otherwise, we say that is a blocking c-defeater of .
(Strength defeat).
Let be a finite set of agent identifiers, the specification of an agent , and , two arguments built from such that s-attacks . We say that s-defeats (equivalently, is an s-defeater of ) and, in particular, is a proper s-defeater of .
(Strength-defense defeat).
Let be a finite set of agent identifiers, the specification of an agent , and , two arguments built from such that sd-attacks . We say that sd-defeats (equivalently, is an sd-defeater of ) iff . In particular, if , we say that is a proper sd-defeater of ; otherwise, we say that is a blocking sd-defeater of .
(Strength-counter-defense defeat).
Let be a finite set of agent identifiers, the specification of an agent , and , two arguments built from such that scd-attacks . We say that scd-defeats (equivalently, is an scd-defeater of ) iff . In particular, if , we say that is a proper scd-defeater of ; otherwise, we say that is a blocking scd-defeater of .
There exist some differences in the way in which the different types of attack are resolved. Specifically, the differences rely on the sets of informants that are compared in each case. For the resolution of c-attacks, it suffices to compare the set of informants from the attacking argument with the set of informants from the disagreement sub-argument. Note that the resolution of c-attacks in this way is analogous to the standard resolution of rebutting attacks in the literature of structured argumentation (see e.g., [25,36]). When resolving s-attacks, it is important to recall the nature of the attacking argument, which provides reasons against the consideration of a given informant used by the attacked argument. To prevent the success of an s-attack, the attacked argument should somehow give reasons for the consideration of that informant, and they should be provided by informants that are not worse than the ones associated with the attacking argument. As a result, since the attacked argument does not give reasons for the consideration of the challenged informant, the s-attack will always succeed (i.e., no comparison between sets of informants is made). Again, the resolution of s-defeats relates to the way in which undercutting attacks are resolved in the literature.
Then, sd-attacks and scd-attacks are handled analogously. In these two cases, there exists a conflict between a backing argument and a detracting argument, where they respectively provide reasons for and against the consideration of a given informant. Therefore, in the resolution of these types of attack, the entire sets of informants associated with the attacking and the attacked argument are compared. Given the resemblance between backing and detracting arguments as proposed here and the backing and undercutting arguments of [19], it can be noted that sd-defeats and scd-defeats somehow relate to the implicit defeats of [19], and are resolved analogously; notwithstanding this, there is a clear difference between them since our approach compares the sets of informant agents of the backing and detracting arguments whereas [19] makes use of a preference relation which might not take into account the information sources of arguments.
Finally note that, in all cases where sets of informants are compared to determine the success of an attack (thus turning it into a defeat), the traditional form of resolution of [3] is adopted. Namely, in our approach, an attack succeeds if the set of informants from the attacked argument is not better (w.r.t. the informant-based comparison criterion) than the set of informants from the attacking argument. In contrast, works like [6,31] consider that preferences play an additional role, serving to repair the attack relation in order to account for conflicts derived from the preferences, even in cases where those conflicts might not have been originally expressed within the attack relation. Among others, they both consider the existence of a defeat from an argument to an argument in a scenario where attacks (and not vice-versa) but is strictly preferred to .
Let us consider , the specification of agent given in Example 3, and the arguments listed in Example 5. Also, suppose that is the single informant credibility criterion from Definition 18. Then, we identify the following defeats (among others): c-defeats and . In both cases, the disagreement sub-argument is such that . Then, since , . In particular, , so is a proper c-defeater of and . Continuing with the resolution of attacks from Example 6, the detracting arguments and proper s-defeat . Similarly, the detracting argument is a proper s-defeater of . On the other hand, the backing argument proper sd-defeats since , and . Finally, the detracting argument is a proper scd-defeater of ; this is because , and .
The defeats are illustrated in Fig. 3. The notation for the arguments is the same as the one used in Fig. 2; on the other hand, defeats are represented with solid arrows between the arguments.
To determine whether an agent can accept a literal L as a belief it is necessary to find out if, from its IBDP , it is possible to build an argument that ends up undefeated after all things considered. Naturally, this will require the consideration of all possible defeaters for . Then, given a defeater , defeaters for it will also have to be considered, as well as the defeaters for those defeaters, and so on. This dialectical analysis leads to characterizing the notion of an argumentation line, which constitutes a sequence of arguments where each argument is a defeater of its predecessor.
(Argumentation line).
Let be a finite set of agent identifiers, the specification of an agent , and an argument built from . An argumentation line for is a sequence of arguments built from , denoted , where each element of the sequence is a defeater of its predecessor ().
Given an argumentation line Λ for an argument , every argument in an odd position (e.g., , , if they exist) is said to be a supporting argument for because they reinstate ; the set of such arguments will be denoted as . Analogously, arguments in even positions of Λ are called interfering arguments, and the corresponding set will be denoted as .
Given the specification of agent introduced in Example 3, the arguments built from its IBDP (Example 5) and the defeats between them listed in Example 7, we obtain (among others) the following argumentation lines for : , whose set of supporting arguments is and its set of interfering arguments is ; , with and ; and , whose set of supporting arguments is and its set of interfering arguments is .
There may exist argumentation lines that lead to fallacious chains of reasoning. In order to avoid those, we impose some restrictions on argumentation lines, to distinguish the acceptable argumentation lines. The first situation we want to avoid is to have infinite chains of reasoning; this is captured by the first and third clauses in Definition 26, by requiring an acceptable argumentation line to be a finite sequence, and to avoid the introduction of repeated arguments and disagreement sub-arguments. Another constraint imposed on acceptable argumentation lines is a kind of consistency within the sets of supporting and interfering arguments, in order to prevent an argument from being defended by another argument that is in conflict with it; this intuition is modeled in the second clause of Definition 26, requiring the sets of supporting and interfering arguments of an argumentation line to be non-contradictory.
Let us now analyze the inclusion of blocking defeaters in an argumentation line. The existence of a blocking defeat from to in an argumentation line implies that the sets of informants of the two arguments were incomparable under the adopted comparison criterion.6
Note that the two sets of informants cannot be equally preferred since valid informant-based comparison criteria are required to be asymmetric, in line with the credibility order they are based.
In addition, if there exists an argument that is a blocking defeater of , it would imply that the comparison criterion could not resolve that conflict either. As a result, if were to be included in the corresponding argumentation line, it would somehow imply that prevails over , just because it has argument supporting it (which, in turn, was not better than ). To avoid these issues, consecutive blocking defeaters are not allowed in an acceptable argumentation line, and this is captured in the fourth clause of Definition 26.
Next, let us move to considering the inclusion of s-defeats in an argumentation line. As expressed before, s-defeats (corresponding to s-attacks, which always succeed), are aimed at challenging the strength of the defeated argument by targeting one of its informant agents. This is because, as discussed earlier, we consider that the strength of an argument is linked to the credibility of its informant agents. However, it should be noted that the existence of an s-defeat towards an argument is not, on its own, sufficient to establish that the defeat actually took place because of challenging the defeated argument’s strength. This is due to the fact that not every informant providing information used for building an argument is relevant for determining its strength. For instance, as expressed in Definition 20, the strength of a claim argument that c-attacks an argument , with being the disagreement sub-argument, will be determined by a subset of its informants such that , where is the entire set of informants of argument ; specifically, as shown by Definition 17, some informants might not be accounted for when comparing the two arguments. Consequently, the strength of an argument is determined in the context of the resolution of an attack and, as a result, we consider that the strength of an argument (possibly established by alternative subsets of its informants) is not absolute; rather, it is relative to the resolution of each attack it is involved in, as different sets of agents may be considered in each case to determine the argument’s strength.
To capture these intuitions, we next introduce the notion of strength-determining set of informants of an argument in an argumentation line, which corresponds to a set of informants of the argument that provides the necessary strength for its inclusion as a defeater of its predecessor in the line. In particular, as will be shown later in Definition 26, this notion will serve to identify s-defeaters that indeed defeat an argument by challenging its strength, because they provide reasons against one of its strength-determining informants. For instance, consider the argumentation line which is a sub-sequence of the line (also, of the line ) from Example 8, where proper c-defeats argument ; in this case, the strength-determining sets of informants of would be and , since the informants in these sets are individually better than every informant used by the disagreement sub-argument . Furthermore, since the strength-determining set of informants of an argument may not be unique, the following definition allows to obtain every set of informants meeting that requirement:
(Strength-determining sets of informants).
Let be a finite set of agent identifiers, the specification of an agent , an argument built from and Λ an argumentation line ending in . We define the strength-determining sets of informants ofin Λ as the elements in the set returned by the function defined in Fig. 4.
Function characterizing the strength-determining sets of informants.
Note that, for the first argument in an argumentation line, the function returns a unitary set containing the entire set of informant agents from the argument; this is so because the argument was not introduced in the line for being a defeater of its predecessor. Hence, since its strength was not accounted for in the resolution of a previous attack, we deem all its informants as able to determine its strength. The same situation occurs when the argument under consideration is a detracting argument that originates an s-defeat; in such a case, we consider its entire set of informants to be the only strength-determining set because no comparison is made for the resolution of the s-attack into an s-defeat. In every other case, each set of informants that is able to make the argument defeat its predecessor (through a proper or a blocking defeat, respectively) will be a strength-determining set of informants.
Taking the notion of strength-determining set of informants into account, we will only consider the inclusion of an s-defeater in an acceptable argumentation line in cases where the detracting argument targets an informant belonging to (at least) one set of strength-determining informants of its predecessor argument in the line; this intuition is captured in the last clause of Definition 26. Finally, regarding sd-defeaters and scd-defeaters, if they appear in an argumentation line it is as a consequence of the existence of a previous s-defeater. Therefore, no additional considerations have to be taken into account for their inclusion in an acceptable argumentation line. As a result, the notion of acceptable argumentation line is formalized below:
(Acceptable argumentation line).
Let be a finite set of agent identifiers, the specification of an agent , an argument built from and an argumentation line for . We say that Λ is an acceptable argumentation line iff it holds that:
Λ is a finite sequence.
The set of supporting arguments and the set of interfering arguments are non-contradictory.
No argument in Λ is an argument appearing earlier in Λ or a disagreement sub-argument of an argument appearing earlier in Λ ().
For every argument of Λ such that is a blocking defeater of , if there exists in Λ, then is a proper defeater of .
For every detracting argument of Λ such that is a proper s-defeater of challenging the informant (hence, ), there exists such that , where is the sub-sequence of Λ ending in .
Given the argumentation lines , and from Example 8, and are acceptable argumentation lines whereas is not because it does not satisfy condition 5 of Definition 26 ( does not belong to any of the strength-determining sets of informants of in ). The set of all acceptable argumentation lines for a given argument are then gathered to conform a tree structure containing all possible lines of reasoning starting with . Next, we formally define this structure by introducing the notion of dialectical tree.
(Dialectical tree).
Let be a finite set of agent identifiers, the specification of an agent and an argument built from . A dialectical tree for, denoted , is a tree such that its nodes are arguments built from and the edges denote defeats between those arguments, and the following conditions hold:
is the root of .
If is an acceptable argumentation line for and there is no argument built from such that is an acceptable argumentation line for , then is a leaf in and Λ is the branch of starting from the root down to the leaf .
Once built, the dialectical tree is marked in order to determine the acceptance status of the root argument. The marking criterion will mark the nodes in the tree as undefeated (U) or as defeated (D). Briefly, leafs of the tree are U, while inner nodes are D if they are defeated by a U claim argument, a U backing argument or a U detracting argument originating a scd-defeat, or if they are defeated by U detracting arguments originating s-defeats that target an informant in each of its strength-determining sets of informants.
(Marking criterion).
Let be a finite set of agent identifiers, the specification of an agent , an argument built from and the dialectical tree for . If N is a node in , it will be marked as:
Dif N is the root of or it is an inner node corresponding to an argument , and either:
N has a child marked U corresponding to a c-defeater, an sd-defeater or an scd-defeater of ; or
for each , N has a child marked U that corresponds to an s-defeater of such that , where is the argumentation line corresponding to the arguments in the branch starting from the root down to the node N.
Uotherwise.
Consider the specification of agent introduced in Example 3 and the arguments listed in Example 5. Figure 5 depicts the marked dialectical tree for argument .7
Note that, to save space, the tree is depicted horizontally.
This tree has two acceptable argumentation lines, namely and from Example 8. The notation for arguments and defeats in the figure is the same as in Fig. 3. Also, each argument is marked in its left vertex as defeated or undefeated, with a circle containing a D or a U, respectively. Observe that the nodes corresponding to are marked U (they are leafs). Then, the node corresponding to is marked D because it has a child marked U corresponding to an scd-defeater. Consequently, since is marked D, is marked U. However, note that is marked U even though it has two children corresponding to s-defeaters ( and ) that are marked U. This is because there is no s-defeater for in targeting the informant (recall that , with ). Finally, since is marked U, the root of is marked D.
A marked dialectical tree embodies a dialectical analysis considering every possible argument an agent can build for and against the root argument in the tree. Hence, if the root argument is marked as U, it means that the conclusion of that argument is warranted, and the agent can accept it as a belief. Moreover, the existence of one argument marked as U in its dialectical tree is sufficient for the agent to accept L as a belief. On the contrary, if every argument for L is marked as D in its own tree, then the literal L will not be warranted and thus, the agent will not accept it as a belief. The notions of warranted argument and warranted literal are formalized next:
(Warranted claim argument – warranted literal).
Let be a finite set of agent identifiers, the specification of an agent and a claim argument built from . We say that warrants the argumentand the literal L iff is marked as U in the dialectical tree .
Consider agent specified in Example 3 and the marked dialectical tree from Example 9. Agent does not warrant the argument since, as depicted in Fig. 5, the root of is marked as D. Consequently, since there are no other claim arguments for “h”, agent does not warrant the literal “h” (hence, it will not accept it as a belief).
Suppose the detracting rule is added to the IBDP of agent , leading to the new agent’s specification . In this scenario, the dialectical tree will be the one illustrated in Fig. 6. Differently from Example 9, is now marked D because for each , has a child marked U that corresponds to an s-defeater targeting an informant in ; specifically, the detracting arguments and are s-defeaters targeting , whereas the detracting argument (with ) is an s-defeater targeting . As a result, the root of is now marked U and thus, agent warrants the argument and the literal “h”, accepting the latter as a belief.
As introduced above, the beliefs an agent has are determined by the set of literals it warrants. Given this, it is to be expected that the set of warranted beliefs from an agent is consistent. In the remainder of this section we will formally show that an agent does not warrant complementary literals, as expressed by Theorem 1. For this purpose, we will first show some intermediate results, given by Lemmas 1 and 2.
Letbe a finite set of agent identifiers,the specification of an agent, andan argument built from. Ifwarrants, then for every claim argumentbuilt fromsuch thatc-defeatsit holds that the root node ofhas a child corresponding to argumentthat is markedU.
Sincec-defeats, by Definitions
24
,
26
and
27
it follows directly that there exists a child N of the root ofcorresponding to. Now we have to prove that N is marked asUin. Given that by hypothesiswarrants, by Definition
29
the root of the dialectical treeis marked asU. By the characterization of dialectical trees, a part ofcan appear as a sub-tree of, rooted in the node N. Then we have the following cases:
(the complete tree) is a sub-tree of, rooted in N. In this case, by Definition
28
, the node N will be marked asUin.
A part of(not the complete tree) is a sub-tree of, rooted in N. In this case, there exists a node S inthat is not a node in the sub-tree ofrooted in N. If the node S corresponds to an interfering argument in, then by Definition
28
it holds that not having S indoes not change the marking of N, as it was marked asUin. Consequently, N will be marked asUin. Let us now suppose that the node S corresponds to a supporting argumentin. Therefore, there exists an acceptable argumentation linecorresponding to a branch of. Then, if the node S corresponding todoes not appear in the sub-tree ofrooted in N and the root ofhas a child corresponding to, it should be the case that the argumentation line(i.e., the argumentation line starting withand continuing with the sub-sequence of Λ ending in) is not acceptable because the inclusion ofviolates some condition of Definition
26
. Hence, it must be the case that either:
The set of interfering arguments of(i.e., the set containingand every other argument in the interfering set of, such as) is contradictory. However, such interfering arguments (includingand) appear in Λ as supporting arguments. Thus, this would imply that Λ is not an acceptable argumentation line because its set of supporting arguments would be contradictory as well. Contradiction.
is an argument appearing earlier inor is a disagreement sub-argument of an argument appearing earlier in. Without loss of generality, these two cases are covered by:is a sub-argument of an argument appearing earlier in. Then, we have to consider the following cases:
is a sub-argument of an argument different fromin. Since the only difference betweenand Λ prior tois the inclusion ofas the first element in the sequence,would be a sub-argument of an argument appearing earlier in Λ, leading Λ to be a non-acceptable argumentation line. Contradiction.
is a sub-argument of. Since by hypothesisis a claim argument, by Definitions
9
and
12
,is also a claim argument. Then, given thatappears right after argumentin Λ and, we have thatc-defeats; consequently, the setis contradictory. Moreover, sinceit holds thatis also contradictory. Hence,(i.e., the sub-sequence ofending in) would not be an acceptable argumentation line, which contradicts the assumption thatwas not acceptable because of the inclusion of.
leads to two consecutive blocking defeats in. Sinceis a supporting argument in Λ and, as previously shown, is different from, it holds thatis at least the third element in the sequence corresponding to Λ. Thus, the consecutive blocking defeats would also occur in Λ, meaning that Λ is not an acceptable argumentation line. Contradiction.
is a detracting argument appearing right after argumentin, and there is no set of informantsholding that, where(i.e., the sub-sequence ofending in). Sinceis a supporting argument in Λ and, as previously shown, is different from, it holds thatis at least the third element in the sequence corresponding to Λ; hence,is at least the second element in Λ. Consequently,is different fromand the argument precedingin(also, in) will appear precedingin Λ as well. By Definition
25
the strength-determining sets of informantswill be obtained by comparing the informants ofwith the informants of the argument preceding it in(also, in), which is the same argument preceding it in Λ. Thus, it holds that, where(i.e., the sub-sequence of Λ ending in). As a result, ifwas targeting a non-determining informant ofin, it was doing so in Λ as well, meaning that Λ is not an acceptable argumentation line. Contradiction.
□
Letbe a finite set of agent identifiers,the specification of an agent, andan argument built from. Ifwarrants, then for every claim argumentbuilt fromthat is c-defeated byor that c-defeatsit holds thatdoes not warrant.
We have to consider two cases:
c-defeats. Since by hypothesiswarrants, by Lemma
1
the root ofhas a child node corresponding tothat is markedU. Therefore, by Definition
28
, the root ofwill be marked asDand thus, by Definition
29
,does not warrant.
c-defeats. Let us suppose by contradiction thatwarrants. If that were the case, by Lemma
1
the root ofwill have a child node corresponding tothat is markedU. Hence, by Definition
28
, the root ofwill be marked asD. Consequently, by Definition
29
would not be warranted by, which contradicts our hypothesis.
□
Letbe a finite set of agent identifiers,the specification of an agent, and L, H two complementary literals. Ifwarrants L, thendoes not warrant H.
Ifwarrants L, by Definition
29
, there exists a claim argumentbuilt fromthat is warranted by. Let us now suppose by contradiction thatwarrants H. Then, by Definition
29
, there would exist a claim argumentthat is warranted by. Hence, by Definition
13
andc-attack each other. Moreover, by Definition
20
, it will be the case thatis a proper c-defeater of,is a proper c-defeater of, orandare blocking c-defeaters of one another. Then, by Lemma
2
this would imply thatdoes not warrant, contradicting the hypothesis that L is warranted by. □
Related work
The literature of argumentation offers a wide and extensive variety of approaches accounting for the notion of argument strength, some of which will be discussed in this section. For instance, approaches like [3,5] resolve conflicts by identifying the stronger arguments through a general preference relation. On the other hand, works like [13,28] define the strength of an argument through a formula yielding a numerical value. This formula is based on the intuition that the strength of an argument is inversely proportional to the strength of its attackers, with the aim of codifying the likelihood of an argument to be ultimately defeated. Similarly, [35] proposed a game-theoretic measure of argument strength, where the strength of each argument is calculated in a way such that if an argument is attacked then its strength falls, but if the attack is in turn attacked, then the strength of the original argument rises. Related to these, approaches to value-based argumentation [10] consider that the strength of an argument depends on the social values it advances, and determining whether the attack of one argument on another succeeds depends on the comparative strength of the values advanced by the arguments concerned. Also, in [41] an approach combining the ASPIC argumentation system and the fuzzy set theory is proposed, where argument strength is computed by using a t-norm aggregating the importance of its involved premises and rules. Similarly to us, the latter attaches information to the basic elements of their system to compute argument strength, which in turn is used to determine acceptable arguments. However, in that work it is not possible to challenge the strength of an argument (even when undercutting defeaters could be seen as similar to our detracting defeaters, an undercut challenges the applicability of a rule instead of the strength of the argument as detracting defeaters do).
We can also identify a group of approaches that, explicitly or implicitly, make use of a notion of credibility or trust to account for the arguments’ strength. Given the close relationship between our proposal and these approaches, we devote the rest of this section to discuss the similarities and differences with some of them.
In [38] and [44] an argumentation formalism is proposed which, as part of its reasoning process, uses information about trust to measure the arguments’ strength. This formalism is described as a set of graphs and, to determine an agent’s beliefs, the authors propose a model that accounts for the trust in the information that is used for building the arguments. Like ours, their approach is presented in a multi-agent setting, where informant agents can have different levels of credibility and these credibilities are used to measure the arguments’ strength. In contrast to our proposal, where each agent has its own credibility order (completely independent from the other agents’), they use a centralized notion of trust that is codified in a shared trust network. This global network holds information about how agents trust each other and can be used to obtain an agent-centric trust network that represents the viewpoint of a particular agent. Although from these graphs it is possible to determine a credibility order for each agent, these orderings are strongly dependent on the connections in the global network.
Similarly to our formalism, each piece of information in [38,44] is linked to an agent which establishes how credible that information is, and the strength of an argument is determined by the credibility of the information it is based on. Then, [38] and [44] use an argumentation inference mechanism to deal with a potentially contradictory belief base. In that context, arguments are built to support pieces of information that can be consistently inferred from the belief base, and the strength measures are used to decide between conflicting arguments. In addition to that, our approach allows to construct arguments for and against the consideration of informant agents. Therefore, differently from theirs, an agent in our approach will be able to reason and argue about the strength of its arguments, in addition to arguing about the domain information stored in its knowledge base.
Another significant difference between [38,44] and the work reported in this paper is that they use numerical values to establish the trust relation among agents, leading to a total order over the set of agent identifiers; in contrast, we use symbolic information in the form of a strict partial order over that set. Such contrast leads to different approaches for measuring the arguments’ strength. On the one hand, they compute the strength of an argument using a formula based on numerical values assigned to the agents that provided the information used for building the argument. On the other hand, the strength of an argument in our approach is not absolute; instead, it is relative to the resolution of the conflicts the argument is involved in, and is established by the sets of strength-determining informants of the argument (according to a valid informant-based comparison criterion). Consequently, in their approach, the central component for determining the arguments’ strength is the formula to be used, whereas in ours it is the adopted informant-based comparison criterion.
The work reported in [17] relates to our proposal in that they also make use of the credibility of informant agents as a source of argument strength. Similarly to us, they associate arguments with their information sources (and multiple sources can be associated with the same argument); however, they abstract away from the origin of such credibility. That is, their credibility function associates an informant with a set of arguments, without specifically attaching the informant to a particular piece of knowledge within the argument. In contrast, we associate each defeasible rule in an argument with its informant agent to conform a defeasible domain object. Then, in [18] the authors extend their work to combine credibilities of informants of facts, assumptions and conclusions in order to determine the arguments’ strength. Like in our approach, defeats among arguments in [17,18] are defined by accounting for a notion of argument strength based on the credibility of their informants. However, a key aspect that differentiates our approach from theirs is the fact that their credibility function establishes a total order over the set of informants, whereas our credibility orders are assumed to be strict partial orders. Finally, as remarked above for other approaches, another difference between our work and [17,18] is the fact that we provide the means to reason about the arguments’ strength by allowing to build arguments for and against their different informant agents; in contrast, the approach of [17,18] does not provide such a possibility.
In this paper we introduced a particular informant-based comparison criterion (see Definition 18) to measure and compare the arguments’ strength. As discussed in Section 4, the single informant credibility criterion resembles the rules priorities criterion (see e.g., [25]) which takes preferences among rules to then define preferences among arguments (which are in turn expressed as sets of rules). That is, the rules priorities criterion lifts preferences over rules to preferences over sets of rules. Similarly, an informant-based comparison criterion makes use of a credibility order (i.e., a relationship over informant agents) to establish preferences over sets of informants; in other words, preferences over informant agents are lifted onto preferences over sets of informants. The literature offers a variety of approaches dealing with the issue of lifting preferences, such as [9,23,33,34] (see also [8] for an extensive overview on the issue of lifting an order relation on a set X to an order on the family of all non-empty subsets of X).
On the one hand, in [34] the authors propose natural extensions of the versions of five axioms introduced in [8] and also state that not every axiom is desirable in every situation. Regarding the axioms or conditions associated with a lifting principle (here, a valid informant-based comparison criterion), we can note the following difference. Whereas in our approach the credibility order (which is then lifted to define an informant-based comparison criterion) should be irreflexive, asymmetric and transitive, [34] assumes an order relation on a set X that has to be reflexive, antisymmetric, transitive and total. For instance, regarding the reflexivity vs. irreflexivity requirement, we can note that a valid informant-based comparison criterion is required to be irreflexive in order to align with the strict nature of the credibility order it is based on. An exhaustive analysis of our approach in the context of their proposed axioms is left as future work, including the study of alternative or additional conditions to be imposed either on credibility orders or on valid informant-based comparison criteria.
In [9] the authors argue that, in the context of structured argumentation, the support provided by an argument for its conclusion is determined by the degree of support of its premisses, and by the degree of support provided by the inference rules applied in its construction. Then, among other things, the authors discuss several alternatives for lifting qualitative orderings on premises and for lifting qualitative orderings on defeasible rules. It should be noted that, when accounting for orderings on different types of elements conforming the arguments, these orderings should be somehow combined when lifting preferences. Our proposal is such that domain knowledge in an IBDP is only expressed through the defeasible domain objects, which associate defeasible rules with their informant agents. In particular, our lifting criteria (the informant-based comparison criteria) only account for one dimension: the credibility order over the informant agents.
When considering the conditions imposed on valid informant-based comparison criteria and the conditions imposed by the different liftings discussed in [9], we can note the following: our conditions are meant to characterize a family of criteria whereas they describe specific kinds of liftings on qualitative orderings of elements. However, we can find a relationship between the liftings studied in [9] and the family of valid criteria characterized here, and we plan to explore this relationship as part of future work. For instance, consider the max-max criterion to lift orderings on premises to orderings on sets of premises X and Y: iff for every there is a such that (with ≻ meaning strictly preferred to). Here, we can see that the max-max criterion satisfies the condition of being based on the original ordering (i.e., at least one pair of elements of X and Y is related by ≻), as in the first clause of Definition 17. Then, it can be shown that the max-max criterion also satisfies the property of yielding irreflexive orderings when applied on transitive orderings over premises. For this, consider a set X and suppose by contradiction that . Hence, for every there exists such that , and this must also be satisfied for such . So, because of the transitivity of the ≻ relation over premises, at some point a cycle will exist, contradicting the fact that it is a strict ordering. Consequently, the max-max lifting criterion satisfies irreflexivity (second clause of Definition 17). Finally, suppose the max-max criterion relates two sets of premises X and Y as follows: and . Here, if , by definition of max-max it holds that for every there exists such that . Then, if also , it must be the case that for every there is such that . So, because of the transitivity of the ≻ relation over pairs of premises, for every it holds that there exists such that (i.e., ), contradicting the fact that the max-max criterion yields irreflexive orderings on sets of premises. Consequently, the max-max criterion also satisfies the third clause of Definition 17 and thus, it could serve as a valid informant-based comparison criterion when considering a credibility order as the base ≻ relation.
Another work that relates to our proposal is [32], where the authors use an argumentation mechanism based on trust as a layer of a belief revision process carried out by agents dealing with (potentially conflicting) opinions about their pairs. In their argumentation approach, trust is used for building a preference ordering amongst arguments, thus codifying their strength. For this, they first aggregate the information about different opinions for the same proposition. Then, using these aggregated propositions, they build arguments whose trustworthiness is assessed using a conjunctive fusion operator over the opinions forming the argument. This assessment considers the number of agents and information pieces that were needed for building the argument. Even though [32] does not account for the fact that arguments can be constructed to challenge an informant agent (hence, other arguments’ strength), it could be interesting to adapt their ideas in our proposal. For that purpose, we can think of two alternatives: either provide a comparison criterion that encodes the above mentioned strategies, or extend the notion of acceptable argumentation line (Definition 26) to consider preferences codifying those strategies.
A qualitative bipolar argumentative modeling of trust is proposed in [39]. Like in our proposal, their approach is qualitative and only a finite number of levels is assumed in the trust scale. On the other hand, in contrast with our proposal, they use a bipolar argumentative approach where trust and distrust can be independently assessed. There, an agent can evaluate its trust into an object X (that can be either a source or another agent) on the basis of two types of information: the observed behavior of X, and the reputation of X according to the other agents. Reputation information is then viewed as an input the agent uses for revising or updating its own trust evaluation, based on its perception. The approach proposed in [39] is such that two kinds of arguments in favor of trusting an agent (either establishing that a good point is reached or a bad point is avoided), as well as two kinds of arguments against trusting an agent (either indicating that a bad point is satisfied or that a good point is not reached) can be constructed. These four kinds of arguments are based on an inference rule and the trust evaluation of the agent, which is represented with an interval over a discrete scale, with the intended meaning that the trust is not larger than nor smaller than . Clearly, our work relates to [39] in that we also allow to build arguments for or against an informant agent; however, our arguments do not encode reasons for trusting or distrusting such an agent, but they provide reasons for or against the consideration of information provided by it. Moreover, as another difference, even though [39] indicates some basic mechanisms leading to the revision of trust values, that paper is mainly focused on evaluating trust rather than on integrating the trust values as a measure of argument strength in an argumentation process.
The authors in [29,30] adopt a symbolic approach to model credibility using two global relations: the trust relation and the distrust relation. These relations, together with the set of agents, constitute a trust system where a pair in the trust (respectively, distrust) relation expresses that agent a trusts (respectively, distrusts) agent b. Their formalism aims at determining whether an agent trusts another taking into account the potential conflicts that may appear when the trust and distrust relations are jointly analyzed in the trust system. To that end, they follow an argumentation approach in which arguments represent a position for an agent to either trust or distrust a peer; those arguments are similar to our backing and detracting arguments, respectively. Additionally, when considering an advanced version of their system, each agent is also provided with a partial order defined amongst its peers, and they use this order to codify the efficacy in which an agent trusts its peers (aiming to model a degree of trustworthiness or reputation). Like us, they use these efficacy orders to provide strength to their arguments. However, the goal of the argumentation system in the two approaches differs. Their aim is to decide if an agent trusts another or not, given the trust and distrust relations. In contrast, we aim at establishing the beliefs an agent has, which correspond to the pieces of information (literals) it warrants. Notwithstanding this, similarly to ours, their proposal seeks to enrich argumentation systems with strength measures that account for (potentially conflicting) information about trust, and use these measures to decide which arguments and conclusions prevail.
The work reported in [4], similarly to ours, presents an agent-based argumentation approach for reasoning about beliefs and information received from other agents. There, beliefs are also used to represent how trustworthy the information sources are to a given agent. They identify six forms of trust that can appear as part of the formulas in the agent belief base. From their belief bases arguments are built, conflicts among them are identified and then resolved. In particular, besides the usual conflicts in structured argumentation, the authors identify several types of attack that arise from the semantics of the six forms of trust considered. Therefore, like us, they allow to challenge and support the credibility or trustworthiness of an informant; nevertheless, there are some differences. In their approach each form of trust is binary: the agent either trusts or does not trust an informant. In contrast, in our approach agents are ordered using a strict partial order and thus, it is possible to establish whether an informant is more credible than another. There are also differences on how trust relates to argument strength. In their approach, trust forms are used for constructing arguments and do not directly affect argument strength. On the other hand, our approach uses the level of credibility as the measure to define the arguments’ strength. They consider the notion of strength when they introduce graded beliefs, where a grade is attached to the beliefs operator. Using these grades they compute the strength of an argument as the weakest link. However, unlike our approach, they do not provide any mechanism to challenge the strength of an argument. Finally, it is worth to mention that these differences also apply to a comparison between our proposal and the work reported in [48].
Conclusion
In this paper we have presented an argumentative reasoning formalism where the credibility of informants plays a central role, as it allows to determine the arguments’ strength. Our formalism was developed in a multi-agent setting where agents share domain knowledge. There, each agent may obtain information from other informant agents and also has an assessment of how credible these informants are. Agents are equipped with the argumentative machinery, allowing them to reason with the potentially conflicting information in their knowledge bases to finally determine their warranted beliefs. In our approach, defeasible rules (which represent domain knowledge) are associated with their informant agents. Also, we introduced two new kinds of rules (backing and detracting rules) in order to be able to argue about the contexts in which the domain knowledge provided by the informant agents should be used or not. In other words, these rules are used to express reasons for and against the consideration of informants, respectively. From all this knowledge, an agent will be able to construct arguments to support its inferences. In addition, each agent has a credibility order among its informant agents and a comparison criterion used to assess the strength of the conflicting arguments built from its knowledge base.
As shown before, our informant-based approach is such that the strength of an argument is determined by the credibility of its informant agents. To that end, the comparison criterion in an agent’s specification is based on the agent’s credibility order. In particular, we have shown that the strength of an argument in our approach is not absolute, but it is relative to the resolution of the conflicts the argument is involved in. Then, it could be the case that some informant providing information for building an argument is relevant for establishing the argument’s strength in some cases, but not in others. In this context, the incorporation of backing and detracting rules allows agents to argue about the arguments’ strength. Specifically, backing rules allow to express reasons for the consideration of informant agents, whereas detracting rules enable to express reasons against the consideration of information provided by them. Using these rules we defined new types of argument which, together with the classic arguments supporting conclusions, are considered by the argumentation machinery to establish the beliefs an agent has. Finally, we have formally shown that the warranting process employed by our argumentative approach is sound, preventing an agent from warranting contradicting conclusions.
It is worth noting that the defeasible domain objects within an informant-based Defeasible Logic Program (IBDP) establish a correspondence between defeasible rules and their informant agents. The fact that, differently from standard DeLP programs, an IBDP does not include strict rules may appear as a limitation of our approach. However, recall that strict rules are provided in DeLP as a representational tool that gives the possibility of expressing the indefeasible nature of the relation between the body and head of such rules, making them indisputable. In contrast, in an IBDP, a domain object is expressed as a pair containing a rule and the informant of that rule. In this context, such rules are always defeasible since they come attached with the credibility of their informant agent. Therefore, our approach only accounts for defeasible rules (hence the name of the defeasible domain objects) since our main focus was on how the arguments’ informants affect and determine the arguments’ strength. Moreover, regarding the absence of strict rules, we would like to remark that a number of different applications of DeLP (e.g., see [1,2,16,24,26,27,47]) have been developed using defeasible logic programs without strict rules. Consequently, we consider that the current characterization of IBDPs without strict rules is not a real limitation for our approach’s expressivity and applicability. Notwithstanding this consideration, and to maintain a general approach, it is possible to extend our approach to account for standard DeLP strict rules and facts, where these elements of the program should not come with an associated informant to reflect their strict and indisputable nature accurately; we plan to do this as part of future work.
Regarding the notion of argument strength accounted for in this work, we can highlight the fact that it is solely based on the credibility of the informant agents. This may lead one to think of this notion of argument strength as too narrow, or too specific. Nevertheless, as discussed in Section 4, an informant-based comparison criterion like the single informant comparison criterion does not stray too far from other existing criteria in the literature of structured argumentation, resembling the rules priorities criterion. At any rate, as part of future work, we will explore a generalization our notion of argument strength in order to combine different comparison criteria at different levels. For instance, we could adopt an approach similar to [15], where the rules priorities criterion is used first to resolve attacks into defeats and, in case of undecidedness, the generalized specificity criterion is considered later. More generally, we will explore the possibility of using the operators defined in [45], which allow to combine multiple argument comparison criteria.
As shown in the existing literature (see [12] for an overview), DeLP is among the four major approaches to structured argumentation. Extending a DeLP program into an IBDP, with the addition of informants of defeasible rules and informant rules allowing to argue for and against the consideration of information provided by the different informant agents, we believe we are contributing to expanding DeLP’s applicability domain. Nevertheless, as part of our future work, we plan to study the possibility of further exploiting these ideas and apply them in the context of other major structured argumentation approaches such as ASPIC+ [36] or ABA [46]. In that regard, we consider that such an exploration could make for exciting advances in the area, and will most definitely bring on new challenges requiring a substantial transformation of those frameworks.
In Section 2 we argued that our approach is restricted to deal with credibility orders relating agents that are sources of information about the same topic. As future work, we intend to extend our approach to account for multi-topic credibility orders (i.e., handle multiple credibility orders, one for each topic). In order to be able to deal with these, we also plan to extend the knowledge representation and reasoning capabilities of an IBDP by, among other things, expanding defeasible domain objects to state their topic explicitly. In particular, such an extension would allow us to better model scenarios like the one described in the example of the medical domain given after Definition 11. Then, for instance, we could represent the information provided by agent about medical treatments as belonging to one topic and, on the other hand, the information provided by stating that he is not a physician as belonging to another topic. In that way, detracting rules could also be extended to include the topic under which the information provided by a given agent is challenged; consequently, for the example given in the paper, the detracting rule could target the defeasible domain objects provided by which correspond to medical advice but not the one referring to not being a physician.
Finally, we would like to discuss other exciting prospects for future research. On the one hand, we plan to study how our approach could be extended to consider trust and distrust relations as presented in [29,30]; briefly, the idea would be to connect such relations with backing and detracting arguments. On the other hand, we will also study how an agent’s credibility order should be updated when the warranted information is taken into account. For that purpose, backing and detracting arguments can also play a central role. For instance, suppose that the credibility order initially establishes that an informant is less credible than another informant . Then, if the argumentative machinery yields an undefeated backing argument for and an undefeated detracting argument for , there could be an indication that the credibility order has to be updated so that becomes more credible than . Furthermore, as another direction for future work, we intend to formally analyze the conditions characterizing our valid informant-based comparison criteria in the context of various lifting principles such as those addressed in [9], and also study (and redefine in the context of our approach, when required) different preference handling principles like the ones discussed in [22].
Footnotes
Acknowledgements
We would like to thank the reviewers for their helpful and insightful comments on the previous versions of this paper. This work has been supported by EU H2020 research and innovation programme under the Marie Sklodowska-Curie grant agreement No. 690974 for the project “MIREL: MIning and REasoning with Legal texts”, by CONICET (grant PIP 112-201701-00871) and by Universidad Nacional del Sur (grants PGI 24/N046 and 24/ZN32).
References
1.
R.A.Agis, S.Gottifredi and A.J.García, Acquiring knowledge from expert agents in a structured argumentation setting, Argument & Computation10(2) (2019), 149–189. doi:10.3233/AAC-190447.
2.
R.A.Agis, S.Gottifredi and A.J.García, An approach for distributed discussion and collaborative knowledge sharing: Theoretical and empirical analysis, Expert Systems with Applications116 (2019), 377–395. doi:10.1016/j.eswa.2018.09.016.
3.
L.Amgoud and C.Cayrol, A reasoning model based on the production of acceptable arguments, Ann. Math. Artif. Intell.34(1–3) (2002), 197–215. doi:10.1023/A:1014490210693.
4.
L.Amgoud and R.Demolombe, An argumentation-based approach for reasoning about trust in information sources, Argument & Computation5(2–3) (2014), 191–215. doi:10.1080/19462166.2014.881417.
5.
L.Amgoud and S.Vesic, A new approach for preference-based argumentation frameworks, Ann. Math. Artif. Intell.63(2) (2011), 149–183. doi:10.1007/s10472-011-9271-9.
6.
L.Amgoud and S.Vesic, Rich preference-based argumentation frameworks, Int. J. Approx. Reasoning55(2) (2014), 585–606. doi:10.1016/j.ijar.2013.10.010.
7.
K.Atkinson, P.Baroni, M.Giacomin, A.Hunter, H.Prakken, C.Reed, G.R.Simari, M.Thimm and S.Villata, Towards artificial argumentation, AI Magazine38(3) (2017), 25–36. doi:10.1609/aimag.v38i3.2704.
8.
S.Barberà, W.Bossert and P.K.Pattanaik, Ranking Sets of Objects, Springer, 2004, pp. 893–977.
9.
M.Beirlaen, J.Heyninck, P.Pardo and C.Straßer, Argument strength in formal argumentation, FLAP5(3) (2018), 629–676.
10.
T.J.M.Bench-Capon, Persuasion in practical argument using value-based argumentation frameworks, J. Log. Comput.13(3) (2003), 429–448. doi:10.1093/logcom/13.3.429.
11.
T.J.M.Bench-Capon and P.E.Dunne, Argumentation in artificial intelligence, Artif. Intell.171(10–15) (2007), 619–641. doi:10.1016/j.artint.2007.05.001.
12.
P.Besnard, A.J.García, A.Hunter, S.Modgil, H.Prakken, G.R.Simari and F.Toni, Introduction to structured argumentation, Argument & Computation5(1) (2014), 1–4. doi:10.1080/19462166.2013.869764.
13.
P.Besnard and A.Hunter, A logic-based theory of deductive arguments, Artif. Intell.128(1–2) (2001), 203–235. doi:10.1016/S0004-3702(01)00071-6.
14.
P.Besnard and A.Hunter, Elements of Argumentation, MIT Press, 2008. doi:10.7551/mitpress/9780262026437.001.0001.
15.
C.E.Briguez, M.C.Budán, C.A.D.Deagustini, A.G.Maguitman, M.Capobianco and G.R.Simari, Argument-based mixed recommenders and their application to movie suggestion, Expert Syst. Appl.41(14) (2014), 6467–6482. doi:10.1016/j.eswa.2014.03.046.
16.
M.Capobianco, C.I.Chesñevar and G.R.Simari, Argumentation and the dynamics of warranted beliefs in changing environments, Autonomous Agents and Multi-Agent Systems11(2) (2005), 127–151. doi:10.1007/s10458-005-1354-8.
17.
C.F.Chang, P.Harvey and A.Ghose, Source sensitive argumentation system, in: ICEIS 2006 – Proceedings of the Eighth International Conference on Enterprise Information Systems: Databases and Information Systems Integration, Paphos, Cyprus, May 23–27, 2006, 2006, pp. 39–46.
18.
C.F.Chang, P.Harvey and A.K.Ghose, Combining credibility in a source sensitive argumentation system, in: Advances in Artificial Intelligence, 4th Helenic Conference on AI, SETN 2006, Heraklion, Crete, Greece, May 18–20, 2006, Proceedings, 2006, pp. 478–481. doi:10.1007/11752912_49.
19.
A.Cohen, A.J.García and G.R.Simari, A structured argumentation system with backing and undercutting, Eng. Appl. of AI49 (2016), 149–166. doi:10.1016/j.engappai.2015.10.001.
20.
A.Cohen, S.Gottifredi, A.J.García and G.R.Simari, A survey of different approaches to support in argumentation systems, Knowledge Eng. Review29(5) (2014), 513–550. doi:10.1017/S0269888913000325.
21.
A.Cohen, S.Gottifredi, L.H.Tamargo and A.J.García, Extending Defeasible Logic Programming with informant-based argumentation, in: Argumentation-Based Proofs of Endearment. Essays in Honor of Guillermo R. Simari on the Occasion of His 70th Birthday, C.I.Chesñevar, M.A.Falappa, E.Fermé, A.J.García, A.G.Maguitman, D.C.Martínez, M.V.Martínez, R.O.Rodríguez and G.I.Simari, eds, College Publications, London, 2018, pp. 73–89.
22.
K.Cyras and F.Toni, Properties of ABA+ for non-monotonic reasoning, in: NMR 2016 – Proceedings of the 16th International Workshop on Non-Monotonic Reasoning, Cape Town, South Africa, 22–24 April 2016, http://nmr2016.cs.uct.ac.za/proceedings_nmr2016_online.pdf.
23.
S.K.Dyrkolbotn, T.Pedersen and J.M.Broersen, On elitist lifting and consistency in structured argumentation, FLAP5(3) (2018), 709–746.
24.
A.J.García, N.D.Rotstein, M.Tucat and G.R.Simari, An argumentative reasoning service for deliberative agents, in: Knowledge Science, Engineering and Management, Second International Conference, KSEM 2007, Melbourne, Australia, November 28–30, 2007, Proceedings, 2007, pp. 128–139. doi:10.1007/978-3-540-76719-0_16.
25.
A.J.García and G.R.Simari, Defeasible logic programming: An argumentative approach, Theory and Practice of Logic Programming4(1–2) (2004), 95–138. doi:10.1017/S1471068403001674.
26.
D.R.García, A.J.García and G.R.Simari, Planning and defeasible reasoning, in: 6th International Joint Conference on Autonomous Agents and Multiagent Systems (AAMAS 2007), Honolulu, Hawaii, USA, May 14–18, 2007, 2007, p. 222.
27.
D.R.García, A.J.García and G.R.Simari, Defeasible reasoning and partial order planning, in: Foundations of Information and Knowledge Systems, 5th International Symposium, FoIKS 2008, Pisa, Italy, February 11–15, 2008, Proceedings, 2008, pp. 311–328. doi:10.1007/978-3-540-77684-0_21.
28.
S.Gottifredi, N.D.Rotstein, A.J.García and G.R.Simari, Using argument strength for building dialectical bonsai, Ann. Math. Artif. Intell.69(1) (2013), 103–129. doi:10.1007/s10472-013-9338-x.
29.
W.T.Harwood, J.A.Clark and J.L.Jacob, Networks of trust and distrust: Towards logical reputation systems, in: Logics in Security, 2010.
30.
W.T.Harwood, J.A.Clark and J.L.Jacob, A perspective on trust, security and autonomous systems, in: New Security Paradigms Workshop, 2010.
31.
S.Kaci, L.W.N.van der Torre and S.Villata, Preference in abstract argumentation, in: Computational Models of Argument – Proceedings of COMMA 2018, Warsaw, Poland, 12–14 September 2018, 2018, pp. 405–412.
32.
A.Koster, A.L.C.Bazzan and M.de Souza, Liar liar, pants on fire; or how to use subjective logic and argumentation to evaluate information from untrustworthy sources, Artif. Intell. Rev.48(2) (2017), 219–235. doi:10.1007/s10462-016-9499-1.
33.
J.Maly, M.Truszczynski and S.Woltran, Preference orders on families of sets – when can impossibility results be avoided? in: Proceedings of the Twenty-Seventh International Joint Conference on Artificial Intelligence, IJCAI 2018, July 13–19, 2018, Stockholm, Sweden, 2018, pp. 433–439.
34.
J.Maly, M.Truszczynski and S.Woltran, Preference orders on families of sets – when can impossibility results be avoided?, J. Artif. Intell. Res.66 (2019), 1147–1197. doi:10.1613/jair.1.11879.
35.
P.Matt and F.Toni, A game-theoretic measure of argument strength for abstract argumentation, in: Logics in Artificial Intelligence, 11th European Conference, JELIA 2008, Dresden, Germany, September 28 – October 1, 2008. Proceedings, 2008, pp. 285–297.
36.
S.Modgil and H.Prakken, The ASPIC framework for structured argumentation: A tutorial, Argument & Computation5(1) (2014), 31–62. doi:10.1080/19462166.2013.869766.
37.
D.Nute, Defeasible reasoning: A philosophical analysis in Prolog, in: Aspects of Artificial Intelligence, J.H.Fetzer, ed., Kluwer Academic Pub., 1988, pp. 251–288. doi:10.1007/978-94-009-2699-8_9.
38.
S.Parsons, E.Sklar and P.McBurney, Using argumentation to reason with and about trust, in: Argumentation in Multi-Agent Systems – 8th International Workshop, ArgMAS 2011, Taipei, Taiwan, May 3, 2011, Revised Selected Papers, 2011, pp. 194–212.
39.
H.Prade, A qualitative bipolar argumentative view of trust, in: Scalable Uncertainty Management, First International Conference, SUM 2007, Washington, DC, USA, October 10–12, 2007, Proceedings, 2007, pp. 268–276. doi:10.1007/978-3-540-75410-7_20.
40.
I.Rahwan and G.R.Simari (eds), Argumentation in Artificial Intelligence, Springer, 2009. doi:10.1007/978-0-387-98197-0.
41.
N.Tamani and M.Croitoru, A quantitative preference-based structured argumentation system for decision support, in: IEEE International Conference on Fuzzy Systems, FUZZ-IEEE 2014, Beijing, China, July 6–11, 2014, IEEE, 2014, pp. 1408–1415.
42.
L.H.Tamargo, A.J.García, M.A.Falappa and G.R.Simari, On the revision of informant credibility orders, Artificial Intelligence212 (2014), 36–58. doi:10.1016/j.artint.2014.03.006.
43.
L.H.Tamargo, S.Gottifredi, A.J.García, M.A.Falappa and G.R.Simari, Deliberative DeLP agents with multiple informants, Inteligencia Artificial, Revista Iberoamericana de Inteligencia Artificial15(49) (2012), 13–30.
44.
Y.Tang, K.Cai, P.McBurney, E.Sklar and S.Parsons, Using argumentation to reason about trust and belief, J. Log. Comput.22(5) (2012), 979–1018. doi:10.1093/logcom/exr038.
45.
J.C.Teze, S.Gottifredi, A.J.García and G.R.Simari, An approach to generalizing the handling of preferences in argumentation-based decision-making systems, Knowl.-Based Syst.189 (2020).
46.
F.Toni, A tutorial on assumption-based argumentation, Argument & Computation5(1) (2014), 89–117. doi:10.1080/19462166.2013.869878.
47.
M.Tucat, A.J.García and G.R.Simari, Using Defeasible Logic Programming with contextual queries for developing recommender servers, in: The Uses of Computational Argumentation, Papers from the 2009 AAAI Fall Symposium, Arlington, Virginia, USA, November 5–7, 2009, 2009.
48.
S.Villata, G.Boella, D.M.Gabbay and L.W.N.van der Torre, A socio-cognitive model of trust using argumentation theory, Int. J. Approx. Reason.54(4) (2013), 541–559. doi:10.1016/j.ijar.2012.09.001.