Next Article in Journal
A Color- and Geometric-Feature-Based Approach for Denoising Three-Dimensional Cultural Relic Point Clouds
Previous Article in Journal
Some Constructions and Mathematical Properties of Zero-Correlation-Zone Sonar Sequences
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Distinguishing the Leading Agents in Classification Problems Using the Entropy-Based Metric

1
Department of Industrial Engineering, Ariel University, Ariel 4076414, Israel
2
Department of Industrial Engineering, Tel Aviv University, Tel Aviv 6997801, Israel
*
Author to whom correspondence should be addressed.
Entropy 2024, 26(4), 318; https://doi.org/10.3390/e26040318
Submission received: 4 March 2024 / Revised: 2 April 2024 / Accepted: 2 April 2024 / Published: 5 April 2024
(This article belongs to the Section Multidisciplinary Applications)

Abstract

:
The paper addresses the problem of distinguishing the leading agents in the group. The problem is considered in the framework of classification problems, where the agents in the group select the items with respect to certain properties. The suggested method of distinguishing the leading agents utilizes the connectivity between the agents and the Rokhlin distance between the subgroups of the agents. The method is illustrated by numerical examples. The method can be useful in considering the division of labor in swarm dynamics and in the analysis of the data fusion in the tasks based on the wisdom of the crowd techniques.

1. Introduction

The behavior of the group of autonomous agents includes a variety of physical and cognitive actions, like collective motion and cooperative decision-making. Each of these actions depends on the individual abilities of the agents, on the division of the agents among the teams, and on the division of labor between the agents and the teams.
To simplify control of the teams’ and the group’s activities, some of the agents are often determined as leaders, whose task is to influence the other agents and promote them to fulfill the mission. Such procedures are widely known as elections in distributed systems, which are considered both in social and political processes and in different contexts of computer science and robots’ control [1,2,3].
Given a group of communicating agents, elections are conducted as follows: The agents consider the candidate agents, and after deliberations, they identify and label a certain agent as the leader. Then, the elected leader influences the other agents and coordinates their activities [1]. Certainly, each team in the group can elect its leader, and then these leaders elect the leader of the group, which results in hierarchical control in the system.
The process that is close to the election of the leader is the selection of the leader [4,5,6,7]. In this process, the leader is selected according to the known characteristics of the agents and their correspondence with the characteristics required for fulfilling the mission of the group. Usually, after the section of the leader, the other agents are considered followers. Note that, in contrast to the election of the leader, the leader selection is not necessarily conducted by the agents but can be processed by the central coordinator or controller of the group.
Now, assume that the group leader or the team leader already exists. Then rises an inverse problem—to distinguish the leaders in the distributed system. Namely, given a group of communicating agents, it is required to identify the leaders, which are the agents who mostly influence the other agents in the group.
In the paper, the problem of distinguishing the leading agents is considered in the context of the classification problem [8,9]. In such a classification, it is assumed that the agents have different levels of expertise, and the cooperative classification is obtained by a certain version of plurality voting.
The most popular method of avoiding the influence of non-competent agents uses the weighted opinions of the agents. This method was implemented in the well-known Dawid–Skene algorithm [10,11], which iterates the expectation of the correct choice with respect to the agents’ expertise and maximizes the likelihood of the agents’ expertise with respect to the expected correct choice.
Another approach to selecting the competent agents was implemented in the algorithms [12,13], which are based on the similarities of the agents’ classifications, and in the algorithm [14], where the competent agents are selected using the expectation bias [15].
After selecting the subgroup of competent agents, the resulting classification is obtained using the opinions of these agents and ignoring the opinions of non-competent agents.
However, studies in social psychology [16], which can be traced back to the well-known experiments by Asch [17,18], demonstrate that the opinion of the group member is highly influenced by the opinions of the other members of the group. Consequently, competent agents are not necessarily the most influential or leading agents. Together with that, it is reasonable to assume that the leader must be competent in certain fields and be elected on the basis of this competence (see the 11th rule by Peterson [19]).
In the paper, we suggest a method of distinguishing the leaders in the group. The method considers the connectivity between the agents and creates the subgroups by maximizing the distances between the partitions formed by these subgroups.
Note that the elections in the group are internal processes in which the leaders are identified following certain criteria known to the agents, while distinguishing the leaders is an external operation in which the agents are characterized by their relations with the other agents. Thus, the distinguished leaders can differ from the elected leaders: elections specify who will govern, and distinguishing specifies who must govern.
The rest of the paper is organized as follows: Section 2 includes a formulation of the problem, and Section 3 considers an example that clarifies the problems of classification, distinguishing the competent agents, and distinguishing the leading agents. Section 4 presents a suggested solution based on the connectivity between the agents and the Rokhlin distance between the agents’ subgroups. Section 5 presents the methods of distinguishing the experts, and Section 6 considers two examples that illustrate the relationship between the group of experts and the group of leaders. Section 7 concludes the discourse.

2. Problem Formulation

The main problem considered in the paper is the problem of distinguishing the leaders in the group of agents. As indicated above, this problem differs from the problem of electing the leader and requires consideration of communication between the agents.
We consider the problem in the context of the classification of given items by a group of agents, where some of the agents are experts in the field of classification and others are deletants.
Consideration of these two problems gives rise to the third problem, which is the problem of relations between the group of leaders and the group of experts.

2.1. Distinguishing the Leaders

Let A = a 1 , a 2 , , a l be a group of communicating agents conducting a certain common mission. Communication between the agents is defined by the directed graph G = V , E in which the vertices v V are associated with the agents and the edges e E are defined by the adjacency matrix R = r i j l × l such that r i j = 1 if agent a i communicates with agent a j and r i j = 0 otherwise.
The problem of distinguishing the leaders in group A is formulated as follows: Given the adjacency matrix R , it is required to recognize a subgroup A * A of the agents such that the agents a * A * have maximal influence in the group A .
Consideration of the agent’s influence in the group is based on the following intuition inspired by the flows on graphs (see, e.g., [20]). Assume that each vertex v V in the graph G is a source of unique colored flow and that the color of the vertex is a mix of its own color and the colors of the incoming flows. Then, the influence of the agent is associated with the impact of its vertex on the colors of the other vertices in the graph G .
The vertices v * V * V associated with the leading agents a * A * are called the leading vertices. The leading agent such that its vertex in the graph G has no predecessors is called a dictator, and the leading agent whose vertex is a separating vertex (see, e.g., [21]) is called a monarch.

2.2. Distinguishing the Experts

Let X = x 1 , , x n be a set of n items, which represent certain objects, concepts, or symbols. Classification problem requires to label the items x i X , i = 1 , 2 , , n , by m labels, 1 < m < n , such that each item is labeled by a single label.
Formally, the problem is to distribute the items x i X , i = 1 , 2 , , n , over m sets C 1 , C 2 , , C m , 1 < m < n , called classes, such that each item x i is included only in one class C j and that there is no item which is not included in some classes. Then, the resulting classification is the partition γ = C 1 , C 2 , , C m of the set X , where C j X , j = 1 , , m , C j C j = for j j , and j = 1 m C j = X .
If classification is conducted by a single agent, then the resulting classification γ depends on the competence of this agent. The quality of classification is defined by the difference between the classification γ and the correct classification γ ˇ . Certainly, the correct classification γ ˇ is not available to the agent and is used for testing the classification methods.
Now assume that the classification is conducted by the indicated above group A = a 1 , a 2 , , a l of agents where each agent a k A , k = 1 , 2 , , l , provides classification represented by the partition γ k = C k , 1 , C k , 2 , , C k , m . By the general assumption of “the wisdom of the crowd” techniques [8,9], some combination of the agents’ classification will provide classification γ , which is as close as possible to the correct classification γ ˇ .
Then, the problem is to aggregate the agents’ partitions γ k into a single partition γ such that it, as best as possible, represents the correct partition γ ˇ .
The simplest method of creating the partition γ is plurality voting. By this method, the item x is included in the class C , which was chosen by most agents a A (the ties are broken randomly). Despite its popularity, this method strongly depends on the competence of the agents, such that non-competent agents can influence the resulting classification.
To avoid such influence, in more sophisticated methods [10,11,12,13,14], the problem is divided into two stages. First, using the agents’ classifications γ k , k = 1 , 2 , , l , the agents competent in certain classes are distinguished, and second, the resulting classification γ is obtained by aggregation of the classes provided by the competent agents.

2.3. Relationship between the Leaders and the Experts

Finally, assume that the group A = a 1 , a 2 , , a l of l communicating agents classifies the set = x 1 , , x n of n items to m classes C 1 , C 2 , , C m , 1 < m < n .
In addition, assume that the group A of the agents includes a non-empty subgroup A * A of leaders and non-empty subgroup A A of experts.
Then, the problem is to check whether there exists a relationship between the set of leaders A * and the set A of experts, and if it exists, what this relationship is.
We assume that in real-world situations, such a relationship is possible, and the leaders are assumed to be experts in at least one class. The problem is to confirm or withdraw this hypothesis.

3. Illustrative Example

Assume that the group A includes l = 9 communicating agents and that communication between the agents is defined by the directed graph G = V , E , where the vertices v V are associated with the agents and the edges e E specify the communication between the corresponding agents. The graph G is shown in Figure 1.
The sets of input and output vertices in this graph are presented in Table 1.
As indicated above, in consideration of the agents’ impacts, each vertex of the graph G is considered a source of the unique colored flow. The color absorbed by the vertex is a mix of its own color and the colors of the incoming flows. The impact of the agent is considered to be the impact of its vertex on the color of the other vertices. Then, intuitively, the vertex with a maximal number of predecessors and successors is associated with the agents with maximal influence.
Following the cardinality of the sets of input and output vertices, it can be supposed that the set of vertices associated with the candidates to the leading agents is V c = v 1 , v 4 , v 5 , v 8 such that a 1 is a dictator and a 4 is a monarch. In addition, because of the maximal number of predecessors, intuitively, vertex v 6 can also be considered a candidate for the leading vertex.
Assume that the agents a from the group A distribute n = 12 items x from the set X over m = 4 classes C . The results of the classification are shown in Table 2.
In the table, the first agent a 1 included the item x 1 into the class C 2 , the item x 2 into the class C 1 , the item x 3 into the class C 3 and so on, such that the partition γ 1 created by the first agent is
γ 1 = x 2 , x 6 , x 8 , x 1 , x 3 , x 9 , x 10 , x 12 , x 4 , x 5 , x 7 , x 11 .
The second agent a 2 included the item x 1 into the class C 2 , the item x 2 into the class C 3 , the item x 3 into the class C 2 and so on, such that the second agent is
γ 2 = , x 1 , x 3 , x 10 , x 2 , x 6 , x 7 , x 8 , x 9 , x 11 , x 12 , x 4 , x 5 ,
and so on. Correct classification is represented by the following partition:
γ ˇ = x 2 , x 6 , x 8 , x 1 , x 3 , x 9 , x 10 , x 12 , x 4 , x 5 , x 7 , x 11 ,
and the partition obtained by the plurality voting is
γ P l = x 2 , x 3 , x 6 , x 10 , x 1 , x 8 , x 11 , x 12 , x 4 , x 7 , x 9 ,
It is seen that the classifications γ k provided by the agents a k , k = 1,2 , , l , are rather far from the correct classification γ ˇ as well as the aggregated classification γ P l obtained by plurality voting.
However, some agents created certain classes that are equivalent to the classes in the correct classification, namely,
  • agent a 1 —class C 1 = x 2 , x 6 , x 8 ,
  • agent a 2 —class C 2 = x 1 , x 3 , x 10 ,
  • agent a 3 —classes C 1 = x 2 , x 6 , x 8 and C 2 = x 1 , x 3 , x 10 ,
  • agent a 4 —class C 3 = x 4 , x 5 , x 11 ,
  • agent a 5 —class C 4 = x 7 , x 9 , x 12 ,
  • agent a 6 —classes C 3 = x 4 , x 5 , x 12 and C 4 = x 7 , x 9 , x 12 .
The other agents a 7 , a 8 and a 9 provided completely erroneous classifications, where despite the correct classification of some items, all obtained classes differ from the classes appearing in the correct classification. Then, aggregating the appropriate classes from the classifications created by the agents a 1 , , a 6 and avoiding classifications created by the agents a 7 , , a 9 , provides correct classification γ = γ ˇ .
In the considered case of random classifications and relations between the agents, a possible set of leaders is A * = a 1 , a 4 , a 5 , a 8 and possible set of experts is A = a 1 , a 2 , a 3 , a 4 , a 5 , a 6 , which means that there is no clear relationship between these sets. However, since in real-world situations, the leader must be competent in at least one field of knowledge, the absence of such a relationship is not obvious, and its consideration is reasonable.

4. Distinguishing the Group of Leaders Using the Entropy-Based Metric

Let A = a 1 , a 2 , , a l be a group of agents and G = V , E be the directed graph representing communication between the agents such that the vertices v V are associated with the agents and the edges e = v i , v j E define communication between the agents a i , a j A , i , j = 1,2 , , l .
Denote by u v , ξ V a predecessor of the vertex v V such that the shortest path from u v , ξ to v in the graph G is of the length ξ , and by u v , ξ V a successor of the vertex v V such that the shortest path from v to u v , ξ in the graph G is of the length ξ . In particular, u v , 1 = u v is a direct predecessor of v and u v , 1 = u v is a direct successor of v . For completeness, we also say that u v , 0 = u v , 0 = v .
It is clear that each v V and all its predecessors are the predecessors of each successor of v , and each v V and all its successors are the successors of each predecessor of v .
For a vertex v V , denote by U v , ξ , the set of its predecessors u v , ξ and by U v , ξ the set of all its successors u v , ξ . The set of direct predecessors is denoted by U v , 1 = U v and the set of direct successors is denoted by U v , 1 = U v .
Given a graph G , let U v be the set of direct predecessors of the vertex v , U u v be the sets of direct predecessors of the vertices u v U v , U u u v be the sets of direct predecessors of the vertices u u v U u v and so on, up to, but not including, the set that already appears among the sets of direct predecessors obtained at the previous steps.
Similarly, let U v be the set of direct successors of the vertex v , U u v be the sets of direct successors of the vertices u v U v , U u u v be the sets of direct successors of the vertices u u v U u v and so on, up to, but not including, the set that already appears among the sets of direct successors obtained at the previous steps.
Finally, for the vertex v , let us form the predecessors’ tree T v and the successors’ tree T v . In the tree T v , the root is associated with the set U v and the leaves at their levels are associated with the sets U u v , U u u v and so on, respectively. Similarly, in the T v , the root is associated with the set U v and the leaves at their levels are associated with the sets U u v , U u u v and so on.
For illustration, the predecessors’ tree T v 8 and the successors’ tree T v 8 of the vertex v 8 in the graph G are shown in Figure 2.
The sets associated with the leaves of the trees T v and T v form, respectively, the predecessor cover τ v and the successor cover τ v of certain subsets of the set V of vertices.
For example, the predecessor and successor covers of the vertex v 8 are τ v 8 = v 1 , v 2 , v 4 , v 7 , v 9 and τ v 8 = v 5 , v 6 , v 6 .
The predecessor cover τ V of the subset V V of vertices is a set
τ V = v V τ v
of the predecessor covers τ v of the vertices v V , and the successor cover τ V of the subset V V of vertices is a set
τ V = v V τ v
of the successor covers τ v of the vertices v V .
For example, for the indicated above set V c = v 1 , v 4 , v 5 , v 8 of vertices, the predecessor and successor covers are τ V c = v 1 , v 2 , v 3 , v 4 , v 8 , v 4 , v 7 , v 9 and τ V c = v 2 , v 4 , v 4 , v 5 , v 6 , v 5 , v 7 , v 8 .
Then, we say that the subset V * V is a set of leading vertices if the distance d τ V * , τ V * between its predecessor cover τ V * and successor cover τ V * is maximal over all possible subsets of the set V .
The agents a * associated with the leading vertices v * V * are called the leaders and the group A * A of leading agents is called the leading group.
Distance d τ V , τ V between the covers τ V and τ V can be calculated using different methods. Here, we suggest the distance measure, which is based on the Rokhlin metric [22]. Since in the considered task, the main stress is on the communication between the agents and on the classification of the data items, the use of such an entropy-based metric is reasonable. Together with that, since the suggested method deals with formal sets of vertices in the graph, the other measures, e.g., the Ornstein distance [23,24], can be applied. For a comparison between the Rokhlin distance and the Ornstein distance, see [25]. Note that both Rokhlin and Ornstein metrics require the defined probability measure on the sets; if such a probability does not exist, then the normalized Hamming distance [12] can be used.
The Rokhlin metric is defined as follows. Let Ω , Q , p be a probability space with a probability measure p on Ω , and let α = Q | Q Q , Q i Q j = , i j , Q α Q = Ω , be a partition of Ω . The entropy of the partition α is the value
H α = Q α p Q log p Q ,
where l o g is base 2 , and it is assumed that p log p = 0 log 0 = 0 . In addition, let β = R | R Q , R i R j = , i j , R β R = Ω , be another partition of Ω . Then the conditional entropy of partition α given partition β is the value
H α | β = R β Q α p Q , R log p Q | R ,
where p Q , R = p Q R and p Q | R = p Q R p R .
The Rokhlin metric [22], which defines the distance between partitions α and β is a sum
d R α , β = H α | β + H β | α ,
For basic properties of this metric and its role in dynamical systems theory, see [26,27]; for additional properties and comparison with the Ornstein metric, see [25].
To apply this metric for measuring the distance between the covers τ V and τ V , note again that each of these sets does not necessarily cover the set of vertices V , but the subsets Q τ V Q V and R τ V R V of this set. Then, let us add to each of these sets the set which completes it to the cover of V .
Namely, the set τ V is completed with the set Q = V \ Q τ V Q and the set τ V is completed with the set R = V \ R τ V R . As a result, the sets
τ V = τ V Q   and   τ V = τ V R
cover the set V of vertices.
Then, it is required to define the probability measure p : V 0,1 on the set of vertices. Since there is no additional information about the agents, we assume that
p v = 1 # V
for each vertex v V , and
p Q = v Q p v = # Q # V
for each subset Q V of vertices.
For the conditional entropy, we have
H τ V | τ V = R τ V Q τ V p Q , R log p Q | R = R τ V Q τ V p Q , R log p Q | R Q τ V p Q , R log p Q | R = R τ V Q τ V p Q , R log p Q | R + p Q , R log p Q | R Q τ V p Q , R log p Q | R p Q , R log p Q | R = R τ V Q τ V p Q , R log p Q | R R τ V p Q , R log p Q | R Q τ V p Q , R log p Q | R p Q , R log p Q | R .
In this formula, the first term is equivalent to the conditional entropy of the sets τ V and τ V , the second term represents the influence of the sets Q and R to the elements of the sets τ V and τ V , and the last term defines the conditional entropy of Q with respect to R .
This definition is a direct extension of the definition of conditional entropy of the partitions. In fact, if the sets τ V and τ V are covers of V , then Q = and R = . Then,
H τ V | τ V = H τ V | τ V = R τ V Q τ V p Q , R log p Q | R ,
and if τ V and τ V are partitions of V , then it is equivalent to the definition of the conditional entropy.
Note that since τ V and τ V are covers of the set V , the conditional entropy H τ V | τ V does not necessarily meet all the properties of the conditional entropy defined for the partitions. However, here, we will not consider specific properties of the conditional entropy of the covers but will use it directly to define the distance between the predecessor cover τ V and the successor cover τ V .
The distance d τ V , τ V between the predecessor cover τ V and the successor cover τ V is defined by the Rokhlin distance between the covers τ V and τ V of the set V of vertices as
d τ V , τ V = H τ V | τ V + H τ V | τ V ,
For example, the distance between the predecessor cover τ V c = v 1 , v 2 , v 3 , v 4 , v 8 , v 4 , v 7 , v 9 and the successor cover τ V c = v 2 , v 4 , v 4 , v 5 , v 6 , v 5 , v 7 , v 8 is calculated as follows:
The completed sets for the covers τ V c and τ V c are Q = v 1 , v 2 , , v 9 \ v 1 , v 2 , v 3 , v 4 , v 7 , v 8 , v 9 = v 5 , v 6 and R = v 1 , v 2 , , v 9 \ v 2 , v 4 , v 5 , v 6 , v 7 , v 8 = v 1 , v 3 , v 9 . Then, the completed covers of the set V of vertices are
τ V c = v 1 , v 2 , v 3 , v 4 , v 8 ,   v 4 , v 7 , v 9 , v 5 , v 6
and
τ V c = v 2 , v 4 , v 4 , v 5 , v 6 , v 5 , v 7 ,   v 8 , v 1 , v 3 , v 9 .
The probability of each vertex v V is p v = 1 9 . Then, conditional entropies H τ V c | τ V c and H τ V c | τ V c are (the zero terms are omitted).
H τ V c | τ V c = p v 2 log p v 2 p v 2 , v 4 p v 4 log p v 4 p v 2 , v 4 p v 4 log p v 4 p v 2 , v 4 p v 7 log p v 7 p v 5 , v 7 p v 5 log p v 5 p v 5 , v 7 p v 1 log p v 1 p v 1 , v 3 , v 9 p v 3 log p v 3 p v 1 , v 3 , v 9 p v 9 log p v 9 p v 1 , v 3 , v 9 = 1 9 log 1 2 1 9 log 1 2 1 9 log 1 2 1 9 log 1 2 1 9 log 1 2 1 9 log 1 3 1 9 log 1 3 1 9 log 1 3 = 1.08 ,
H τ V c | τ V c = p v 2 log p v 2 p v 1 , v 2 p v 1 log p v 1 p v 1 , v 2 p v 4 log p v 4 p v 3 , v 4 , v 8 p v 4 log p v 4 p v 3 , v 4 , v 8 p v 8 log p v 8 p v 3 , v 4 , v 8 p v 3 log p v 3 p v 3 , v 4 , v 8 p v 7 log p v 7 p v 7 , v 9 p v 9 log p v 9 p v 7 , v 9 p v 5 log p v 5 p v 5 , v 6 = 1 9 log 1 2 1 9 log 1 2 1 9 log 1 3 1 9 log 1 3 1 9 log 1 3 1 9 log 1 3 1 9 log 1 2 1 9 log 1 2 1 9 log 1 2 = 1.26 ,
The distance between the predecessor and successor covers of the set V c = v 1 , v 4 , v 5 , v 8 of vertices is
d τ V c , τ V c = 1.08 + 1.26 = 2.34 .
For comparison, the distance between the indicated above predecessor cover τ v 8 = v 1 , v 2 , v 4 , v 7 , v 9 and the successor cover τ v 8 = v 5 , v 6 , v 6 of the vertex v 8 is
d τ v 8 , τ v 8 = 1.96 .
Then, the group A c = a 1 , a 4 , a 5 , a 8 of the agents associated with the vertices of the group V c = v 1 , v 4 , v 5 , v 8 is preferable as a group of leaders than a group A c = a 8 , which includes only one agent a 8 associated with the vertex v 8 .
Hereby, we defined the group A * A of leading agents and suggested the criterion for its recognition among the other agents in the group A . The same procedure can be continued over the group A * and then recurrently over the obtained groups up to distinguishing a unique leading agent.
The algorithmic solution to the problem of distinguishing the leading agents is a complex task which requires an exhaustive search among all possible subsets of the agents from the group A or, that is, the same, among all possible subsets of vertices from the set V . Together with that, certain heuristics omitting the vertices with a relatively small number of predecessors and successors can strongly decrease the number of candidate solutions.

5. Distinguishing the Group of Experts

Assume that the group of agents A = a 1 , a 2 , , a l considers the set X = x 1 , , x n of n items and each agent a k A , k = 1,2 , , l , provides partition γ k = C k , 1 , C k , 2 , , C k , m of the set X to m classes. The resulting classification is an aggregated partition γ = C 1 , C 2 , , C m created from the agents’ partitions γ k , k = 1,2 , , l , and to obtain the correct partition γ , it is required to recognize partitions provided by the competent agents and avoid partitions provided by non-competent agents.
Distinguishing the experts is based on the assumption that the agents with the same competence in the same fields provide similar classifications of the items related to their field of expertise and can provide different classifications of the items that are outside of the scope of their competence [12]. In other words, we follow the well-known phrase by Father Dominique Bouhours ([28] (p. 125), punctuation and grammar preserved):
“Great Minds often think alike on the same Occasions, and we are not always to suppose, that such Thoughts are borrow’d from one another when exprest by Persons of the same heroick Sentiments.”
Following this assumption, agent a k A is considered a weak expert in a certain class C j if the agent’s partition γ k includes C j and there exist the other agents a k , a k , A , such that their partitions γ k , γ k ,… include C j . If the partition γ k class C j is at the same position as in the partitions γ k , γ k ,…, then the agent a k is called strong expert or expert, for briefness.
The number of agents with equivalent classes C j required for specifying the agent as an expert varies and depends on the number l of agents in the group. Following general statistical assumptions, we say that the number of such agents is at least 10 % of l and, for small groups, is not less than 2 .
As indicated above, there exist several algorithms of classification that implement the difference between the opinions of competent and non-competent agents [10,11,12,13]. In particular, the Distance-Based Collaborative Classification (DBCC) algorithm [12] directly considers the normalized Hamming distance d H γ k , γ k | j between the partitions γ k = C k , 1 , C k , 2 , , C k , m and γ k = C k , 1 , C k , 2 , , C k , m with respect to each class C j
d n H γ k , γ k | j = # C k , j C k , j / # C k , j + # C k , j ,
where C k , j C k , j = C k , j C k , j \ C k , j C k , j is a symmetric difference between the classes C k , j and C k , j .
If on the set X of items, a probability measure p : X 0,1 is defined, then using this measure, the distance between the partitions γ k and γ k of X with respect to the class C j can be defined as
d p H γ k , γ k | j = p C k , j C k , j ,
or in the form of the Rokhlin metric as
d p R γ k , γ k | j = p C k , j \ C k , j log p C k , j \ C k , j p C k , j \ C k , j log p C k , j \ C k , j .
Below, we assume that the experts are already distinguished by these or other methods and consider the relationship between the group of leaders and the group of experts.

6. Relationship between the Group A * of Leaders and the Group A of Experts

Let us return to the above example and assume that the distinguished group of experts is A = a 1 , a 2 , a 3 , a 4 , a 5 , a 6 . Our aim is to check whether this group of experts is also a group of leaders.
Denote by V V , the set of vertices in the graph G associated with the agents from the group A of experts. Then, following the presented above procedure of distinguishing the group of leading agents, the distance between the predecessor partition τ V and the successor partition τ V is
d τ V , τ V = H τ V | τ V + H τ V | τ V = 0.89 + 0.70 = 1.59 .
It is seen that the predecessor partition τ V and successor partition τ V of the set V are closer than the predecessor partition τ V c and successor partition τ V c of the previously distinguished set V c = v 1 , v 4 , v 5 , v 8 of vertices associated with the agents from the set A c of candidate leaders (distance d τ V c , τ V c = 2.34 ). Moreover, partitions τ V and τ V are closer than the partitions τ v 8 and τ v 8 of the vertex v 8 (distance d τ v 8 , τ v 8 = 1.96 ).
Now, let us consider the classifications provided by the candidate group A c = a 1 , a 4 , a 5 , a 8 of leaders. We have
γ 1 = x 2 , x 6 , x 8 , x 1 , x 3 , x 9 , x 10 , x 12 , x 4 , x 5 , x 7 , x 11 , γ 4 = x 3 , x 6 , x 7 , x 10 , x 12 , x 8 , x 9 , x 4 , x 5 , x 11 , x 1 , x 2 , γ 5 = x 3 , x 4 , x 10 , x 5 , x 6 , x 8 , x 11 , x 1 , x 2 , x 7 , x 9 , x 12 , γ 8 = x 3 , x 10 , x 2 , x 8 , x 1 , x 4 , x 5 , x 6 , , x 11 , x 12 , x 7 , x 9 .
Despite the competence of agent a 1 in class C 1 of agent a 4 in class C 3 and of agent a 5 in class C 5 , the resulting plurality voting partition is (item x 2 is labeled randomly)
γ P l = x 3 , x 6 , x 10 , x 2 , x 8 , x 1 , x 5 , x 11 , x 12 , x 4 , x 7 , x 9 ,
which is far from the correct partition γ ˇ .
Thus, in the considered example with random relations between the agents a k and a k and randomly chosen classifications γ k , k , k = 1,2 , , l , the group of experts strongly differs from the group of leading agents.
However, as indicated above, it is reasonable to assume that the agent elected to be a leader or a member of the group of leaders is competent in certain fields [19].
Following this assumption, let us consider the other example and form a group of leaders starting with the agents’ expertise [13]. Assume that the group A of l = 8 agents classifies n = 9 items x from the set X over m = 5 classes C . The agents’ classifications are shown in Table 3.
By the Wisdom in the Crowd (WICRO) algorithm [13], the agents are divided into clusters based on the number of agreements about the classes for each item x . The clusters obtained by the agents are summarized in Table 4.
Following the table, a 1 and a 2 are the agents that agree that the item x 1 should be in the class C 1 and the item x 3 should be in the class C 3 ; a 1 , a 2 , and a 7 are the agents that agree that the item x 2 should be in the class C 2 and so on.
In addition, it is seen that the agents a 1 and a 2 appear both in the cluster a 1 , a 2 and in the cluster a 1 , a 2 , a 7 together with the agent a 7 . Thus, we assume that the agents a 1 and a 2 are predecessors and successors of each other and both are predecessors of the agent a 7 . The same holds for the agents a 5 and a 6 and the agent a 4 .
Also, the agents a 3 and a 4 appear in two clusters a 1 , a 3 , a 4 and a 2 , a 3 , a 4 , while each of the agents a 1 and a 2 appear in only one of these two clusters. So, we assume that the agents a 3 and a 4 are predecessors and successors of each other and both are predecessors of the agents a 1 and a 2 .
Finally, the agent a 8 does not appear in any cluster, so we assume that this agent is a successor of all other agents.
Associating the agents with the vertices and the relations between the agents with the edges, one obtains the directed graph G = V , E shown in Figure 3.
Similarly to Table 1, the sets of input and output vertices in this graph are presented in Table 5.
Following this graph, the clear leader is the agent a 4 associated with the vertex v 4 and additional leaders are the agents a 1 and a 2 associated with the vertices v 1 and v 2 . Together with that, according to the structure of the graph, none of the agents is a dictator or monarch.
The further application of the majority voting in the group A * = a 1 , a 2 , a 4 of leading agents results in the classification that is correct for items x 1 , x 2 , x 3 and x 5 , is incorrect for items x 6 , x 7 and x 8 , and with probability 1 3 can be correct for each of the items x 4 and x 9 .

7. Conclusions

The paper considered the problem of distinguishing the leaders in the group of autonomous agents.
In the paper, we suggested a definition of the leading agents, which are the agents that maximally divide the group. For calculating the distances between the subgroups of the agents, we use the entropy-based Rokhlin metric, which was extended for measuring the distances between the covers of the sets.
In the framework of classification problems, the paper considers the relationship between the competent agents and the leading agents and presents an example of distinguishing the leaders based on their expertise in certain fields of knowledge.
The suggested method can be used in programming the division of labor in the swarm activity dynamics and in the analysis of the data fusion in the records obtained by the wisdom of the crowd techniques.
Further research will include verification of the method on a wider range of data and consideration of the relations between the properties of the graphs, the groups of the distinguished leading agents, and the levels of their expertise.

Author Contributions

Conceptualization, E.K. and I.B.-G.; methodology, E.K. and I.B.-G.; software, E.K.; formal analysis, E.K.; investigation, E.K.; writing—original draft preparation, E.K.; writing—review and editing, I.B.-G.; supervision, I.B.-G.; project administration, I.B.-G. All authors have read and agreed to the published version of the manuscript.

Funding

This researchwas partially supported by the Koret Foundation’s “Digital Living 2030” fund.

Institutional Review Board Statement

Not applicable.

Data Availability Statement

No new data were created or analyzed in this study. Data sharing is not applicable to this article.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Garcia-Molina, H. Elections in a distributed computing system. IEEE Trans. Comput. 1982, 31, 48–59. [Google Scholar] [CrossRef]
  2. Kim, T.W.; Kim, E.H.; Kim, J.K.; Kim, T.Y. A leader election algorithm in a distributed computing system. In Proceedings of the Fifth IEEE Computer Society Workshop on Future Trends of Distributed Computing Systems, Cheju, Republic of Korea, 28–30 August 1995; pp. 481–485. [Google Scholar]
  3. Guo, M.; Tumova, J.; Dimarogonas, D.V. Cooperative decentralized multi-agent control under local tasks and connectivity constraints. In Proceedings of the 53rd IEEE Conference on Decision and Control, Los Angeles, CA, USA, 15–17 December 2014; pp. 75–80. [Google Scholar]
  4. Clark, A.; Bushnell, L.; Poovendran, R. On leader selection for performance and controllability in multi-agent systems. In Proceedings of the 2012 IEEE 51st IEEE Conference on Decision and Control (CDC), Maui, HI, USA, 10–13 December 2012; pp. 86–93. [Google Scholar]
  5. Walker, P.; Amraii, S.A.; Lewis, M.; Chakraborty, N.; Sycara, K. Control of swarms with multiple leader agents. In Proceedings of the 2014 IEEE International Conference on Systems, Man, and Cybernetics (SMC), San Diego, CA, USA, 5–8 October 2014. [Google Scholar]
  6. Fitch, K. Optimal Leader Selection in Multi-Agent Networks: Joint Centrality, Robustness and Controllability. Ph.D. Thesis, Princeton University, Princeton, NJ, USA, 2016. [Google Scholar]
  7. Lewkowicz, M.A.; Agarwal, R.; Chakraborty, N. Distributed algorithm for selecting leaders for supervisory robotic swarm control. In Proceedings of the 2019 International Symposium on Multi-Robot and Multi-Agent Systems (MRS), New Brunswick, NJ, USA, 22–23 August 2019; pp. 112–118. [Google Scholar]
  8. Whitehill, J.; Ruvolo, P.; Wu, T.; Bergsma, J.; Movellan, J. Whose vote should count more: Optimal integration of labels from labelers of unknown expertise. Adv. Neural Inf. Process. Syst. 2009, 22, 2035–2043. [Google Scholar]
  9. Hamada, D.; Nakayama, M.; Saiki, J. Wisdom of crowds and collective decision making in a survival situation with complex information integration. Cogn. Res. 2020, 5, 48. [Google Scholar] [CrossRef] [PubMed]
  10. Dawid, A.P.; Skene, A.M. Maximum likelihood estimation of observer error-rates using the EM algorithm. J. Roy. Stat. Soc. Ser. C 1979, 28, 20–28. [Google Scholar] [CrossRef]
  11. Sinha, V.B.; Rao, S.; Balasubramanian, V.N. Fast Dawid-Skene: A fast vote aggregation scheme for sentiment classification. In Proceedings of the 7th KDD Workshop on Issues of Sentiment Discovery and Opinion Mining, London, UK, 20 August 2018. [Google Scholar]
  12. Ghanaiem, A.; Kagan, E.; Kumar, P.; Raviv, T.; Glynn, P.; Ben-Gal, I. Unsupervised classification under uncertainty: The distance-based algorithm. Mathematics 2023, 11, 4784. [Google Scholar] [CrossRef]
  13. Ratner, N.; Kagan, E.; Kumar, P.; Ben-Gal, I. Unsupervised classification for uncertain varying responses: The wisdom-in-the-crowd (WICRO) algorithm. Knowl.-Based Syst. 2023, 272, 110551. [Google Scholar] [CrossRef]
  14. Kagan, E.; Novoselsky, A. Unsupervised classification by iterative voting. Crowd Sci. 2023, 7, 63–67. [Google Scholar] [CrossRef]
  15. Galton, F. Vox populi. Nature 1907, 75, 450–451. [Google Scholar] [CrossRef]
  16. Aronson, E. The Social Animal, 10th ed.; Aronson, J., Ed.; Worth Publishers: New York, NY, USA, 2008. [Google Scholar]
  17. Asch, S. Effects of group pressure upon the modification and distortion of judgments. In Groups, Leadership, and Men; Guetzdow, H., Ed.; Carnegie Press: Lancaster, PA, USA, 1951; pp. 117–190. [Google Scholar]
  18. Asch, S. Opinions and social pressure. Sci. Am. 1955, 193, 31–35. [Google Scholar] [CrossRef]
  19. Peterson, J. 12 Rules of Life. An Antidote to Chaos; Random House: Toronto, ON, Canada, 2018. [Google Scholar]
  20. Cormen, T.; Leiserson, C.; Rivest, R. Introduction to Algorithms; The MIT Press: Cambridge, MA, USA, 1990. [Google Scholar]
  21. Ore, O. Theory of Graphs; American Mathematical Society: Providence, RI, USA, 1962. [Google Scholar]
  22. Rokhlin, V.A. Lectures on the entropy theory of measure-preserving transformations. Russ. Math. Surv. 1967, 22, 1–52. [Google Scholar] [CrossRef]
  23. Ornstein, D.S. Measure preserving transformations and random processes. Am. Math. Mon. 1971, 78, 833–840. [Google Scholar] [CrossRef]
  24. Ornstein, D.S. Ergodic Theory, Randomness, and Dynamical Systems; Yale University Press: New Haven, CT, USA; London, UK, 1974. [Google Scholar]
  25. Kagan, E.; Ben-Gal, I. Probabilistic Search for Tracking Targets; Wiley & Sons: Chichester, UK, 2013. [Google Scholar]
  26. Sinai, Y.G. Introduction to Ergodic Theory; Princeton University Press: Princeton, NJ, USA, 1977. [Google Scholar]
  27. Sinai, Y.G. Topics in Ergodic Theory; Princeton University Press: Princeton, NJ, USA, 1994. [Google Scholar]
  28. Bouhours, D. The Arts of Logick and Rhetorick, Illustrated by Examples Taken out of the Best Authors, Antient and Modern, in All the Polite Languages, Interpreted and Eplain’d by That Learned and Judicious Critick; Clark, J., Hett, R., Pemberton, J., Ford, R., Gray, J., Eds.; ECCO Print: London, UK, 1728. [Google Scholar]
Figure 1. Example of the graph defining communication between the agents.
Figure 1. Example of the graph defining communication between the agents.
Entropy 26 00318 g001
Figure 2. Example of the predecessors’ and successors’ trees: (a) the predecessors’ tree T v 8 and (b) the successors’ tree T v 8 of the vertex v 8 in the graph shown in Figure 1.
Figure 2. Example of the predecessors’ and successors’ trees: (a) the predecessors’ tree T v 8 and (b) the successors’ tree T v 8 of the vertex v 8 in the graph shown in Figure 1.
Entropy 26 00318 g002
Figure 3. Graph of communication between the agents with respect to the agents’ clustering.
Figure 3. Graph of communication between the agents with respect to the agents’ clustering.
Entropy 26 00318 g003
Table 1. The sets of input and output vertices in the graph G .
Table 1. The sets of input and output vertices in the graph G .
Vertex v 1 v 2 v 3 v 4 v 5 v 6 v 7 v 8 v 9
Input set v 1 v 1 , v 2 v 4 , v 4 , v 8 v 5 , v 8 v 4 v 7 , v 9
Output set v 1 , v 4 v 4 v 5 v 5 , v 7 v 6 v 8 v 5 , v 6 v 8
Table 2. Example of n = 12 items distributed by l = 9 agents over m = 4 classes. Partitions γ k , k = 1,2 , , 9 represent the agents’ classifications; partition γ ˇ represents correct classification; and partition γ P l represents the result of plurality voting.
Table 2. Example of n = 12 items distributed by l = 9 agents over m = 4 classes. Partitions γ k , k = 1,2 , , 9 represent the agents’ classifications; partition γ ˇ represents correct classification; and partition γ P l represents the result of plurality voting.
γ 1 γ 2 γ 3 γ 4 γ 5 γ 6 γ 7 γ 8 γ 9 γ ˇ γ P l
x 1 222431 1 3 4 22
x 2 131431 1 2 3 11
x 3 322111 2 1 3 21
x 4 444313 4 3 1 34
x 5 444323 4 3 2 34
x 6 131122 2 3 1 11
x 7 433144 4 4 3 44
x 8 131222 1 2 3 12
x 9 334244 2 4 4 44
x 10 322111 2 1 1 21
x 11 434323 4 3 3 33
x 12 333144 1 3 4 43
Table 3. Example of n = 9 items distributed by l = 8 agents over m = 5 classes. Partitions γ k , k = 1,2 , , 8 represent the agents’ classifications and partition γ ˇ represents correct classification.
Table 3. Example of n = 9 items distributed by l = 8 agents over m = 5 classes. Partitions γ k , k = 1,2 , , 8 represent the agents’ classifications and partition γ ˇ represents correct classification.
γ 1 γ 2 γ 3 γ 4 γ 5 γ 6 γ 7 γ 8 γ ˇ
x 1 1 1 2 4 3 5 4 3 1
x 2 2 2 3 4 3 4 2 5 2
x 3 3 3 2 1 4 2 5 2 3
x 4 5 1 4 4 3 3 2 1 4
x 5 5 2 5 5 1 4 4 3 5
x 6 5 2 2 2 3 1 1 4 1
x 7 4 3 1 5 2 2 4 4 2
x 8 2 1 5 2 3 2 1 2 3
x 9 5 1 2 4 4 4 3 2 4
Table 4. Clusters of the agents with respect to the number of agreements about the classes of the items.
Table 4. Clusters of the agents with respect to the number of agreements about the classes of the items.
Agent’s ClusterChosen Class Number
x 1 a 1 , a 2 1
x 2 a 1 , a 2 , a 7 2
x 3 a 1 , a 2 3
x 4 a 5 , a 6 3
x 5 a 1 , a 3 , a 4 5
x 6 a 2 , a 3 , a 4 2
x 7 a 5 , a 6 2
x 8 a 5 , a 6 3
x 9 a 4 , a 5 , a 6 4
Table 5. The sets of input and output vertices in the graph G shown in Figure 3.
Table 5. The sets of input and output vertices in the graph G shown in Figure 3.
Vertex v 1 v 2 v 3 v 4 v 5 v 6 v 7 v 8
Input set v 2 , v 3 v 1 , v 4 v 4 v 3 , v 5 , v 6 v 6 v 5 v 1 , v 2
Output set v 2 , v 7 , v 8 v 1 , v 7 , v 8 v 1 , v 4 , v 8 v 2 , v 3 , v 8 v 4 , v 6 , v 8 v 4 , v 5 , v 8 v 8 V \ v 8
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Kagan, E.; Ben-Gal, I. Distinguishing the Leading Agents in Classification Problems Using the Entropy-Based Metric. Entropy 2024, 26, 318. https://doi.org/10.3390/e26040318

AMA Style

Kagan E, Ben-Gal I. Distinguishing the Leading Agents in Classification Problems Using the Entropy-Based Metric. Entropy. 2024; 26(4):318. https://doi.org/10.3390/e26040318

Chicago/Turabian Style

Kagan, Evgeny, and Irad Ben-Gal. 2024. "Distinguishing the Leading Agents in Classification Problems Using the Entropy-Based Metric" Entropy 26, no. 4: 318. https://doi.org/10.3390/e26040318

APA Style

Kagan, E., & Ben-Gal, I. (2024). Distinguishing the Leading Agents in Classification Problems Using the Entropy-Based Metric. Entropy, 26(4), 318. https://doi.org/10.3390/e26040318

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop