1. Introduction
Social Network Analysis (SNA) has increasingly gained attention in the research community over the past few years, emerging as a method of analysis of intra-team passing networks in team sports [
1,
2,
3,
4,
5]. In football, passing networks between players (nodes) reflect how the entities are connected and bring light to how each entity within that group interacts with the others [
1,
2,
4,
6]. Group and individual metrics are used to describe the team’s interactive behaviors using different degree measures to depict how connected the team is, as well as to identify the influential players within the team. By applying degree centrality and betweenness and closeness metrics to the global interaction of players, it is possible to assess the intermediary role of each player in distributing the ball during the game, allowing the identification of players with more relevant roles within the team [
1,
3,
7,
8,
9,
10].
uPATO is a dedicated software tool for network analysis in team sports [
11,
12]. In the last few years, the use of network analysis tools has been applied to team sports to understand how collective and individual performance may be optimized. These tools consider the network of passes between players during a game, which defines an adjacency matrix. Then, this matrix can be analyzed based on graph theory, which allows us to know which nodes (i.e., players) are more important and more involved in the game.
Another approach that can be performed to study ball passing networks created during the game is by using a Markov chain. According to this approach, the transition probability
, represents a directed edge between player
j to player
i, and all the edges between possible player are given in advance. In this, every pass is independent of the previous sequence of passes [
13,
14]. Yamamoto [
13], for example, has analyzed passing sequences in football using a Markov chain approach to estimate the predictability of certain passing courses and to determine when there is a chance of error in the information system—i.e., when a sequence of passes fails and the team loses possession.
Determining the degree of entropy of the probability matrices associated to Markov chains has not yet been explored in team sports analysis. According to entropy theory, higher levels of entropy show a greater unpredictability in how the system behaves. In our case, this means that the player (or the team) has a passing pattern that is more difficult to predict, and therefore more difficult to defend.
The aim of this paper is to present several novel mathematical models for pattern analysis in team sports, particularly football, to analyze the level of entropy in passing networks using a Markov chain approach. The novelty of this study relies in an innovative approach that proposes to estimate the likelihood of one player passing to, or receiving from, any other in the weighted directed network, as well as an estimation of the level of unpredictability of that player. In practical terms, higher levels of unpredictability will be reflected in a greater difficulty in predicting where to intercept the pass. Conversely, players with more predictable destinations will be more easily defendable and countered.
2. Novel Mathematical Models for Entropy of Nodes and Weighted Directed Networks
In this section, we present mathematical concepts that are based on information theory and probability theory that can be applied to networks when such networks can be considered weighted digraphs or weighted directed networks. The relation between the nodes of a network presented in this paper is described, in graph theory, by weighted digraphs [
12,
15,
16]. Therefore, we consider that
is the weighted adjacency matrix of a weighted digraph with n nodes,
.
Based on the mathematical concept of conditional probability involving two random variables
X and
Y, such that the pair of transmitter and receiver
is the joint distribution of the Markov chain associated with the weighted adjacency matrix,
, of a weighted digraph with n nodes,
, [
17](p. 13)–[
18,
19,
20] (p. 156), we can define the concept of the Markov chain transition matrix.
Definition 1. Thereal matrix of order n,, is called the transition matrix of the Markov chain associated with theof awithn nodes. Each element ofis obtained by:where the two random variables X and Y are such that the pair of transmitter and receiveris the joint distribution of the Markov chain associated with theof,. Considering the mathematical concept of
n-step transition probabilities [
20] (p. 157) and a weighted digraph with
n nodes, we can define the concept of
k-step node transition.
Definition 2. Given a, the transition matrix of the Markov chain associated with the weighted adjacency matrix,, of a weighted digraph,, with n nodes, eachelement of matrixis called the k-step node transition, since they give the probabilities of transitions fromtoin k time steps, whereis the power k of matrix,and.
Remark 1. In Definition 2, if, then=, and sois called node transition.
Based on Theorem 8.3 [
20] (p. 156) and a weighted digraph with
n nodes, we can define the probability of each node after the k times step in the network.
Definition 3. Given a, thetransition matrix of the Markov chainassociated with theof awith n nodes, the probability of all the nodes in the network after k times steps is obtained by:whereand. Considering the mathematical concept of entropy of a random variable, defined on the sample space of a random experiment and taking on a finite number [
17] (p. 51) [
21] (p. 19), and a weighted digraph with
n nodes, we define the concept of node out-entropy.
Definition 4. Given a weighted digraph,with n nodes,is called the node out-entropy of a node, and is determined by:
whereare the elements ofassociated with theof, and,. Remark 2. In Definition 4, if we replacewith, we obtain the concept node in-entropy,.
Based on the mathematical concept of conditional entropy [
21] (p. 22) [
22] (p. 70) and two random variables X and Y, such that the pair of transmitter and receiver
is the joint distribution of the Markov chain associated with the weighted adjacency matrix of a weighted digraph with
n nodes, we can define the concept of node transition out-entropy.
Definition 5. Given a weighted digraph,withn nodes, theis called node transition out-entropy of a nodeand is determined by:whereis the total links of, andnode out-entropy of node,. Remark 3. In Definition 5, ifwe replacewith, andwith, we obtain the concept of node transition in-entropy,.
Considering the mathematical concept of relative entropy between two random variables
X and
Y, such that the pair of transmitter and receiver
is the joint distribution of the Markov chain associated with the weighted adjacency matrix of a weighted digraph with
n nodes [
21] (p. 248) [
23] (p. 25), we can define the concept of the relative out-entropy between two nodes.
Proposition 1. Given a weighted digraph,with n nodes and two distributions,and, with two random variables X and Y, such that the pair of transmitter and receiveris the joint distribution of the Markov chain associated with theof, the relative out-entropy between two nodesand,, is obtained by:whereis thenode out-entropy of a node, andandarethe elements ofassociated with theof. Proof. The relative entropy between two distributions [
23] (p. 247) is defined as:
Considering
and
, we can write:
By expression 3, we can write:
□
Remark 4. In Proposition 1, ifwe replacewithandwith, we obtain the concept relative in-entropy,.
Considering the mathematical concept of conditional entropy between two random variables
X and
Y, such that the pair of transmitter and receiver
is the joint distribution of the Markov chain associated with the weighted adjacency matrix of a weighted digraph with
n nodes [
21] (p. 22), [
22] (p. 70), we can define the concept of network transition out-entropy.
Proposition 2. Given a weighted digraph,with n nodes and two random variables X and Y, such that the pair of transmitter and receiveris the joint distribution of the Markov chain associated with theof,is the network transition out-entropy of aand is determined by:whereis thenode transition entropy of a node. Proof. Consider two random variables X and Y, such that the pair of transmitter and receiver is the joint distribution of the Markov chain associated with the of .
The conditional entropy [
22] (p. 70) is defined as:
□
Remark 5. In Proposition 2, ifwe replacewithandwith, we obtain the concept network transition in-entropy,.
Considering the mathematical concept of joint entropy [
19] (p. 394) [
21] (p. 22) [
22] (p. 69) and a weighted digraph with
n nodes, we can define the concept of total network entropy.
Proposition 3. Given a weighted digraph,with n nodes and two random variables X and Y, such that the pair of transmitter and receiveris the joint distribution of the Markov chain associated with theof, the total network entropy,, is obtained by:whereis the total links of,. Proof. Consider two random variables
X and
Y, such that the pair of transmitter and receiver
is the joint distribution of the Markov chain associated with the
of
. The joint entropy [
19] (p. 236) is defined as:
□
Considering the mathematical concept of relative entropy between two random variables
X and
Y, such that the pair of transmitter and receiver
is the joint distribution of the Markov chain associated with the weighted adjacency matrix of a weighted digraph with
n nodes [
21] (p. 247) [
23] (p. 25), we can define the concept of network relative out-entropy.
Proposition 4. Given a weighted digraph,with n nodes and two random variables X and Y, such that the pair of transmitter and receiveris the joint distribution of the Markov chain associated with theof. The network relative out-entropy,, is obtained by:where. Proof. Consider two distributions,
and
, with two random variables
X and
Y, such that the pair of transmitter and receiver
is the joint distribution of the Markov chain associated with the
of
. By the concept of relative entropy [
21] (p. 249), we obtain:
□
Remark 6. In Proposition 4, if we consider, we obtain the concept network relative in-entropy,.
3. A Case Study
The 2015/2016 Champions League final is presented as the case study for the presentation and interpretation of the metrics proposed. The two opposing teams were Real Madrid (RM) and Atletico Madrid (AM). After a 1-1 draw during regular time, RM became champion with a 5-3 win after a penalty shoot-out. For the notational analysis of each pass sequence (transition between nodes), the uPATO platform was used [
11,
12]. From this, adjacency matrices were computed to calculate the Markov chains, represented in
Appendix A as the probability of state out-transition matrices—i.e., the probability of each player (node) passing (transitioning information) to any other player (node) present in the game (network). Note that, before computing the Markov chains, a Laplace Smoothing [
18] was performed upon the weighted adjacency matrices. This smoothing technique serves the purpose of avoiding the pitfall of estimating chances that appear impossible. In our case, we use the special case of Laplace Smoothing, where 1 was added to every connection between two distinct nodes, also called add-one smoothing.
The RM starting 11 were the following: player 1: Keylor Navas (1); player 2: Dani Carvajal (15); player 3: Sergio Ramos (4); player 4: Pepe (3); player 5: Marcelo (12); player 6: Casemiro (14); player 7: Luka Modric (19); player 8: Toni Kroos (8); player 9: Karim Benzema (9); player 10: Gareth Bale (11); and player 11: Cristiano Ronaldo (7). The AM starting 11 constituted the following players: player 1: Jan Oblak (13); player 2: Juanfran (20); player 3: Stefan Savic (15); player 4: Diego Godin (2); player 5: Filipe Luís (3); player 6: Augusto Fernández (12); player 7: Koke (6); player 8: Gabi (14); player 9: Antoine Griezmann (7); player 10: Fernando Torres (9); and player 11: Saúl Ñiguez (17).
Table A1 and
Table A2, presented in the
Appendix A section, show RM and AM’s state out-transition probability matrices, there are two nodes with probabilities over 0.3. For RM, transitioning between player 2 and 10 had a 0.33 chance of occurring, and there was a 0.34 chance between player 3 and 5. This might mean that when the ball is in 2, the game is vertical through the center, in search for a creative midfielder. Conversely, player 3 searches for a supported, winged play. In AM’s game, passing between 2 and 8 had a chance of 0.30, and between 5 and 11 has a chance of 0.37.
The state in-transition probabilities were also calculated (
Table A3 and
Table A4), representing the probability of each player receiving the ball from each other player. Contrarily to the out-transition probabilities, these tables should be read column-wise. Here, the reception probabilities in RM’s game only show player 5 receiving the ball from player 3, with a 0.32 probability of occurring. In AM’s game, the same happened in only one case. In practical terms, for RM’s game, a key interaction between two players (player 3 and 5) may have been identified. Specifically, player 5 predominantly receives the ball from 3, and the latter predominantly passes to 5. In AM’s game, it is player 6 that has a greater chance of receiving the ball from 11.
Having this information might be useful for the opposing team—for example, in order to adapt their defensive tactics to reduce the chances of interaction between players. Knowing how team players interact with each other may identify the prime pathways to feed the attacking players and eventually reduce the chances of creating goal opportunities. This type of empirical observations requires, in our opinion, further analysis in order to produce ratios and cut-off values.
Applying a nonlinear-based approach to state transition matrices in football has not been done yet. Theoretically, higher levels of entropy reflect greater variability of node transitions. Node Transition Entropy was calculated for each node of the team to represent the degree of transition variability of each player, based on the probability of it happening. The more positive a value is, the more that node contributes to the overall entropy of the network when compared to the other nodes.
In the examples below (
Table A5 and
Table A6), we selected the cut-off value of |0.90| for the Relative Out-Transition Entropies. RM showed more values above our stipulated limit, showing that there were more players with more chaotic passing probabilities. Regarding the AM relative entropy values, only two players showed more chaotic behavior in their passing probabilities, and this difference is against the goalkeeper (node 1). This shows that AM players were more consistent, and therefore more predictable, in their passing patterns. It is important to note that the defined Relative Out/In-Transition Entropies can both have negative and positive values. Negative values indicate that the first node in the pair is more chaotic—i.e.,
means that
is more chaotic than
. The inverse also applies, where if
,
is less chaotic than
. In the case of
, both nodes have the same chaocicity, meaning that they add the same value to the network as a whole.
Similarly, in-transition entropy reflects the degree of variability in receiving the ball (
Table A7 and
Table A8). Applying the same criteria, we can note that for RM, two players (player 5 and 8) showed chaoticity in the probability of receiving passes from other teammates. In line with what has been previously observed, AM players also had smaller entropy values for this metric, again showing a more predictable behavior. In this case, no AM player presented values above the selected cut-off value.
The final step for this case study is to calculate team metrics for each team, presented in
Table 1. Network Relative Out-Entropy (NROE) and Network Relative In-Entropy (NRIE) show the consistency of the network. Here, higher values of entropy reflect a smaller consistency of interactions between players or, in other terms, a more unpredictable passing pattern. In our football example, both values were higher for Real Madrid, which means that the team was more unpredictable both in passing and in receiving the ball from the other players. It has been shown that unpredictable passing patterns may create more goal-scoring opportunities and contribute to ball possession [
1,
24]. This may be due to the fact that the team has more players involved in each passing sequence, increasing the unpredictability of each play. In this case, success may be due to either the number of passes or the time it possessed the ball, and not to higher entropy values [
25]. That was not the case, as RM had fewer passes than AM. We may regard entropy, therefore, as a positive feature in the game. Real Madrid’s unpredictability in passing may have increased the chances of goal-scoring opportunities, potentially having contributed to the win.
The concept of entropy to model social interactions has already been used [
26,
27,
28]. Newman and Vilenchik (2019) used the concept of relative entropy to model the interactions of players passing the ball in football, having found that when comparing two opposing teams, higher entropy values lead to more chances of creating goal-scoring opportunities [
27]. The authors limit their analysis to whether an interaction between players occurred, without defining the type of interaction—passing or receiving the ball. In other words, direction did not matter. In football, however, variability in passing patterns may come from these two actions and they may lead to different conclusions: a team can be more unpredictable in passing than in receiving the ball. Fewer players passing the ball to many players is different from many players passing the ball to a few. Concomitantly, game strategy must take this into account and adapt accordingly.
The division between the Network Transition Out-Entropy (NTOE) and Network Transition In-Entropy (NTIE) separates the teams’ out and in interactions. NTOE, therefore shows the entropy of the network in sending the information to other nodes. Conversely, NTIE aims to show the entropy of receiving information from other nodes. For our football example, NTOE and NTIE are the entropy levels of the players’ probability of passes and receptions, respectively.
NTOE and NTIE were slightly higher for Real Madrid, showing again that this team was more unpredictable in the interactions. Against an opponent, higher levels of entropy may prevent the defending team from learning the patterns the other team. It is worth noticing that the division in the out and in entropies allows, in our opinion, a deeper understanding of the team’s dynamics by extending the network analysis to more than the existence (or not) of interactions between nodes, taking into account the direction of information and its probability of occurring.
Total Network Entropy (TNE) reflects the degree of variability and therefore the unpredictability of the team as a whole. Real Madrid was slightly more unpredictable than Atletico, with a TNE of 6.44 vs. 6.22. The unpredictability of passing patterns however, does not happen isolated from other actions; it is, rather, dependent on the interactions between rivals and teammates, space occupation, time, tactics, and game situation [
10,
29]. This leads to the need to constantly analyze how the team behaves during the different phases of the game, as they may behave differently. For instance, if, when winning, ball possession is kept using the same passing patterns or variability, the unpredictability decreases.