Next Article in Journal
Analysis of Quantum Correlations Obtained Using Local Optimal Universal Asymmetric Cloners
Next Article in Special Issue
An Efficient Quantum Secret Sharing Scheme Based on Restricted Threshold Access Structure
Previous Article in Journal
Position/Attitude Cascade Regulation of Nonholonomic Four-Wheeled Mobile Robot via Double-Loop Sliding-Mode Control Mechanism
Previous Article in Special Issue
Path Planning Research of a UAV Base Station Searching for Disaster Victims’ Location Information Based on Deep Reinforcement Learning
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Multi-Objective Multi-Instance Learning: A New Approach to Machine Learning for eSports

Department of Computer Engineering, Dokuz Eylul University, Izmir 35390, Turkey
*
Author to whom correspondence should be addressed.
Entropy 2023, 25(1), 28; https://doi.org/10.3390/e25010028
Submission received: 22 November 2022 / Revised: 18 December 2022 / Accepted: 20 December 2022 / Published: 23 December 2022
(This article belongs to the Special Issue Advances in Information Sciences and Applications)

Abstract

:
The aim of this study is to develop a new approach to be able to correctly predict the outcome of electronic sports (eSports) matches using machine learning methods. Previous research has emphasized player-centric prediction and has used standard (single-instance) classification techniques. However, a team-centric classification is required since team cooperation is essential in completing game missions and achieving final success. To bridge this gap, in this study, we propose a new approach, called Multi-Objective Multi-Instance Learning (MOMIL). It is the first study that applies the multi-instance learning technique to make win predictions in eSports. The proposed approach jointly considers the objectives of the players in a team to capture relationships between players during the classification. In this study, entropy was used as a measure to determine the impurity (uncertainty) of the training dataset when building decision trees for classification. The experiments that were carried out on a publicly available eSports dataset show that the proposed multi-objective multi-instance classification approach outperforms the standard classification approach in terms of accuracy. Unlike the previous studies, we built the models on season-based data. Our approach is up to 95% accurate for win prediction in eSports. Our method achieved higher performance than the state-of-the-art methods tested on the same dataset.

1. Introduction

Electronic sports (eSports) is a general term used to describe online digital games that are played professionally or amateurishly by teams and watched by a large number of audiences. Naturally, eSports is a significant research area in both the scientific community and industry, in terms of not just size, but also commercial value [1]. eSports games generate huge amounts of statistical match data that are publicly available, allowing us to extract significant insights. Multiplayer Online Battle Arena (MOBA) provides a great opportunity for machine learning (ML) with the availability of high-volume and high-dimensional data. The data dimension is high since it involves many different attributes in three main categories (pre-game features, in-game features, and post-game features) such as player-related information (e.g., champion character, role, and position), match-related information (e.g., season, platform, duration, and software version), team-related information (i.e., kills, assists, deaths, damages, healing, wards, vision score, level, rewards, bans, and golds earned/spent for each player separately), information regarding coaches and trainers, characters to be selected, items to be purchased, and statistics (e.g., blood, tower, inhibitor, dragon, baron, and win–lose). Formally, standard eSports analytics was defined by Schubert et al. [2] as follows: the process of using eSports-related data to find and visualize useful and meaningful patterns/trends to assist with decision-making processes. This definition highlights a fundamental aspect: the opportunity of using machine learning techniques to predict match outcomes.
The eSports industry has already become a highly profitable industry with total revenue of $159.3 billion [3] and over 645 million audiences [4]. The prediction of eSports competition has important impacts on market size and growth, revenue, sponsorship, and media coverage [5]. Therefore, forecasting match outcomes is requested by all the stakeholders such as professional players, amateurs, coaches, trainers, organizers, sponsors, audiences (fans), and media workers. In this way, they can develop tactics to gain advantages in eSports competitions [6,7]. To meet this demand, ML models have been developed in several studies [8,9,10]. In other words, the strengths of ML methods in making predictions related to eSports have been proven in previous studies [11,12].
The previous studies [6,7,8,9,10,11,12,13,14,15] related to prediction in eSports have several limitations. First, some studies [8,9] have been mainly focused on in-game predictions, aiming to inform audiences and players. Second, some of them [16,17] have built models for player skill prediction, and so they used sensors such as eye trackers, keyboard/mouse loggers, electroencephalography (EEG) headsets, pulse-oximeters, heart rate monitors, and chair seat/back sensors, which limit the area of its application due to hardware requirements. Third, they [9,10,18] used data collected over long time periods; however, eSports games are continuously updated, and these major game changes can remarkably alter the fundamental characteristics of the games. Since major updates might make previous data obsolete, we carried out a season-based analysis in order to overcome this limitation. Last and foremost, the previous studies [8,15,16,17,19] are player-centric and use a standard (single-instance) classification technique. However, the strong performance of a single player does not guarantee a win for the team, and a weak performance of a single player does not guarantee a loss [7]. Team-level classification is required since team cooperation is essential in completing game missions and achieving final success. To be able to predict the match outcome by considering multiple players in a team, multiple instance classification is required; however, no previous prediction models have been adopted for multi-instance learning.
Multi-instance learning (MIL) [20] is a special type of machine learning (ML) where multiple training samples are assigned into bags and only one class label is assigned for all the samples in a bag. In other words, it learns from a set of training bags that involves multiple feature vectors. In our study, each bag corresponds to a team and contains five separate feature vectors for five players in the team. If the team wins the game, its bag is associated with the class label 1, and otherwise, the label is 0. We propose using MIL since collaboration among players on each team remarkably affects which team will win or lose.
The novelty and contributions of this article can be summarized as follows. (i) It proposes a new approach, called Multi-Objective Multi-Instance Learning (MOMIL). (ii) It is the first study that applies the multi-instance learning technique to predict the match outcome in eSports. (iii) It considers the multi-objective concept in MIL for the first time, since each team member has a specific role (objective) in eSports matches. Importantly, the aim of our study is to provide an algorithmic contribution towards increasing the classification performance of models compared to the previous studies [13,21,22,23,24]. (iv) Unlike previous studies, we built the models on season-based data due to the regularity of changes and updates in the game versions. Thereby, our study is also original in that it considers model usage lifespan first-time and presents extended results and analyses for win prediction. (v) Our method achieved higher performance than the state-of-the-art methods [13,21,22,23,24] tested on the same dataset.
Entropy is the elementary measure used in this study when building decision trees. The experiments that were carried out on a publicly available LoL dataset show that the proposed approach outperforms the standard classification approach in terms of accuracy. Using multi-instance learning algorithms, our approach is up to 95% accurate for win prediction in LoL seasons.
The remainder of the article is organized as follows. In the following section, we provide a comprehensive literature review of ML in eSports. Furthermore, we give background information on LoL and provide definitions of multi-objective and multi-instance learning. Section 3 explains the proposed MOMIL approach. Section 4 describes four different experiments and the experimental results. The Section 5 summarizes the study and suggests possible future works.

2. Related Work

2.1. Literature Review

eSports is currently referred to as one of the major international and popular sports with millions of amateur and professional players and spectators. Unlike original sports, eSports can be performed without being dependent on place and time in its nature; therefore, a large number of matches have been held every day, resulting in a huge amount of data and opportunities for ML studies.
Table 1 shows a comparison of our study with the previous studies [6,7,8,9,10,11,12,13,14,15,16,17,18,19,23,24,25,26,27,28,29,30,31,32]. It provides a brief description, an overview of algorithms, the task performed, the genre of the game, and whether multi-instance learning was adopted or not. Many studies have been focused on classification (CF) [6,7,8,9,10,12,13,15,16,17,19,23,24,28,31] and regression (R) [11,14,18,27,32] tasks; however, recently, clustering (CL) [29,30], and association rule mining (ARM) [24] tasks have received increasing attention from researchers.
eSports analytics research has focused on different problems such as predicting match outcomes [6,7,10,12,13,14,15], recommending items [23,24], predicting the ranking of players [18], classifying eSports games [28], identifying roles [29], predicting player churn [32], clustering player-centric networks [30], predicting player skill, and re-identification [16,17]. Since win prediction is one of the most important problems and it has commercial value, in this study, we focused on this problem. However, match outcome prediction in eSports is very different from the win prediction in physical sports [33,34].
In the literature, a variety of ML algorithms have been used for prediction problems in eSports. Much of the previous work [6,7,8,9,10,16,23,24,27] used Linear or Logistic Regression (LR); however, other regression methods such as Mixed-Effects Cox Regression (MECR) [32] and Bayesian Hierarchical Regression (BHR) [14] have also been applied to solve a prediction problem in eSports. The main drawback of linear regression is that it can model linear dependencies in the data. On the other hand, the disadvantage of logistic regression is the difficulty of detecting complex relationships between data instances. The most widely used classification techniques are Neural Networks (NN) [8,11,12,23,24], Decision Tree (DT) [23,24,27,28,31], Classification and Regression Tree (CART) [7,12], K-Nearest Neighbors (KNN) [8,13,15,16,31], Support Vector Machine (SVM) [15,16,27], and Naive Bayes (NB) [16,28]. However, building a single classifier with these techniques may not be strong and stable enough. This issue can be overcome by using an ensemble model. As types of ensemble learning, the Random Forest (RF) [6,10,16,18,28,31], Gradient Boosting Machines (GBM) [6,10,18], and Extremely Randomized Trees (ERT) [17] methods were tested in some studies. The advantage of ensemble learning over the single classifier is the ability to combine the prediction outputs from multiple estimators to improve generalization ability and robustness. An incorrect prediction of an ensemble member can be corrected by other members thanks to the majority voting. However, one of the drawbacks of ensemble learning techniques is that they increase computation time and model complexity. The sequential nature of team fights in eSports may make non-deep learning models less suitable. Therefore, deep learning techniques such as Long Short-Term Memory (LTSM) [25], Deep Neural Network (DNN) [19], and Convolutional Neural Networks (CNN) [12,23,26] have also been used in predicting match outcomes. Principal Component Analysis (PCA) [31], Linear Discriminant Analysis (LDA) [15], and Quadratic Discriminant Analysis (QDA) [15] methods have also been applied to eSports data. The drawback of PCA and LDA techniques is that they define a linear projection of the data, and therefore, the scope of their application is somewhat limited. On the other hand, QDA has the advantage of separating non-linear data.
Multi-instance learning [35] is one of the most exciting technologies that implement entropy in computer science [36]. The entropy of a dataset is a good measure to determine its impurity or unpredictability. In this study, we used entropy to characterize the structure of the data when building a decision tree. In addition to multi-instance models, multi-objective [37] and multi-strategy models [38] have also been successfully developed in the literature.
Some studies in the literature have mainly used pre-game features [10] or in-game features [6], while we consider post-game features. Pre-game features are generated in the character and item selection phase before a match starts. For example, Araujo et al [23,24] recommend giving the most suitable item set to each team member to increase the performances of their characters. In-game features are used for live (real-time) predictions. For instance, Katona et al. [19] predicted whether a player will die within the next five seconds. Post-game features are generated to summarize the game, notably at the end of the game, such as kills, rewards, damages, gold earned by each player, and match duration. Kadan et al. [31] subdivided games into several intervals and predicted the round results.
The majority of work in eSports has focused on constructing ML models for well-known games, including League of Legends (LoL) [8,9,11,12,13,16,23,24,28,30,32], Defense of the Ancients 2 (Dota 2) [6,7,10,14,19,28,29], PlayerUnknown’s Battlegrounds (PUBG) [18,27], Counter-Strike (CS) [17,28,31], and StarCraft [15]. Since LoL is the most popular game in eSports in the world, in this study, we focused on this game.
Our study differs from existing studies in many respects. First, our study is the first study that uses multi-instance learning in eSports. Second, data analysis techniques in the previous studies are mainly classical ML algorithms such as LR, SVM, RF, DR, NN, and KNN, while we used different algorithms such as Multi-Instance Tree Inducer (MITI), Multi-Instance Rule Inducer (MIRI), Two-Level-Classification (TLC), and Multi-Instance Wrapper (MIWrapper) algorithms. Third, it is designed to be season-based since each season involves many changes in the meta-game that affect winner prediction. In this way, we overcome the limitations of current studies, meaning our approach generalizes well to different game versions.

2.2. Background Information

2.2.1. League of Legends

LoL is a MOBA game, where the blue team and the red team compete against each other on a single map (called Summoner’s Rift) to destroy the opposing base (called Nexus) first. The map contains three main roads (called lanes), which are connected by a base of each team. The game typically consists of five players on each team, and each player (called summoner) controls a single character (called champion). Once the match begins, each player selects a champion to play over 120 champions and goes to their respective lane after buying some starting items with golds. Items are objects (e.g., weapons, armor) that are used for providing an improvement on champions. Gold or experience points (XP) is the in-game currency for an LoL match to buy performance-enhancing items. There are also computer-controlled monsters such as Dragon, Baron, and Rift Herald. Throughout a match, the players gain gold and experience in a variety of ways such as killing enemy champions or monsters and destroying defense buildings (called Towers) of the enemy team. When a player is killed, their champion is reanimated after a time-out increasing in accordance with the champion’s XP. Each match lasts approximately 20–40 min on average. Finally, whichever team destroys the Nexus in the opponent’s base first obtains the victory. Players are ranked according to their skill levels (called ladder) as follows: Bronze, Silver, Gold, Platinum, Diamond, Master, and Challenger.
LoL is a team game and the final performance in the game under investigation depends on the relationship among players in the team [7]. At the beginning of the game, each player in a five-member team selects a single character, which has various advantages and disadvantages to contribute to the overall strategy of the team [39]. The ability of a character can become more valuable with its harmony with other characters. In other words, the ability of a character can be dependent on the interaction with another characters. As it is a team-based game, each player in the team typically plays a certain role (objective) in a match, similar to traditional sports. For example, in the LoL game, team members play in a cooperative manner by selecting a position, including top lane, mid lane, jungle, attack damage carry, and support. Players in a team influence each other by means of cooperation; e.g., one player can defend another player who is attacking the enemy. The studies [7,39,40] in the literature have emphasized the importance of teamwork and compromise. For example, Lan et al. [39] indicated that the outcome of a game is determined both by each player’s behavior and by the interactions among players. Gu et al. [40] also noted that cooperation among teammates should be considered for match outcome prediction. As emphasized by a number of studies [7,39,40], the objectives, positions, and abilities of five players are all together relevant for the team’s success. Since team collaboration is critical for success, we propose a team-level classification approach in this paper.

2.2.2. Multi-Objective and Multi-Instance Learning

Multi-objective learning is a type of learning in which each member has a special role (task) and has a probability of reaching the target goal, and the final result is determined by jointly considering the features of members. For example, in an eSports match, each player is responsible for one task on a team and has a success rate, players influence each other by means of cooperation, and the team with the better performance wins.
Multi-instance learning (MIL) is a special case of learning where the algorithm learns from bags of instances, rather than single instances. The aim of MIL is to classify bags according to several unseen instances.
Figure 1 illustrates the difference between the standard classification and multi-instance classification. In traditional classification, each instance (a feature vector) is assigned to a certain class label. Nevertheless, in multi-instance classification, a training set consists of bags containing several feature vectors, and each bag is associated with a class label. In other words, a multi-instance classification algorithm learns from a dataset that contains bags of training instances, rather than single training instances. In our study, each bag corresponds to a team in the LoL game and contains five separate feature vectors (match statistics) for five players belonging to the team. If a team wins the game, its bag is associated with the class label 1, and otherwise the label is 0.

3. Materials and Methods

3.1. The Proposed Approach

Using the standard (single-instance) classification to estimate the match outcome of an eSports game has attracted considerable attention [6,7,10,12,13,14,15]; however, multi-instance classification is as yet unknown. Multi-instance classification is, however, required because it provides the ability to jointly consider multiple instances (here multiple players in a team) when classifying. In order to bridge this gap, this paper proposes a new approach: MOMIL.
MOMIL learns from multi-instance data, builds a classification model by simultaneously considering multiple instances, and then uses the model to predict the winning probability of a team for unseen vectors. Considering data from multiple players jointly can help to capture relationships among players and to explore team-level mechanics that are particularly relevant since most of the games in eSports are team-based games. Our study especially investigates the winning predictors in LoL.
Figure 2 presents the general overview of the proposed MOMIL approach, which consists of five main stages. (i) In the first stage, named data collection, the raw data are collected from the game environment, including information about matches, champions, players, items purchased, and statistics. (ii) In the second stage, features are generated, which summarize the games, including kills, assists, deaths, damages, rewards, and golds earned by each player. (iii) In the data pre-processing stage, single-instance data are transformed into multi-instance data. Since each team in LoL is composed of five players, every five sequential records in the training set were merged in a straightforward manner, from solo to full team composition. (iv) In the next step, predictive models were built by using multi-instance learning algorithms such as Multi-Instance Support Vector Machines (MISVM), Multiple Instance Logistic Regression (MILR), Multi-Instance AdaBoost (MIBoost), Multi-Instance Tree Inducer (MITI), and Multi-Instance Rule Inducer (MIRI). (v) In the fifth stage, the predictive models are tested to evaluate their performances by using various metrics such as accuracy, recall, precision, F-measure, and the area under the curve of the receiver operating characteristic (AUC-ROC). In this stage, the k-fold cross-validation technique is used for validation, in which the data are divided into k equal subsets, using k 1 folds as the training set and one fold as the test set. Finally, in the win prediction stage, the match outcome is predicted according to a given sample.
The aim of this study is to develop an intelligent model that discovers useful patterns and rules to be able to estimate the match outcomes of eSports games. Our approach (MOMIL) improves win prediction as it makes team-level analysis by taking into account the objectives of the players in the team together during the classification. It investigates the correlation between match outcome (class label) and multiple objectives simultaneously rather than individually.
The comparison of our approach against traditional ones can be summarized as follows. Rather than player-centric prediction, we performed a team-centric prediction. Instead of using standard single-instance classification algorithms, we used multi-instance learning algorithms. Our approach considers the multi-objective concept since each team member has a specific role (objective) in eSports matches. Unlike making predictions over long time periods, we built the models on season-based data since each season involves many changes in the environment.

3.2. Formal Description of the Proposed Approach

Let X is a d-dimensional instance such that X = ( x 1 , x 2 , , x d ) . Let O objective types such that O = { O 1 , O 2 , , O m } . A bag B includes a set of pairs of instances and their obejectives such that B = ( ( X 1 , O 1 ) , ( X 2 , O 2 ) , , ( X m , O m ) ) , where m is the number of intances in a bag, as well as the count of objectives. Training dataset D contains a set of pairs of bags and their corresponding class labels such that D = { < B 1 , C 1 > , < B 2 , C 2 > , , < B n , C n > } , where C i is the class label of the bag B i . The target class attribute has k labels such that C i { C 1 , C 2 , . , C k } for i = 1 , 2 , , k . For instance, in a binary classification, the class labels of the bags can be one or zero, i.e., C 1 = 1 and C 2 = 0 . The aim is to build a multi-objective multi-instance classifier model M to successfully label given query bags. The multi-instance classifier predicts the output at the bag-level and makes a decision for a given query bag. In fact, a traditional classifier is a special type in which each bag includes only one instance, as well as only one objective B i = ( X i ) .
Definition 1.
MOMIL is an approach that is applied to a dataset that consists of a set of bags D = { < B i , C i > } i = 1 n , wgere each bag contains a set of instances, each of which has an objective such that B = ( ( X 1 , O 1 ) , ( X 2 , O 2 ) , , ( X m , O m ) ) . First, a multi-objective multi-instance dataset (D) is generated from a traditional (single-instance) training set (S), and then a classifier model (M) is trained directly on such data. In this way, the MOMIL approach jointly considers relationships and cooperation among data instances during classification.
Algorithm 1 presents the pseudo-code of the proposed MOMIL method. In the first loop, the single-instance data (S) are transformed into multi-instance data (D). When each team in the game is composed of m players, every sequential m records in the data are merged in a straightforward manner, from solo to full team composition. In this way, the player-level match statistics are taken into account as single-instance data, while team-level features are considered as multi-instance data. After that, the algorithm builds the model (M) on multi-instance data. The constructed model maps input vectors representing game statistics of the players in the teams to output labels (win or loss). Finally, in the last loop, the winners are estimated for the given test data. At the ith iteration, the test query T i is classified by the model M and the predicted class labels are stored in a data structure (Y). The time complexity of the proposed MOMIL approach is O ( L ( n ) + n m ) , where m is the number of players in a team, n is the number of bags in the training set, and L is the time required for the learning process on n bags.
Algorithm 1: Multi-Objective Multi-Instance Learning (MOMIL)
Inputs:
S: single-instace data S = { ( x 1 , y 1 ) , ( x 2 , y 2 ) , . , ( x n m , y n m ) }
T: test set containing b bags
m: the number of players in a team
n: the number of bags in the training set
Outputs:
Y: predicted class labels on given query bags
Begin:
for: i = 1 tondo
Entropy 25 00028 i001
In this study, entropy was used to measure the homogeneity of the dataset when building a decision tree. At each node of the tree, the splitting criterion is the normalized information gain (difference in entropy). The entropy of a dataset is high if the number of samples in the classes is close to each other. On the other hand, the entropy value is small if there is a class that includes most of the samples in the dataset. The algorithm tries to minimize the entropy since a small entropy means that an instance can be classified with high probability. The entropy of a training data set S is calculated as follows:
E n t r o p y ( S ) = i = 1 k P i l o g 2 P i
where P i refers to the ratio of the samples belonging to class i and k is the number of the classes.

3.3. Advantages of the Proposed Approach

The proposed MOMIL approach has many advantages as follows:
  • The proposed approach utilizes multi-instance learning to predict the match outcome in eSports for the first time. Thereby, it expands the standard classification task in the field of eSports.
  • MOMIL is designed for team-level classification. This property increases the performance of the predictive model, since team collaboration is essential in completing game missions and achieving final success.
  • eSports games are continuously updated, and the major game changes make the previous models obsolete. The proposed model overcomes this limitation by using data collected over short time periods.
  • Another advantage of MOMIL is its implementation simplicity. After converting single-instance data into multi-instance data in a straightforward manner, any multi-instance learning algorithm can be applied easily.
  • An important advantage of the proposed MOMIL method is that it is designed for analyzing any type of data that are suitable for win prediction. Therefore, it can be easily applied without having any background information. It does not require any specific knowledge or assumption for the given data. Therefore, it can be widely applied to many different eSports games such as LoL, Dota 2, Destiny, PUBG, and Counter-Strike.
  • One of the key advantages is that it can be used for feature engineering which means identifying the most significant features for win prediction among the available features in the training bag set.
  • Another advantage of MOMIL is its ability to deal with non-linear and complex win prediction problems.

3.4. Multi-Instance Classification Algorithms

In this study, the following nine multi-instance learning algorithms were tested and compared with each other.
  • Multi-Instance AdaBoost (MIBoost) [41]: As an upgraded version of the AdaBoost algorithm, it takes into consideration the geometric mean of the posterior of instances inside a bag. Naturally, the class label of a bag is predicted in two steps. In the first step, the algorithm finds instance-level class probabilities for the separate instances in a bag. In the second step, it combines the instance-level probability forecasts into bag-level probability for assigning a label to this bag.
  • Multi-Instance Support Vector Machine (MISVM) [42]: First, the algorithm assigns the class label of the bag to each instance in this bag as its initial class label. Afterward, it applies the SVM solution for all the instances in positive bags and then reassigns the class labels of instances based on the SVM result. The objective in MISVM can be written as follows:
    min y i min w , b , φ 1 2 w 2 + c i φ i s . t . i : y i ( ( w , x i ) + b ) 1 φ i , φ i 0 , y i { 1 , 1 }
    where b is the bias vector, w is the projection matrix, { φ i } i = 1 N are the slack variables of the support vector machine, and y i is the class label of the training pattern x i .
  • Multi-Instance Logistic Regression (MILR) [43]: It performs collective multi-instance assumption by using logistic regression (LR) when evaluating bag-level probabilities. The instance-level class probability S i j with parameters θ = ( w , b ) is calculated as follows:
    S i j = 1 1 + e ( w . B i j + b )
    where w is a vector of feature weights, b is a bias parameter, B i j is the jth instance of the ith bag. The bag-level probability is estimated by using a softmax function, which combines the probabilities of the instances in the bag as follows:
    S i = p ( C i = 1 | B i ) = s o f t m a x α ( S i 1 , , S i n ) = j = 1 n S i j e α S i j j = 1 n e α S i j
    where α is a constant related to the softmax approximation, p ( C i j = 1 | B i ) is the posterior probability that the ith bag is positive for binary classification, and n is the number of bags in the training set.
  • Multi-Instance Tree Inducer (MITI) [44]: This algorithm builds a decision tree in a best-first strategy by using a simple priority queue. It applies a splitting criterion to divide each internal node data point into subsets. At the beginning of the algorithm, a weight w i b is assigned to each instance x i b such that 1 / | b | which is the inverse of the size of the bag b. After that, a weighted Gini-impurity metric is calculated for a set S as follows:
    G i n i ( S ) = x i b S w i b 2 x i b S w i b 2 x i b S + w i b 2
    where S and S + denote the negative and positive instances in S, respectively.
  • Multi-Instance Rule Inducer (MIRI) [45]: MIRI is an algorithm inspired by MITI. It is a multi-instance learner that benefits from partial MITI trees and generates a compact set of classification rules.
  • Summary Statistics for Propositionalization (SimpleMI) [46]: This method maps the instances in a bag to a single feature vector by analyzing statistical properties. Therefore, the basic idea is to transform a bag into a vector form first and then classify it by using a standard learner.
  • Multi-Instance Wrapper (MIWrapper) [47]: This algorithm assigns a weight to the instances in the bags. A class probability is calculated by the propositional model for each instance in the bag. After that, the predicted probabilities of instances are then averaged to assign a label to the bag.
  • Citation K-Nearest Neighbours (CitationKNN) [48]: It is an adapted version of KNN for MIL problems. To predict the class label of a query bag, the algorithm considers not only the nearest bags (called references), but also other bags that regard the query bag as one of their nearest neighbors (called citers). CitationKNN uses the Hausdorff measure to calculate the distance between two bags as follows:
    D i s t ( B , B ) = min 1 i p 1 j q ( D i s t ( b i , b j ) ) = min b B min b B b b
    where B and B indicate two bags, i.e., B = { b 1 , b 2 , , b p } and B = { b 1 , b 2 , b q } ; p and q are the number of instances in each bag, respectively; b and b are two different feature vectors; and b i and b j are instances from each bag.
  • Two-Level-Classification (TLC) [49]: In the first level, the instances in each bag are re-represented by a meta-instance by encoding the relationships among them. In the second level, the algorithm induces a function to capture the interactions between the meta-instance and the class label of the bag. TLC requires the selection of a partition generator (i.e., C4.5) and a classifier (i.e., LogitBoost) with decision stumps [50].

4. Experimental Studies

To demonstrate the effectiveness of the proposed MOMIL approach, the experiments were carried out on a publicly available LoL dataset. The algorithms in the multi-instance learning package in WEKA [50] were used with default parameters. As a base classifier, the C4.5 Decision Tree (DT) algorithm was preferred due to its efficiency. This algorithm calculates entropy, which is a powerful measure to determine how a tree node splits data.
We used the 10-fold cross-validation technique to evaluate the performances of classifiers. The performances of the classifiers were compared in terms of accuracy and F-measure. Accuracy is a metric that is widely used in many applications [16,28,51] to measure the success of a model. Formally, accuracy is the proportion of correct predictions to total predictions. It is a useful measure of the degree of the predictive power of the classifier and how it may generally perform. In addition to accuracy, we also compared the performances of the methods in terms of F-measure. Although accuracy is a useful metric in classification performance, it alone is not sufficient to determine the quality of the prediction since it makes no distinction between the classes. The F-measure is a useful measurement since it takes into consideration three quantities: false-negative (FN), true-positive (TP), and false-positive (FP). F-measure is a summary performance measurement as it is the harmonic mean of precision and recall. Thereby, this metric represents both precision and recall by a single score.
Given the continuous changes and updates in eSports games, the strong connection between the features and target match outcome can remarkably limit model usage lifespan. For this reason, in this study, prediction models have evolved from season to season. Our approach, which builds the models on season-based data, is practically feasible since the models reflect game mechanics changes.
We conducted four experiments for demonstrating the efficiency of the proposed MOMIL approach.
  • Experiment 1—To show the superiority of MOMIL, we compared the multi-instance classification methods with their standard (single-instance) counterparts.
  • Experiment 2—We determined the best MIL algorithm by comparing alternative ones.
  • Experiment 3—We identified the most important factors that affect the victory in the matches in LoL.
  • Experiment 4—We compared our results with the results presented in the state-of-the-art studies [13,21,22,23,24] on the same dataset.

4.1. Dataset Description

To show the efficiency of MOMIL, the experiments were carried out on an LoL dataset publicly available in the Kaggle data repository (https://www.kaggle.com/paololol/league-of-legends-ranked-matches (accessed 21 November 2022 )). The dataset contains statistical match information of 1,834,520 players from the regions of North America and Europe. The raw data contain several tables with information about matches, players, champions, items purchased, team bans, and statistics. We combined the tables by using the join operation. After that, we extracted 40 features that captured relevant information to build a win-prediction model, including kills, assists, deaths, damages, rewards, and gold earned by each player. The objective was to predict the outcome of an LoL match based on team performances in the previous games.
In the data pre-processing stage, the complete data set was divided into subsets for season-based analysis (seasons 3, 4, 5, 6, and 7) since each season of the game has many changes that affect the winning strategy. The game is normally played with 10 players; for this reason, we removed the match records that contained fewer players. Although the raw dataset is not provided as multi-instance data, we transformed it for this purpose. The individual player statistics were taken into account as single-instance data, while team-level features were considered as multi-instance data. For winner prediction, the multi-instance method successfully maps input vectors representing game statistics of the players in the teams to output labels (win or loss). The winner can then be estimated for given query vectors of the players in a team.

4.2. Comparison of Single-Instance and Multi-Instance Classification

In the first experiment, to show the superiority of the proposed MOMIL approach, we compared the multi-instance learning methods with their standard (single-instance) counterparts, including MIBoost vs. AdaBoost, MISVM vs. SVM, MILR vs. LR, SimpleMI vs. DT, MIWrapper vs. DT, CitationKNN vs. KNN, and TLC vs. LogitBoost. It should be noted here that we compared both MIWrapper and SimpleMI with DT, because they apply a standard learner to multi-instance data, but in different ways, and in this study, we selected DT as the base learner for both of them. For this reason, we took into consideration MIWrapper vs. DT comparison, as well as SimpleMI vs. DT comparison.
Table 2 shows the results for each season separately in terms of accuracy (%). From the experimental results, it is clearly seen that multi-instance (MI) learning algorithms are better than their single-instance (SI) versions for all seasons. For example, MISVM (93.28%) achieved higher accuracy than SVM (88.99%) for season 3. On average, the multi-instance methods (MIBoost, MISVM, MILR, SimpleMI, MIWrapper, CitationKNN, and TLC) improved the classification accuracy by approximately 9.5%, 4.5%, 6%, 13.5%, 7.5%, 0.3%, and 11.3% compared to the single-instance methods (AdaBoost, SVM, LR, DT, KNN, and LogiBoost), respectively. Thereby, the results indicated that the proposed MOMIL approach could construct a robust model with high prediction accuracy. This is because it would make sense to see better performance in classification by applying multi-instance learning since cooperation among team members in LoL is crucial for success. In other words, MOMIL benefits from the collaboration property of multi-instance data and increases the performance of the predictive model.
One of the key properties of the MOMIL method is that it can be applied to any multi-instance dataset without having background information on the game. It is seen from Table 2 that the classification accuracy values of MI algorithms change from season to season. For example, MIBoost is more likely to correctly predict LoL match outcomes for season 7 (95.88%) than for season 4 (89.68%). Therefore, it experimentally confirmed that the performance of the multi-instance algorithm can be affected by the characteristics of seasons.
Although the MI algorithms have higher accuracy than their SI versions, the results were evaluated by using a statistical test to show that the differences in classification performances are statistically significant. The p-values for each season were calculated by using the Mann–Whitney-U test. According to the calculated p-values (0.0297, 0.0405, 0.0213) for seasons (3,5,7), 4, and 6, respectively, it can be concluded that the performance results are statistically significant because all the p-values are under the significance threshold level ( α = 0.05).
Figure 3 shows the comparative results for MI and SI algorithms in terms of F-measure. According to the results, the MI methods achieved a higher F-measure value compared to the SI methods. The F-measure value ranges between 0 and 1, where 1 is the best value. In other words, the higher F-measure value, the better the classification performance.
As can be seen from Figure 3, the F-measure values obtained by the MI algorithms are closer to 1 than the SI algorithms for all seasons. The F-measure value difference between MI and SI is also remarkable for almost all algorithms, except CitationKNN. For instance, MILR (0.951) is significantly better than LR (0.882) for season 7. MI outperformed SI by increasing the F-measure by at least 0.008 points and at most 0.147 points on the seasons. One possible explanation for this improvement is that the MI methods analyze data at the team level, rather than only at the player level, since team cooperation is critical for match success. In other words, MI jointly considers the players in a team during the classification, since all five players are relevant for the team’s success and collaboration among players on each team significantly affects which team will win or lose. As a result, the proposed MOMIL approach can be successfully used for win prediction.

4.3. Comparison of Multi-Instance Classification Algorithms

In the second experiment, we determined the best MIL method by comparing alternative ones. Figure 4 shows the average results in terms of accuracy (%). Based on the results, it can be noted that the SimpleMI method achieved the highest accuracy (94.01%) in terms of the seasonal average. The MILR and MIBoost methods followed it with accuracy values of 93.94% and 93.67%, respectively. However, the CitationKNN method performed poorly in comparison to the other methods. This is probably because of the fact that the closest references are likely similar to the closest citations and utilizing these did not improve prediction much.

4.4. Factor Analysis

In the third experiment, we identified the most important factors that affect the victory of matches in LoL. Our aim is not only to build a well-performing win-prediction classifier but also to describe the prediction by means of feature importance. We used the OneR algorithms to evaluate the features since they rank features according to the minimum-error rate on the training set. Through pair-wise comparison, it estimates how each independent feature is correlated with the class attribute. In decreasing order, Figure 5 shows the importance of features for the win prediction in LoL matches. As can be seen, feature ranks range between 50.13 and 68.19. According to the results, the most important factors for win prediction are the number of turret kills, deaths per min, and assists per minute. The largest killing spree follows them as one of the important features. After killing-based features, gold earned by players is also ranked among the top 10 features.

4.5. Comparison of Our Study with the State-of-the-Art Studies

We compared our results with the results presented in the previous studies [13,21,22,23,24] on the same datasets. As shown in Table 3, the proposed approach outperforms the previous methods presented in [13,21,22,23,24]. For example, on average, MOMIL achieved significantly higher accuracy than the method in [13], which also focuses on win prediction.
In previous studies, deep learning techniques have also been applied to different eSports datasets. The results obtained in their studies can be summarized as follows. Lan et al. [39] proposed a CNN+RNN model for predicting win–loss outcomes and achieved 87.85% accuracy. Gu et al. [40] applied DNN to a different LoL dataset and obtained an accuracy of 62.09%. Similarly, Do et al. [52] also used DNN and reported that game outcomes can be predicted with 75.1% accuracy. Kim and Lee [53] proposed a deep learning model based on a bidirectional LSTM and reported 58.07% accuracy for win–loss prediction in LoL.

5. Conclusions and Future Work

The objective of this study was to successfully predict the outcome of an eSports match by using machine learning methods. For this purpose, we proposed a new approach, called MOMIL, which considers the task of analyzing data not only to handle the win prediction probability of individual players but also to explore the winning probability of a team as a whole. In particular, each player is related to all the teammates; therefore, our approach makes a team-level analysis, where the feature vectors of the players in a team are considered together.
In the experiments, our approach was applied to a publicly available LoL dataset that has a variety of input features that represent different respects of matches, including kills, assists, deaths, damages, rewards, and gold earned by each player. We built the models on season-based data since each season of the game has many changes that affect the win-prediction model.
The main findings from this research can be summarized as follows:
  • The experimental results showed that the multi-instance-based classification approach outperformed the standard classification approach for winner prediction in terms of accuracy and F-measure.
  • The proposed MOMIL approach achieved up to 95% accuracy for match outcome prediction in LoL seasons.
  • Among multi-instance learning algorithms, the SimpleMI method achieved the highest accuracy in terms of the seasonal average. The MILR and MIBoost methods followed it.
  • The most important factors for win prediction are the number of turret kills, deaths per min, and assists per minute.
Our study especially investigated the winning predictors in League of Legends. However, as future work, it can also be applied to similar MOBA games such as Dota 2 and Counter-Strike.

Author Contributions

Conceptualization, K.U.B.; methodology, K.U.B.; software, D.B.; validation, D.B.; formal analysis, D.B.; investigation, K.U.B.; resources, K.U.B.; data curation, K.U.B.; writing—original draft preparation, D.B.; writing—review and editing, K.U.B.; visualization, K.U.B. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The LoL dataset is publicly available at the following website: https://www.kaggle.com/datasets/paololol/league-of-legends-ranked-matches (accessed on 20 November 2022).

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Reitman, J.G.; Anderson-Coto, M.J.; Wu, M.; Lee, J.S.; Steinkuehler, C. Esports research: A literature review. Games Cult. 2020, 15, 32–50. [Google Scholar] [CrossRef]
  2. Schubert, M.; Drachen, A.; Mahlmann, T. Esports analytics through encounter detection. In Proceedings of the MIT Sloan Sports Analytics Conference, Boston, MA, USA, 11 March 2016; pp. 1–18. [Google Scholar]
  3. Saiz-Alvarez, J.M.; Palma-Ruiz, J.M.; Valles-Baca, H.G.; Fierro-Ramirez, L.A. Knowledge Management in the Esports Industry: Sustainability, Continuity, and Achievement of Competitive Results. Sustainability 2021, 13, 10890. [Google Scholar] [CrossRef]
  4. Zhang, Y.; Liu, R. Economic Sources behind the Esports Industry. In Proceedings of the 7th International Conference on Financial Innovation and Economic Development, Harbin, China, 14–16 January 2022; pp. 643–648. [Google Scholar]
  5. Keiper, M.C.; Manning, R.D.; Jenny, S.; Olrich, T.; Croft, C. No reason to LoL at LoL: The addition of esports to intercollegiate athletic departments. J. Study Sports Athletes Edu. 2017, 11, 143–160. [Google Scholar] [CrossRef]
  6. Hodge, V.J.; Devlin, S.; Sephton, N.; Block, F.; Cowling, P.; Drachen, A. Win prediction in multi-player esports: Live professional match prediction. IEEE Trans. Games 2021, 13, 368–379. [Google Scholar] [CrossRef] [Green Version]
  7. Xia, B.; Wang, H.; Zhou, R. What contributes to success in MOBA games? An empirical study of Defense of the Ancients 2. Games Cult. 2019, 14, 498–522. [Google Scholar] [CrossRef]
  8. Smerdov, A.; Somov, A.; Burnaev, E.; Zhou, B.; Lokuwicz, P. Detecting video game player burnout with the use of sensor data and machine learning. IEEE Internet Things J. 2021, 8, 16680–16691. [Google Scholar] [CrossRef]
  9. Maymin, P.Z. Smart kills and worthless deaths: ESports analytics for League of Legends. J. Quant. Anal. Sports 2021, 17, 11–27. [Google Scholar] [CrossRef]
  10. Viggiato, M.; Bezemer, C.-P. Trouncing in Dota 2: An investigation of blowout matches. In Proceedings of the 16th AAAI Conference on Artificial Intelligence and Interactive Digital Entertainment, Worcester, MA, USA, 19–23 October 2020; pp. 294–300. [Google Scholar]
  11. Kho, L.C.; Kasihmuddin, M.S.M.; Mansor, M.A.; Sathasivam, S. Logic mining in League of Legends. Pertanika J. Sci. Technol. 2020, 28, 211–225. [Google Scholar]
  12. Kang, S.-K.; Lee, J.-H. An e-sports video highlight generator using win-loss probability model. In Proceedings of the 35th Annual ACM Symposium on Applied Computing, Brno, Czech Republic, 30 March–3 April 2020; pp. 915–922. [Google Scholar]
  13. Cardoso, G.M.M. Predicao do Resultado Utilizando KNN: Analise do Jogo League of Legends. Master’s Thesis, Instituto Federal de Educacao, Bambui, Brazil, 2019. [Google Scholar]
  14. Clark, N.; Macdonald, B.; Kloo, I. A Bayesian adjusted plus-minus analysis for the esport Dota 2. J. Quant. Anal. Sports 2020, 16, 325–341. [Google Scholar] [CrossRef]
  15. Sanchez-Ruiz, A.A.; Miranda, A. A machine learning approach to predict the winner in StarCraft based on influence maps. Entertain. Comput. 2017, 19, 29–41. [Google Scholar] [CrossRef]
  16. Smerdov, A.; Zhou, B.; Lukowicz, P.; Somov, A. Collection and validation of psycophysiological data from professional and amateur players: A multimodal esports dataset. arXiv 2021, arXiv:2011.00958. [Google Scholar]
  17. Khromov, N.; Korotin, A.; Lange, A.; Stepanov, A.; Burnaev, E.; Somov, A. Esports athletes and players: A comparative study. IEEE Pervasive Comput. 2019, 18, 31–39. [Google Scholar] [CrossRef]
  18. Liu, J.X.; Huang, J.X.; Chen, R.Y.; Liu, T.; Zhou, L. A two-stage real-time prediction method for multiplayer shooting e-sports. In Proceedings of the 20th International Conference on Electronic Business, Hong Kong, China, 5–8 December 2020; pp. 9–18. [Google Scholar]
  19. Katona, A.; Spick, R.; Hodge, V.J.; Demediuk, S.; Block, F.; Drachen, A.; Walker, J.A. Time to die: Death prediction in Dota 2 using deep learning. In Proceedings of the IEEE Conference on Games, London, UK, 20–23 August 2019; pp. 1–8. [Google Scholar]
  20. Carbonneau, M.-A.; Cheplygina, V.; Grangera, E.; Gagnon, G. Multiple instance learning: A survey of problem characteristics and applications. Pattern Recognit. 2018, 77, 329–353. [Google Scholar] [CrossRef] [Green Version]
  21. Li, S.; Duan, L.; Zhang, W.; Wang, W. Multi-attribute context-aware item recommendation method based on deep learning. In Proceedings of the 5th International Conference on Pattern Recognition and Artificial Intelligence, Chengdu, China, 19–21 August 2022; pp. 12–19. [Google Scholar]
  22. Duan, L.; Li, S.; Zhang, W.; Wang, W. MOBA game item recommendation via relationaware graph attention network. In Proceedings of the IEEE Conference on Games (CoG), Beijing, China, 21–24 August 2022; pp. 338–344. [Google Scholar]
  23. Villa, A.; Araujo, V.; Cattan, F.; Parra, D. Interpretable contextual team aware item recommendation application in multiplayer online battle arena games. In Proceedings of the 14th ACM Conference on Recommender Systems, Rio de Janeiro, Brazil, 22 September 2020; pp. 503–508. [Google Scholar]
  24. Araujo, V.; Rios, F.; Parra, D. Data mining for item recommendation in MOBA games. In Proceedings of the 13th ACM Conference on Recommender Systems, Copenhagen, Denmark, 16–20 September 2019; pp. 393–397. [Google Scholar]
  25. Ke, C.H.; Deng, H.; Xu, C.; Li, J.; Gu, X.; Yadamsuren, B.; Klabjan, D.; Sifa, R.; Drachen, A.; Demediuk, S. DOTA 2 match prediction through deep learning team fight models. In Proceedings of the IEEE Conference on Games, Beijing, China, 21–24 August 2022; pp. 96–103. [Google Scholar]
  26. Hitar-Garcia, J.-A.; Moran-Fernandez, L.; Bolon-Canedo, V. Machine Learning Methods for Predicting League of Legends Game Outcome. IEEE Trans. Games 2023, 1, 1–11. [Google Scholar] [CrossRef]
  27. Ghazali, N.F.; Sanat, N.; Asari, M.A. Esports analytics on playerunknown’s battlegrounds player placement prediction using machine learning approach. Int. J. Hum. Technol. Interact. 2021, 5, 1–10. [Google Scholar]
  28. Dikananda, A.R.; Ali, I.; Rinaldi, R.A. Genre e-sport gaming tournament classification using machine learning technique based on decision tree, naive Bayes, and random forest algorithm. In Proceedings of the Annual Conference on Computer Science and Engineering Technology, Medan, Indonesia, 23 September 2020; pp. 1–8. [Google Scholar]
  29. Demediuk, S.; York, P.; Drachen, A.; Walker, J.A.; Block, F. Role identification for accurate analysis in Dota 2. In Proceedings of the 15th AAAI Conference on Artificial Intelligence and Interactive Digital Entertainment, Atlanta, Georgia, 8–12 October 2019; pp. 130–138. [Google Scholar]
  30. Mora-Cantallops, M.; Sicilia, M.-A. Player-centric networks in League of Legends. Soc. Netw. 2018, 55, 149–159. [Google Scholar] [CrossRef]
  31. Kadan, A.M.; Li, L.; Chen, T. Modeling and analysis of features of team play strategies in eSports applications. Mod. Inf. Technol. IT-Educ. 2018, 14, 397–407. [Google Scholar]
  32. Demediuk, S.; Murrin, A.; Bulger, D.; Hitchens, M.; Drachen, A.; Raffe, W.L.; Tamassia, M. Player retention in League of Legends: A study using survival analysis. In Proceedings of the Australasian Computer Science Week Multiconference, New York, NY, USA, 29 January–2 February 2018; pp. 1–9. [Google Scholar]
  33. Yao, W.; Wang, Y.; Zhu, M.; Cao, Y.; Zeng, D. Goal or Miss? A Bernoulli Distribution for In-Game Outcome Prediction in Soccer. Entropy 2022, 24, 971. [Google Scholar] [CrossRef]
  34. Li, S.-F.; Huang, M.-L.; Li, Y.-Z. Exploring and Selecting Features to Predict the Next Outcomes of MLB Games. Entropy 2022, 24, 288. [Google Scholar] [CrossRef]
  35. Xiang, Y.; Chen, Q.; Wang, X.; Qin, Y. Distant Supervision for Relation Extraction with Ranking-Based Methods. Entropy 2016, 18, 204. [Google Scholar] [CrossRef] [Green Version]
  36. Zhu, Z.; Zhao, Y. Multi-Graph Multi-Label Learning Based on Entropy. Entropy 2018, 20, 245. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  37. Deng, W.; Ni, H.; Liu, Y.; Chen, H.; Zhao, H. An adaptive differential evolution algorithm based on belief space and generalized opposition-based learning for resource allocation. Appl. Soft Comput. 2022, 127, 109419. [Google Scholar] [CrossRef]
  38. Deng, W.; Zhang, L.; Zhou, X.; Zhou, Y.; Sun, Y.; Zhu, W.; Chen, H.; Deng, W.; Chen, H.; Zhao, H. Multi-strategy particle swarm and ant colony hybrid optimization for airport taxiway planning problem. Inf. Sci. 2022, 612, 576–593. [Google Scholar] [CrossRef]
  39. Lan, X.; Duan, L.; Chen, W.; Qin, R.; Nummenmaa, T.; Nummenmaa, J. A Player Behavior Model for Predicting Win-Loss Outcome in MOBA Games. Proceedings of 14th International Conference on Advanced Data Mining and Applications, Nanjing, China, 16–18 November 2018; pp. 474–488. [Google Scholar]
  40. Gu, Y.; Liu, Q.; Zhang, K.; Huang, Z.; Wu, R.; Tao, J. NeuralAC: Learning Cooperation and Competition Effects for Match Outcome Prediction. In Proceedings of the 35th AAAI Conference on Artificial Intelligence, Palo Alto, CA, USA, 2–9 February 2021; pp. 4072–4080. [Google Scholar]
  41. Xu, X.; Frank, E. Logistic regression and boosting for labeled bags of instances. In Proceedings of the Pacific-Asia Conference on Knowledge Discovery and Data Mining, Sydney, Australia, 26–28 May 2004; pp. 272–281. [Google Scholar]
  42. Andrews, S.; Tsochantaridis, I.; Hofmann, T. Support vector machines for multiple-instance learning. In Proceedings of the 15th International Conference on Neural Information Processing Systems, Cambridge, MA, USA, 9–14 December 2002; pp. 577–584. [Google Scholar]
  43. Ray, S.; Craven, M. Supervised versus multiple instance learning: An empirical comparison. In Proceedings of the 22nd International Conference on Machine Learning, New York, NY, USA, 7–11 August 2005; pp. 697–704. [Google Scholar]
  44. Blockeel, H.; Page, D.; Srinivasan, A. Multi-instance tree learning. In Proceedings of the 22nd International Conference on Machine Learning, New York, NY, USA, 7 August 2005; pp. 57–64. [Google Scholar]
  45. Bjerring, L.; Frank, E. Beyond trees: Adopting MITI to learn rules and ensemble classifiers for multi-instance data. In Proceedings of the Australasian Joint Conference on Artificial Intelligence, Perth, Australia, 5–8 December 2011; pp. 41–50. [Google Scholar]
  46. Dong, L. A comparison of multi-instance learning algorithms. Master’s Thesis, The University of Waikato, Hamilton, New Zealand, 2006. [Google Scholar]
  47. Frank, E.; Xu, X. Applying Propositional Learning Algorithms to Multi-Instance Data; Technical Report; University of Waikato: Hamilton, New Zealand, 2003. [Google Scholar]
  48. Wang, J.; Zucker, J. Solving the multiple-instance problem: A lazy learning approach. In Proceedings of the International Conference on Machine Learning, San Francisco, CA, USA, 29 June–2 July 2000; pp. 1119–1126. [Google Scholar]
  49. Weidmann, N.; Frank, E.; Pfahringer, B. A two-level learning method for generalized multi-instance problems. In Proceedings of the European Conference on Machine Learning, Dubrovnik, Croatia, 22–26 September 2003; pp. 468–479. [Google Scholar]
  50. Witten, I.H.; Frank, E.; Hall, M.A. Data Mining: Practical Machine Learning Tools and Techniques; Morgan Kaufmann: Cambridge, MA, USA, 2016. [Google Scholar]
  51. Yu, Y.; Hao, Z.; Li, G.; Liu, Y.; Yang, R.; Liu, H. Optimal search mapping among sensors in heterogeneous smart homes. Math. Biosci. Eng. 2022, 20, 1960–1980. [Google Scholar] [CrossRef]
  52. Do, T.D.; Wang, S.I.; Yu, D.S.; McMillian, M.G.; McMahan, R.P. Using Machine Learning to Predict Game Outcomes Based on Player-Champion Experience in League of Legends. In Proceedings of the 16th International Conference on the Foundations of Digital GamesAugust, Montreal, QC, Canada, 3–6 August 2021; pp. 1–5. [Google Scholar]
  53. Kim, C.; Lee, S. Predicting Win-Loss of League of Legends Using Bidirectional LSTM Embedding. KIPS Trans. Software Data Eng. 2020, 9, 61–68. [Google Scholar]
Figure 1. Standard classification versus multi-instance classification.
Figure 1. Standard classification versus multi-instance classification.
Entropy 25 00028 g001
Figure 2. The general overview of the proposed MOMIL approach.
Figure 2. The general overview of the proposed MOMIL approach.
Entropy 25 00028 g002
Figure 3. Comparison of single-instance (SI) and multi-instance (MI) classification methods in terms of F-measure.
Figure 3. Comparison of single-instance (SI) and multi-instance (MI) classification methods in terms of F-measure.
Entropy 25 00028 g003
Figure 4. Comparison of multi-instance classification algorithms.
Figure 4. Comparison of multi-instance classification algorithms.
Entropy 25 00028 g004
Figure 5. The importance of features for win prediction.
Figure 5. The importance of features for win prediction.
Entropy 25 00028 g005
Table 1. Comparison of our study with the previous studies.
Table 1. Comparison of our study with the previous studies.
Ref.YearDescriptionTaskAlgorithmGameMILSeason
CFRCL LoLDota2Other Based
[26]2023Predicting match outcome SVM, NB, KNN, NN XX
[25]2022Predicting match outcome LSTM XX
[8]2021Predicting whether the player will win or lose LR, KNN, NN XX
[9]2021Predicting in-game win probability LR XX
[27]2021Predicting the placement of players DT, LR, SVM PUBGXX
[28]2021Classifying eSports games DT, NB, RFCSXX
[14]2020Predicting match outcome BHR XX
[11]2020Modelling the results (win/lose) of the game NN XX
[16]2020Predicting the player skill and re-identification LR, KNN, SVM, RF, NB XX
[12]2020Predicting match outcome CNN, NN, CART XX
[18]2020Predicting the players’ rankings RF, GBM PUBGXX
[10]2020Predicting match outcome GBM, RF, LR XX
[23]2020Recommending items DT, LR, NN, CNN XX
[13]2019Predicting match outcome KNN XX
[29]2019Identifying roles K-Means XX
[17]2019Predicting the player skill ERT CSXX
[7]2019Predicting match outcome CART, LR XX
[6]2019Predicting match outcome LR, RF, GBM XX
[19]2019Predicting a player will die in the next 5 s DNN XX
[24]2019Recommending items Apriori, Eclat, DT, LR, NN XX
[30]2018Clustering player-centric networks Affinity Propagation XX
[31]2018Predicting round results DT, RF, KNN, PCA CSXX
[32]2018Predicting player churn MECR XX
[15]2017Predicting match outcome LDA, QDA, SVM, KNN, Weighted KNN Star CraftXX
ProposedPredicting match outcome MIBoost, MISVM, MILR, MIRI, MITI, SimpleMI, MIWrapper, CitationKNN, TLC
Table 2. Comparison of single-instance (SI) and multi-instance (MI) classification methods for each season in terms of accuracy (%).
Table 2. Comparison of single-instance (SI) and multi-instance (MI) classification methods for each season in terms of accuracy (%).
MIBoostAdaBoostMISVMSVMMILRLRSimpleMIDT
Season 395.6986.0993.2888.9995.8889.3294.4581.98
Season 489.6881.9889.2885.5891.2986.0391.2978.18
Season 595.2385.1693.4588.2295.3988.1994.7482.16
Season 691.8782.8791.3787.4192.0788.0194.3879.74
Season 795.8885.0693.0387.9295.0688.1795.2181.14
MIWrapperDTCitation
KNN
KNNTLCLogit
Boost
Season 389.2681.9879.5378.9293.6381.75
Season 484.5878.1875.8774.9188.7480.27
Season 589.4382.1678.2778.7193.9381.50
Season 688.5579.7477.5176.3793.6781.61
Season 789.0681.1477.7678.7093.1681.49
Table 3. Comparison of our study with state-of-the-art studies.
Table 3. Comparison of our study with state-of-the-art studies.
Ref.YearAlgorithmExisting
Method
Proposed
Method
F-MeasureF-Measure
[21]2022Multi-attribute Context-aware Item Rec. (MCIR)0.6160.903
F-MeasureF-Measure
[22]2022Relation-aware Graph Attention Network0.6260.903
F-MeasureF-Measure
[23]2020Decision Tree
Logistic Regression
Artificial Neural Network
Convolutional Neural Networks
Team-aware Transformer-based Item Rec. (TTIR)
0.379
0.468
0.566
0.586
0.596
0.903
F-MeasureF-Measure
[24]2019Apriori
Equivalence Class Transform.
Decision Tree
Logistic Regression
Artificial Neural Network
0.48
0.48
0.38
0.43
0.53
0.903
AccuracyAccuracy
[13]2019K-Nearest Neighbors58.00%90.23%
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Birant, K.U.; Birant, D. Multi-Objective Multi-Instance Learning: A New Approach to Machine Learning for eSports. Entropy 2023, 25, 28. https://doi.org/10.3390/e25010028

AMA Style

Birant KU, Birant D. Multi-Objective Multi-Instance Learning: A New Approach to Machine Learning for eSports. Entropy. 2023; 25(1):28. https://doi.org/10.3390/e25010028

Chicago/Turabian Style

Birant, Kokten Ulas, and Derya Birant. 2023. "Multi-Objective Multi-Instance Learning: A New Approach to Machine Learning for eSports" Entropy 25, no. 1: 28. https://doi.org/10.3390/e25010028

APA Style

Birant, K. U., & Birant, D. (2023). Multi-Objective Multi-Instance Learning: A New Approach to Machine Learning for eSports. Entropy, 25(1), 28. https://doi.org/10.3390/e25010028

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop