Next Article in Journal
Quantum and Quantum-Inspired Stereographic K Nearest-Neighbour Clustering
Previous Article in Journal
A New Truncated Lindley-Generated Family of Distributions: Properties, Regression Analysis, and Applications
Previous Article in Special Issue
Analysis of Weather Factors on Aircraft Cancellation using a Multilayer Complex Network
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Community-Based Matrix Factorization (CBMF) Approach for Enhancing Quality of Recommendations

1
Algorithms and Complexity Theory Lab, Department of Computer Science and Engineering, SRM University-AP, Amaravati 522502, India
2
Department of Computing, Sheffield Hallam University, Howard Street, Sheffield S1 1WB, UK
*
Author to whom correspondence should be addressed.
Entropy 2023, 25(9), 1360; https://doi.org/10.3390/e25091360
Submission received: 1 August 2023 / Revised: 8 September 2023 / Accepted: 13 September 2023 / Published: 20 September 2023
(This article belongs to the Special Issue Foundations of Network Analysis)

Abstract

:
Matrix factorization is a long-established method employed for analyzing and extracting valuable insight recommendations from complex networks containing user ratings. The execution time and computational resources demanded by these algorithms pose limitations when confronted with large datasets. Community detection algorithms play a crucial role in identifying groups and communities within intricate networks. To overcome the challenge of extensive computing resources with matrix factorization techniques, we present a novel framework that utilizes the inherent community information of the rating network. Our proposed approach, named Community-Based Matrix Factorization (CBMF), has the following steps: (1) Model the rating network as a complex bipartite network. (2) Divide the network into communities. (3) Extract the rating matrices pertaining only to those communities and apply MF on these matrices in parallel. (4) Merge the predicted rating matrices belonging to communities and evaluate the root mean square error (RMSE). In our experimentation, we use basic MF, SVD++, and FANMF for matrix factorization, and the Louvain algorithm is used for community division. The experimental evaluation on six datasets shows that the proposed CBMF enhances the quality of recommendations in each case. In the MovieLens 100K dataset, RMSE has been reduced to 0.21 from 1.26 using SVD++ by dividing the network into 25 communities. A similar reduction in RMSE is observed for the datasets of FilmTrust, Jester, Wikilens, Good Books, and Cell Phone.

1. Introduction

To predict recommendations to the users based on past behavior and their preferences, a technique in recommender systems named collaborative filtering is used [1]. The main objective is to suggest items or content to the users by considering their interactions or resemblances with other users. Due to the continuous growth in data availability and the interconnected nature of diverse systems, these techniques hold immense importance in revealing patterns, relationships, and significant insights [2].
Matrix factorization (MF) is a technique utilized in collaborative filtering to decompose a matrix of user-item ratings into lower-rank matrices capturing the latent factors underlying the data [3,4]. The user-item rating matrix serves as a representation of user ratings assigned to different items [5]. The matrix is frequently sparse as users typically only rate a small subset of items. MF techniques strive to complete the missing entries in the matrix by decomposing it into lower-rank matrices, one capturing the user’s underlying preferences and the other reflecting the item’s latent characteristics [6]. The latent representations of users along with items can be used to estimate future ratings or calculate missing ratings once the matrix has been factorized.
In network analysis, community detection stands as a fundamental task, endeavoring to identify cohesive subsets of nodes, referred to as communities or modules within a network [7]. These communities in networks signify groups of nodes that display stronger interconnections amongst themselves compared to connections with nodes outside the community [8]. These communities provide insightful information on the structure, organization, and dynamics of complex systems [9,10]. The applications of community detection across various domains, such as social network analysis, biological networks, online forums, and recommendation systems, have garnered substantial attention [11]. Through the identification of communities, we gain insights into the underlying interaction patterns, discover influential groups, and develop a deeper comprehension of the network’s functionality [12].

1.1. Applications

Matrix factorization techniques have been widely applied across diverse domains, demonstrating their versatility and effectiveness. Some of the notable applications of the matrix factorization techniques include
1.
Natural Language Processing (NLP): Within the field of NLP, techniques for matrix factorization have been used in a variety of tasks in topic modeling, text classification, etc. [13,14]. Through the decomposition of the document matrix, MF algorithms can reveal latent representations that effectively capture the underlying semantic structure of the textual data.
2.
Social Network Analysis: MF techniques have been utilized in social network analysis to unveil communities of individuals sharing common interests or behaviors [15,16]. SVD++ facilitates the identification of friends, influencers, or interest-based communities by enhancing user experience and engagement on social media platforms. FANMF is used to uncover community structures and identify influential nodes within the network. FANMF can effectively detect the group of nodes by exposing hidden relationships and structures by factorizing them into nonnegative adjacency matrices.
3.
E-commerce: In the realm of e-commerce platforms, SVD++ is extensively utilized to deliver personalized product recommendations to users [17,18]. By including implicit feedback, supplementary data, and user-item ratings, SVD++ can adeptly capture user preferences and item characteristics for accurate product recommendations [19].
4.
Streaming Services: Streaming platforms, including music or video services or SVD++, provide personalized content recommendations to the users. SVD++ significantly enhances the discovery and recommendation of relevant and captivating content. It ensures that the users are presented with content aligned with their individual tastes and preferences [20].
5.
Image Processing: FANMF finds application in image processing tasks, where the image data are factored into nonnegative matrices. By extraction, it improves image quality and facilitates the analysis of visual data. By decomposition, matrix factorization algorithms are capable of distinguishing noise from the underlying structure. It also completes missing parts and extracts significant features for analysis and representation [21].
6.
Nutritional Recommendation: In the realm of nutritional recommendations, matrix factorization entails structuring dietary information into a user-item matrix. This matrix uncovers hidden factors linked to individual tastes and nutritional traits, facilitating the delivery of personalized dietary advice. By accounting for variables such as taste preferences, dietary constraints, and health objectives, this approach aids individuals in devising well-balanced diets, promoting healthier and custom-tailored eating habits [22].
In essence, the applications of MF techniques have a broad scope of transforming the methods through which we analyze, comprehend, and leverage complex data [23]. The ongoing evolution of these techniques holds the potential to unlock fresh possibilities and propel advancements in numerous domains. It ultimately benefits individuals, organizations, and society at large.

1.2. Problem Statement

Schafer et al. [24] describe the problem of recommender systems as follows:
In the context of a set of users, denoted as U = { u 1 , u 2 , , u a } , a set of items or products, denoted as I = { I 1 , I 2 , , I b } , and the corresponding user-item ratings, where R i j represents the rating assigned by user u i to item I j , the primary objective of recommender systems is to suggest a new item I j to a user u i that the user has not yet encountered or consumed.
As an example, let us take a look at the matrix depicted in Figure 1. In this matrix, each row corresponds to a user, and each column represents an item, such as books. The specific value R i j within the matrix indicates the ratings provided by user u i for item I j . The main objective of recommender systems lies in forecasting the values within the rating matrix that have not yet been assigned or rated.
Content-based recommendation and collaborative filtering represent two widely adopted approaches for solving the problem of recommender systems. Matrix factorization (MF), is a prevalent model-based collaborative filtering technique. The main challenge with MF is its computational intensity. The objective of this study is to enhance the recommendation quality while concurrently reducing the time complexity associated with MF techniques. To achieve this goal, we propose representing the rating matrix as a complex network. Complex networks inherently manifest a community structure. We aim to leverage the community information in order to parallelize the matrix factorization technique.
Recommender systems are natural instances of weighted complex bipartite networks. A bipartite network is characterized by having exactly two distinct node types and the presence of edges connecting nodes of different types. More formally, a bipartite network is defined as G = ( V 1 V 2 , E ) , where V 1 and V 2 denote two sets of node types, and E represents the set of edges connecting nodes from V 1 to nodes from V 2 . You can observe an illustrative example of this in Figure 2, which represents a scenario where users make purchases on e-commerce platforms like Amazon, portraying the interaction as a bipartite network.
We propose that the community information available in the network can be successfully incorporated into MF techniques. This approach not only enhances the quality of recommendations but also provides a parallel framework based on community division to the MF technique, which addresses the problem of computational complexity of MF. This approach requires MF techniques as well as community detection algorithms. The following section briefly reviews the two.

2. Literature Review

In recent years, the MF method has garnered significant attention as a widely adopted and successful method for rating prediction in recommendation systems. The Netflix Prize competition, which was started in 2006, is one early piece of work noteworthy in relation to MF in recommender systems [25]. The fundamental matrix factorization model creates user and item latent feature matrices from the rating matrix, enhancing the accuracy of rating predictions through the understanding of possible connections between users and items.
As the information interconnection era has emerged, the basic MF model no longer satisfies the demands of recommender systems. It thus leads to the emergence of numerous variants of this model. Excluding all the nonnegative entries in latent features, a model nonnegative matrix factorization (NMF) was initiated by Paatero and Tapper in 1994 that improves the accuracy of the model [26]. Mnih et al. in the year 2007 proposed probabilistic matrix factorization (PMF), which utilizes probabilistic modeling to effectively capture uncertainties in user-item ratings, resulting in recommendations that are more reliable and precise [27]. The significance of the singular value decomposition (SVD) technique was introduced by Mastorakis in 1857 and researchers started to apply SVD to the recommendation domain from 2006 [28]. The method has the ability to perform robust MF, enabling the identification of hidden features and effective dimensionality reduction in data analysis.
By integrating explicit and implicit feedback to enhance recommendation accuracy and personalization, Koren et al. in the year 2008 introduced an advanced singular value decomposition (SVD++) method [29]. Through the incorporation of user-item ratings and implicit feedback, SVD++ boosts the performance of the recommender systems, enabling better capture of user preferences and more effective recommendation generation. In 2015, Shi et al. introduced a method named pairwisely constrained nonnegative symmetric matrix factorization (PCSNMF) that incorporates the symmetric community structures found in undirected networks but also leverages pairwise constraints derived from ground-truth group information [30]. In 2017, deep matrix factorization (DMF) was introduced by Xiangnan et al. and the deep learning with MF [31] emerges. This method enables the discovery of intricate patterns and the extraction of complex features from large-scale datasets. Kipf and Welling in the year 2018 combined the power of MF and graph convolutional neural networks to capture both collaborative filtering patterns and graph structures in recommendation systems [32]. By incorporating the graph information, the recommendation accuracy is improved by leveraging the connectivity and relationships among users and items. In 2019, factorized asymmetric nonnegative matrix factorization (FANMF) was introduced by Tosyali et al. by considering the asymmetric relationships between users and items [33]. It led to enhanced recommendation quality, a deeper understanding of user preferences, and ultimately provided personalized user experiences in various data analysis tasks.
Community detection is another concept used by our proposed approach. The existing community detection algorithms are reviewed here. The Girvan–Neuman algorithm uses edge betweenness centrality to iteratively remove edges and identify communities in a network [34]. In 2007, Raghavan et al. developed a label propagation algorithm. The label propagation algorithm updates the node labels iteratively based on the majority of their neighbors’ community labels to effectively detect communities in a network  [35]. Modularity is one of the metrics used for evaluating the quality of community detection. It assesses the extent to which a network can be partitioned into discrete communities or modules by analyzing the connectivity patterns between nodes. The Louvain algorithm was proposed by Blondel et al. The Louvain algorithm has the ability to efficiently optimize modularity, making it highly effective in detecting communities in large-scale networks while maintaining computational speed [36]. In 2011, Pons and Latapy introduced the Walktrap algorithm, which is notable for its use of random walks to assess node similarities for locating communities in large networks [37]. The Leiden algorithm was developed in 2019 by Traag et al. with iterative application results in a partition when local best assignments are made to all community subsets  [38]. The following section briefly discusses the MF techniques and community detection algorithm chosen in our experimentation highlighting the cause of selection.

3. Methodology

In this work, we propose a parallel framework for matrix factorization. This section first provides an elaborate exposition elucidating the various matrix factorization methods, and then discusses the community detection algorithms.

3.1. Matrix Factorization

3.1.1. Basic Matrix Factorization

Simon Funk’s work during the Netflix Prize competition in 2006 played a significant role in popularizing the use of matrix factorization algorithms for recommender systems [25].
Given a rating matrix R of size a × b representing ratings given by a users on b items, where a huge number of ratings are unknown, the matrix factorization technique performs the following five steps.
  • Step 1: Initialize the entries of user latent feature matrix M and item latent feature matrix N of sizes a × k and b × k , respectively, with random values. k is the number of latent features, tuned experimentally with different values of k.
  • Step 2: Multiply the matrices M and N to obtain the predicted rating matrix with non-empty cells having some predicted ratings, as shown below in Equation (1).
    R ˜ = M N T
    r 11 ˜ . . r 1 b ˜ . . . . . . . . . . . . r a 1 ˜ . . r a b ˜ = m 11 . . m 1 k . . . . . . . . . . . . m a 1 . . m a k n 11 . . n 1 b . . . . . . . . . . . . n k 1 . . n k b
                                         a × b                 a × k             k × b
    Note that the given rating matrix R is approximately equal to M N T as shown in Equation (2).
    R M N T
  • Step 3: Compute the deviation between actual and predicted ratings as shown in Equation (3), where m a is of order m a × k and n b is of order n b × k .
    r a b m a n b T
  • Step 4: Minimize the error in the prediction. It is common to use Equation (4) to compute the squared error.
    m i n a , b ( r a b m a n b T ) 2
    In order to avoid overfitting the squared error, the regularization term is added as shown in Equation (5).
    m i n a , b ( r a b m a n b T ) 2 + α ( | | m a | | 2 + | | n b | | 2 )
    The impact of the regularization is controlled by a constant α . | | . | | is the frobenius norm. The approximation of this value is calculated using stochastic gradient descent or alternating least squares. For each rating in the training data, the prediction error is calculated using the stochastic gradient descent method as displayed in Equation (6).
    e a b = r a b m a n b T
  • Step 5: The following update rules shown in Equation (7) are used to update the matrices M and N to minimize squared error.
    n b n b + β ( e a b m a α n b ) m a m a + β ( e a b n b α m a )
    The representation of a matrix for the above Equation (7) are as follows:
    For the equation
    n b n b + β ( e a b m a α n b )
    the matrix representation is
    n 11 . . n 1 b . . . . . . . . . . . . n k 1 . . n k b n 11 . . n 1 b . . . . . . . . . . . . n k 1 . . n k b + β e a b m 11 . . m 1 k . . . . . . . . . . . . m a 1 . . m a k β α n 11 . . n 1 b . . . . . . . . . . . . n k 1 . . n k b
    For the equation
    m a m a + β ( e a b n b α m a )
    the matrix representation is
    m 11 . . m 1 k . . . . . . . . . . . . m k 1 . . m a k m 11 . . m 1 k . . . . . . . . . . . . m k 1 . . m a k + β e a b n 11 . . n 1 b . . . . . . . . . . . . n k 1 . . n k b β α m 11 . . m 1 k . . . . . . . . . . . . m a 1 . . m a k
  • The steps 3, 4, and 5 are repeated until either the number of iterations is fixed or the error reaches 0.
This approach has a time complexity of O ( a b k ) , where a , b are the number of users and items, and k is the number of latent features.

3.1.2. SVD++

SVD++ is an extension of the singular value decomposition method  [29].
In the SVD++ method, implicit feedback is added to the user’s latent feature vector [39,40]. The implicit feedback is computed in the form of two matrices, user feedback matrix P and item feedback matrix Q. The user feedback matrix P is a matrix of the same size as R. At first, the value of P = [ p a b ] ( u a , I b ) is set to be 1 if r a b is observed; it is 0 otherwise. Later, the entries of P are normalized row-wise as follows. Let I j be the item that the user u i has rated; each nonzero entry in the j t h row of P is computed as 1 | I j | . Matrix Q is the same as the item latent feature vector of order b × k . Once the feedback matrices P and Q are available, the dot product of P and Q matrices is computed and added to the user latent feature matrix. This user latent feature matrix computed is used in the procedure of SVD++ by adding an implicit feedback matrix. Then, the deviation between the original rating matrix R, i.e., ( m a n b ) , and the predicted rating matrix R ˜ , i.e., ( ( m + P Q ) a n b ), is computed as the RMSE value as shown in Equation (8).
The time complexity for the SVD++ method is O ( a b k ) , where a , b are the count of users and items, k is the number of latent features.

3.1.3. Factorized Asymmetric nonnegative Matrix Factorization (FANMF)

In 2019, the FANMF method was introduced by Tosyali [33]. The main purpose of the FANMF method is to handle nonnegative data that have an asymmetric nature. Data in the real world are often asymmetrical, where the rows and columns are related in this way and are not symmetrical [41]. In the FANMF method, a rating matrix is decomposed into two nonnegative latent feature matrices. FANMF is different from the traditional methods, as it accommodates the feature matrices of different dimensions that enable us to capture the data that are asymmetric in nature. It also adds a sparsity constraint along with the nonnegativity constraint, which contains only a limited number of nonzero elements.
FANMF extends NMF to asymmetric scenarios where the user and item bias are considered in addition to the user-item interactions to make the model more accurate and improve the recommendation quality. User item bias includes the inherent inclinations of users towards specific items or the inherent attractiveness of items to users, independent of their previous interactions or behaviors.
The original rating is represented as the product of user and item rating vectors as shown in Equation (3). Update the values of M and N by using a multiplicative update algorithm that contains update values [42].
M = M · × ( ( R · / ( M × N + ( R = = 0 ) ) ) × N T ) N = N · × ( M T × ( R · / ( M × N + ( R = = 0 ) ) ) )
where M · × N is the dot product of M and N similar to M · /N is the dot division of M and N, i.e., element wise division. M × N is the product of two matrices M and N. M T is the transpose of the matrix M. The term R = = 0 is included in the denominator to prevent division by zero.
Make the dot product of the updated latent feature matrices, which is R ˜ . The time complexity for the matrix factorization method is given as O ( a b k ) , where a , b are the number of users and items, and k is the number of features that were extracted.
After computing the predicted rating matrix using any of the methods specified above, the error in prediction is computed using root mean square error (RMSE) given in Equation (8).
R M S E = 1 N r a b r ˜ a b 2
where r a b is the original rating, r ˜ a b is the predicted rating, and N is the total number of predictions.
In this work, we introduce parallelism in matrix factorization with the help of community information. Various community detection algorithms are available in the literature to divide a complex network into communities.

3.2. Community Detection

The Girvan–Newman algorithm is one of the first popular algorithms that uses edge betweenness centrality to detect communities in networks, achieved through iterative removal of edges with high betweenness. The shortcoming of the above-mentioned algorithm is while it excels at identifying overlapping communities, it faces the drawback of being computationally expensive, particularly for large graphs, due to repeated edge removals and recalculations of betweenness. The Louvain algorithm addresses these shortcomings by adopting an efficient approach centered around optimizing modularity. Through iterative modularity optimization with node movements, it achieves fast and scalable community detection, effectively capturing communities even in large-scale networks. Louvain’s superior optimization of modularity enables it to outperform Girvan–Newman in terms of speed and scalability, making it the preferred choice for real-world community detection tasks involving massive and complex networks. Therefore, we use the Louvain algorithm to divide the rating network into communities. The next subsection gives the details of Louvain.

Louvain Algorithm

The Louvain algorithm works effectively in covering large communities or groups that are densely interconnected in a network. It employs a modularity optimization process, iteratively refining the network’s community structure to maximize modularity.
The Louvain algorithm has a six-step procedure to find the appropriate communities. Each node in the initial step is attached to its own community. Next, rearrange the nodes between the communities that iteratively optimize the modularity score, which elevates the standards of the community structure. By iteratively considering each node, evaluate the modularity score by relocating the nodes to nearby communities. It is followed until there is no change in the modularity score. A new graph is constructed by combining the nodes in each community by connecting the nodes and edges of different communities. All these steps are repeated until there is no change in the modularity score observed.
The Louvain algorithm has become popular because of its efficiency in detecting communities in the network by achieving a high modularity score, all while maintaining computational efficiency [9]. A higher modularity score signifies a more robust community structure, characterized by tightly interconnected nodes within communities while having fewer connections between communities [43,44]. The calculation of modularity score is shown in Equation (9). This indicates the algorithm’s successful identification of distinct and internally cohesive communities, demonstrating their meaningful organization within the network. The Louvain community detection method takes O ( n log n ) time complexity, where n is the graph’s number of nodes.
In community detection algorithms, modularity plays a significant impact in determining the standard of the communities that are formed. Girvan and Newman introduced the modularity measure to assess the standards of the communities that are formed and defined modularity as shown in Equation (9) [34].
G = 1 2 m ( i , j ) A i j k i k j 2 m δ c i , c j
where m is the total number of edges in the graph, A i j is the edge weight between the nodes i and j, k i and k j are the degree of nodes i and j, respectively, c i , c j are the communities to which nodes i and j belong, and δ ( c i , c j ) is the Kronecker delta function which is equal to 1 if c i = c j and 0 otherwise. The term [ A i j k i k j 2 m ] represents the difference between the observed number of edges between the nodes i and j and the expected number of edges under a null model where edges are placed at random. The values of modularity (G) vary between 1 and + 1 .

4. Proposed Method

In this section, a brief description of the motivation behind the proposed method and in detail about the Community-Based Matrix Factorization (CBMF) approach.

4.1. Motivation

In any type of MF method, the rating matrix is constructed by using users and items. The predicted rating matrix is obtained by using different update rules for the latent feature matrices in each kind of MF. The variation between the original and the predicted rating matrix is evaluated using root mean square error. The matrix factorization method predicts a rating for every user item combination. However, there are some cases such that a user may never consume a set of products. Because MF predicts a rating for them, they are recommended to users. This creates a lot of false recommendations. There are only a limited group of items that a user may consume. We propose that by modeling the matrix into a bipartite network containing user and item nodes, users and items can be clustered into tight communities using community detection algorithms. By limiting MF to only the communities, the prediction of false recommendations will significantly reduce. This approach also reduces the time complexity of MF methods.

4.2. Community-Based Matrix Factorization (CBMF) Approach

With this motivation, we propose a Community-Based Matrix Factorization (CBMF) approach, which is given in Algorithm 1.
Algorithm 1: Community-Based Matrix Factorization (CBMF)
Entropy 25 01360 i001
The proposed CBMF constructs a bipartite graph B G containing user and item nodes. The bipartite graph contains a node for each user as well as an item. An edge is formed from the user node to the item node if the user rates the item. The Louvain community detection algorithm is applied on the constructed bipartite graph to divide into communities. Extract rating matrices pertaining only to the nodes in the communities. Note that these rating matrices are very small in size compared to the original rating matrix. Now, apply the MF method to each of these smaller rating matrices in parallel. Combine all the predicted rating matrices into a single prediction rating matrix. Compute RMSE between the original rating matrix and the predicted by the CBMF approach. The procedure is depicted in Figure 3.

4.3. Time Complexity

For the MF method, let a × b be a rating matrix of a users, b items, and k latent features. The iterations are the same for every method and are taken as constant. The time complexity required for calculating the MF method is typically considered to be O ( a b k ) [45]. The Louvain community detection method takes O ( n log n ) time complexity, where n is the number of nodes of graph [46].
The bipartite graph B G containing users and items is given as input for the Louvain algorithm, with the nodes as users and items. Here, in this case, ( a + b ) will be the nodes in the graph. Therefore, the time complexity of the Louvain algorithm will be O ( ( a + b ) log ( a + b ) ) . A bipartite graph B G is constructed from and is divided into appropriate community structures by using the Louvain algorithm (say c 1 , c 2 , , c c ). From each community, a small rating matrix will be obtained (say R 1 , R 2 , , R c ). Apply MF time complexity for each rating matrix and we get a time complexity of O ( a 1 b 1 k 1 ) , O ( a 2 b 2 k 2 ) ,⋯, O ( a c b c k c ) ; consider the maximum of all these time complexities. Suppose R p community rating matrix has the maximum value and the time complexity is O ( a p b p k p ) .
Hence, the overall time complexity for the MF method using the Louvain community detection method is O ( ( a + b ) log ( a + b ) ) + O ( a p b p k p ) .
Similarly for the SVD++ and FANMF methods, the time complexity is O ( a b k ) . Therefore, the total time complexity that is required for the calculation of MF, SVD++, and FANMF methods will be O ( ( a + b ) log ( a + b ) ) + O ( a p b p k p ) .

5. Dataset Statistics

We have evaluated the approach of CBMF on different datasets from different domains. Datasets, namely MovieLens 100K, Film Trust, Jester, Good Books, Wikilens, and Cell Phone Recommendation, are taken. The dataset statistics of all these datasets are shown in Table 1. From all the six datasets we have taken, datasets, namely MovieLens 100K [47], Good Books [48], and Cell Phone Recommendation [49], are taken from the Kaggle repository. The other three datasets, namely Film Trust [50], Jester [51], and Wikilens [52], are taken from the Konect repository.
All of these datasets’ distribution plots are displayed in Figure 4. The plots are drawn by taking the ratings; the X-axis represents users, while the Y-axis represents the count of each rating provided by those users. In the dataset MovieLens 100K, the users are the people and the items are the movies in the dataset. The rating distribution is from 1 to 5 where the people have rated movies. It is observed in the plot that 34,174 of the users had given a higher rating of four for the movies, and a low rating of one is given by 6110 the users out of 100,000 ratings. For the dataset Film Trust, the users are the people and the items are the films having a rating distribution from 0.5 to 4 where the people have rated films. It is observed in the plot that 9170 users had given a rating of five for films, and a low rating of 0.5 is given by the users 1060 out of 35,494 ratings.
In the dataset Jester, the users are the people, and the items are jokes having a rating distribution from 10 to + 10 where the people have rated the jokes. It is observed that there are above 4000 users who have rated + 10 in the dataset out of 1,048,575 ratings.
The Wikilens dataset is shown in Figure 4, where the users are the people and the items are the Wikipedia articles. The rating distribution is given from 0.5 to 5, where the users rated the items in Wikipedia. From the plot, it is observed that a higher rating of four is given by 5721 users, and a low rating of 1.5 is given by 730 users out of 26,936 ratings. In the dataset Good Books, people are the users, and the items are the books having a rating distribution from 1 to 5. The rating is given by the users on different books. From the plot, we can observe that 376,467 of the users have given a higher rating of four for the books, and a low rating of one is given by 24,676 users out of 1,048,575 ratings. For the Cell Phone Recommendation dataset, the users are the people and the items are the cell phone ids having a rating distribution from 1 to 10. The ratings are given by people for different cell phone ids. From the plot, it is observed that a higher rating of eight is given by 196 users, and a low rating of three is given by 30 users out of 990 ratings.
As community detection is a major part of CBMF, we evaluate the modularity of different datasets using the Louvain algorithm for different communities. The modularity of six different datasets using 25 communities is shown in Figure 5. It is observed from all the datasets that the modularity score is increasing as the communities are increased and maintained constant after a certain number of communities. The highest modularity score is maintained by the Wikilens dataset, and the lowest modularity score is given by the MovieLens 100K dataset. The time taken for calculating the modularity score is shown in Table 2. It is observed that the Good Books dataset takes more time to divide the data into 25 communities, and the least time is taken by the Cell Phone Recommendation dataset. The time taken to execute depends on the size of the dataset and the implementation of the computational experiment.
The specifics regarding the hardware and software specifications employed in our implementation are as follows: The system boasts impressive specifications, featuring an 11th Gen Intel Core i9 processor clocked at 2.5 GHz, a formidable 64 GB of RAM, and a 64-bit operating system. Furthermore, Python is the programming language, renowned for its simplicity and effectiveness in a wide array of programming and data analysis tasks.

6. Result Analysis

This section presents a series of experiments designed to demonstrate the validity of the hypothesis proposed in the paper. Using the Louvain community detection algorithm, three different matrix factorization methods (MF, SVD++, and FANMF) are employed to predict the missing entries in the rating matrix across six distinct datasets as shown in Table 1.
The RMSE results of CBMF against their non-community variations for three MF techniques on six datasets are tabulated in Table 3.
Because the best number of communities is not uniform for all networks, Table 3 contains the RMSE results of CBMF where the networks show high modularity in community division using the Louvain algorithm. For instance, MovieLens 100K, Film Trust, and Jester networks have the best modularity when they are divided into 25 communities. Good Books shows high modularity when divided into 20 communities, and Wikilens and Cell Phone when divided into 15 communities. In each network, for each MF technique, CBMF outperformed the non-CBMF technique. Further experimental results are given in detail below.
The computation of the root mean square error (RMSE) as shown in Equation (8) involves varying the number of latent features at values of k as 2 , 10 , 20 , 30 , 50 . Considering the random nature of the Louvain algorithm, each MF method is iterated for 25 communities and 25 iterations. The evaluation of different patterns in RMSE values, along with the total time taken to assess the MF method for six datasets, is presented. The resulting graph depicts the relationship between the number of communities on the X-axis and the corresponding RMSE value on the Y-axis. Given that the Cell Phone Recommendation dataset comprises only 99 users and 33 items, the value of k should be the minimum of number of users and items. Therefore, the iteration for k is limited to 30.
Figure 6 illustrates the variation in RMSE values across six distinct datasets when employing the MF method with different k values on a set of 25 communities. A notable observation from the graph is that, in all datasets, the absence of the Louvain algorithm results in the highest RMSE value when c is 1. However, as the number of communities increases through the utilization of the Louvain algorithm, i.e., c values range from 2 , 3 , , 25 , a decline in the RMSE value is observed. Notably, the MovieLens 100K, Good Books, and Cell Phone Recommendation datasets exhibit an increase in the RMSE value with higher k values. Conversely, in the Film Trust and Jester datasets, the RMSE value stabilizes after reaching a certain number of communities, regardless of the k value. Furthermore, in the Wikilens dataset, there exhibit a constant relationship between the RMSE value for different communities as well as the k values. Overall, it is evident that the MF method consistently yields better RMSE values with lower k values across all datasets.
Figure 7 displays the computational time required to execute the MF method on six distinct datasets, employing different k values across a set of 25 communities. Across all datasets, a noticeable observation is that the computation time for the MF method, without the utilization of community division, exceeds that of the method with community division. As the number of communities increases, a reduction in computational time is observed. Interestingly, the evaluation time of the method remains constant as the number of communities expands, regardless of the specific k values employed.
Figure 8 depicts the computational time required to calculate the root mean square error (RMSE) for the MovieLens 100K dataset when employing MF with both Non-Community-Based Matrix Factorization (Non-CBMF) and Community-Based Matrix Factorization (CBMF) methodologies. It is evident from the figure that RMSE computation without employing the CBMF approach consumes a significant amount of time. For the case of k value 2 latent features, the RMSE calculation takes approximately 529.34 s. In contrast, when utilizing the CBMF approach with the same number of latent features (for k value 2), the time required is significantly reduced to a mere 9.74 seconds. This marked contrast in computation time becomes even more pronounced as the number of latent features increases. Moreover, an analogous trend is observed as the number of communities increases, resulting in an extended RMSE computation time.
Figure 9 presents the RMSE values across six distinct datasets when employing the SVD++ method with varying k values on a set of 25 communities. Notably, at community 1, a significantly high RMSE value is observed across all datasets. However, as the communities are divided, a decline in the RMSE value is observed, reaching a stable state after a certain number of communities. It is worth noting that in the SVD++ method, subtle variations in the RMSE value can be observed for different k values across all datasets.
Figure 10 displays the computational time required to execute the SVD++ method on six distinct datasets, employing different k values across a set of 25 communities. Noteworthy observations can be made across various datasets. In the MovieLens 100K and Good Books datasets, an increase in the number of communities leads to a decrease in the evaluation time of the method. The maximum evaluation time is observed at community 1. Conversely, in the Jester dataset, the maximum time is taken to evaluate the SVD++ method at c value 1, with slight fluctuations in time as the communities increase. Furthermore, for the Film Trust, Wikilens, and Cell Phone Recommendation datasets, it is observed that initially, evaluating the method without dividing into communities requires less time compared to the division into communities. However, after reaching a certain number of community divisions (i.e., k value 18 for Film Trust and Wikilens datasets), the evaluation time becomes less than that at the time taken without community division and remains constant. Similarly, in the case of the Cell Phone Recommendation dataset, the evaluation time becomes less than that at c value 1, and starting from c value 7 constant time is maintained for evaluation.
The obtained RMSE values from applying the FANMF method to six distinct datasets, using different values of k across a set of 25 communities, are illustrated in Figure 11. In all datasets, it is evident that increasing the k value results in a decrease in the RMSE, regardless of the specific communities. Significantly, in the Film Trust, Wikilens, and Cell Phone Recommendation datasets, it is notable that the RMSE values stabilize after a certain number of communities. This stability is observed across all k values, indicating a consistent RMSE value. Moreover, across all datasets, it can be observed that a higher number of latent features leads to lower RMSE values when employing the FANMF method.
The computational time needed to run the FANMF method on six different datasets using various k values over a set of 25 communities is shown in Figure 12. A notable observation across all datasets is that the computation time for the community 2 exceeds that of the community 1. However, after reaching a certain number of communities, the computation time decreases compared to the community 1 for the Film Trust, Wikilens, and Good Books datasets. Conversely, for the MovieLens 100K and Jester datasets, a constant computation time is maintained, which is higher than that of the community 1.

7. Conclusions

In this study, we have delved into the realm of parallel computation. We proposed Community-Based Matrix Factorization (CBMF), which parallelizes matrix factorization utilizing community information. The experimental results obtained from our study provide compelling evidence of the scalability and speedup attained through CBMF, surpassing the performance of sequential implementations. These findings underscore the immense potential of parallel computation in effectively tackling the resources in the distributed environment. Looking ahead, the introduction of our parallel computation framework unlocks fresh opportunities for the efficient processing of large datasets. It empowers researchers and practitioners to extract valuable insights from intricate networks. This framework serves as a stepping stone for future advancements in providing accurate recommendations across diverse domains. It lays the foundation for future innovations and breakthroughs in the realm of data-driven solutions.

Author Contributions

Concept and methodology by S.T. Equal contribution by M.K.E., T.J.L. and H.S., in compiling the primary data and implementing the model. S.T. ran the simulations. Findings, Validation by S.T. and M.K.E., Analysis by H.S. and T.J.L. Manuscript is written by S.T., M.K.E. and T.J.L., H.S. made the whole correspondence with the journal. All authors have read and approved to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

Abbreviations

Abbreviations used in this paper are as follows:
BGBipartite Graph
DMFDeep Matrix Factorization
FANMFFactorized Asymmetric Nonnegative Matrix Factorization
MFMatrix Factorization
NLPNatural Language Processing
NMFnonnegative Matrix Factorization
PCSNMFPairwisely Constrained Nonnegative Symmetric Matrix Factorization
RMSERoot Mean Square Error
SVD++Advanced Singular Value Decomposition

References

  1. Adomavicius, G.; Tuzhilin, A. Toward the next generation of recommender systems: A survey of the state-of-the-art and possible extensions. IEEE Trans. Knowl. Data Eng. 2005, 17, 734–749. [Google Scholar] [CrossRef]
  2. Felfernig, A.; Jeran, M.; Ninaus, G.; Reinfrank, F.; Reiterer, S.; Stettinger, M. Basic approaches in recommendation systems. In Recommendation Systems in Software Engineering; Springer: Berlin/Heidelberg, Germany, 2014; pp. 15–37. [Google Scholar]
  3. Hintz, J. Matrix Factorization for Collaborative Filtering Recommender Systems. 2015. Available online: https://www.cs.utexas.edu/~ans/pubs/hintz_f15.pdf (accessed on 31 July 2023).
  4. Kumar Bokde, D.; Girase, S.; Mukhopadhyay, D. Role of matrix factorization model in collaborative filtering algorithm: A survey. arXiv 2015, arXiv:1503.07475. [Google Scholar]
  5. Koren, Y.; Bell, R.; Volinsky, C. Matrix factorization techniques for recommender systems. Computer 2009, 42, 30–37. [Google Scholar] [CrossRef]
  6. Mehta, R.; Rana, K. A review on matrix factorization techniques in recommender systems. In Proceedings of the 2017 2nd International Conference on Communication Systems, Computing and IT Applications (CSCITA), Mumbai, India, 7–8 April 2017; pp. 269–274. [Google Scholar]
  7. Abdrabbah, S.B.; Ayachi, R.; Amor, N.B. Collaborative filtering based on dynamic community detection. In Proceedings of the 2nd Workshop on Dynamic Networks and Knowledge Discovery, Nancy, France, 15 September 2014. [Google Scholar]
  8. Kumar, P.; Chawla, P.; Rana, A. A review on community detection algorithms in social networks. In Proceedings of the 2018 4th International Conference on Applied and Theoretical Computing and Communication Technology (iCATccT), Mangalore, India, 6–8 September 2018; pp. 304–309. [Google Scholar]
  9. Bedi, P.; Sharma, C. Community detection in social networks. Wiley Interdiscip. Rev. Data Min. Knowl. Discov. 2016, 6, 115–135. [Google Scholar] [CrossRef]
  10. Du, N.; Wu, B.; Pei, X.; Wang, B.; Xu, L. Community detection in large-scale social networks. In Proceedings of the 9th WebKDD and 1st SNA-KDD 2007 Workshop on Web Mining and Social Network Analysis, San Jose, CA, USA, 12 August 2007; pp. 16–25. [Google Scholar]
  11. Karataş, A.; Şahin, S. Application areas of community detection: A review. In Proceedings of the 2018 International Congress on Big Data, Deep Learning and Fighting Cyber Terrorism (IBIGDELFT), Ankara, Turkiye, 3–4 December 2018; pp. 65–70. [Google Scholar]
  12. Lalwani, D.; Somayajulu, D.V.; Krishna, P.R. A community driven social recommendation system. In Proceedings of the 2015 IEEE International Conference on Big Data (Big Data), Santa Clara, CA, USA, 29 October–1 November 2015; pp. 821–826. [Google Scholar]
  13. Guo, W.; Gao, H.; Shi, J.; Long, B.; Zhang, L.; Chen, B.C.; Agarwal, D. Deep natural language processing for search and recommender systems. In Proceedings of the 25th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining, Anchorage, AK, USA, 4–8 August 2019; pp. 3199–3200. [Google Scholar]
  14. Musto, C.; De Gemmis, M.; Lops, P.; Semeraro, G. Generating post hoc review-based natural language justifications for recommender systems. User Model. -User-Adapt. Interact. 2021, 31, 629–673. [Google Scholar] [CrossRef]
  15. Chen, S.; Owusu, S.; Zhou, L. Social network based recommendation systems: A short survey. In Proceedings of the 2013 International Conference on Social Computing, Alexandria, VA, USA, 8–14 September 2013; pp. 882–885. [Google Scholar]
  16. Sun, Z.; Han, L.; Huang, W.; Wang, X.; Zeng, X.; Wang, M.; Yan, H. Recommender systems based on social networks. J. Syst. Softw. 2015, 99, 109–119. [Google Scholar] [CrossRef]
  17. Kim, J.K.; Cho, Y.H.; Kim, W.J.; Kim, J.R.; Suh, J.H. A personalized recommendation procedure for Internet shopping support. Electron. Commer. Res. Appl. 2002, 1, 301–313. [Google Scholar] [CrossRef]
  18. Wei, K.; Huang, J.; Fu, S. A survey of e-commerce recommender systems. In Proceedings of the 2007 International Conference on Service Systems and Service Management, Chengdu, China, 9–11 June 2007; pp. 1–5. [Google Scholar]
  19. Kim, K.j.; Ahn, H. A recommender system using GA K-means clustering in an online shopping market. Expert Syst. Appl. 2008, 34, 1200–1209. [Google Scholar] [CrossRef]
  20. Hasan, M.R.; Jha, A.K.; Liu, Y. Excessive use of online video streaming services: Impact of recommender system use, psychological factors, and motives. Comput. Hum. Behav. 2018, 80, 220–228. [Google Scholar] [CrossRef]
  21. Park, D.H.; Kim, H.K.; Choi, I.Y.; Kim, J.K. A literature review and classification of recommender systems research. Expert Syst. Appl. 2012, 39, 10059–10072. [Google Scholar] [CrossRef]
  22. Fewtrell, M.; Bronsky, J.; Campoy, C.; Domellöf, M.; Embleton, N.; Mis, N.F.; Hojsak, I.; Hulst, J.M.; Indrio, F.; Lapillonne, A.; et al. Complementary feeding: A position paper by the European Society for Paediatric Gastroenterology, Hepatology, and Nutrition (ESPGHAN) Committee on Nutrition. J. Pediatr. Gastroenterol. Nutr. 2017, 64, 119–132. [Google Scholar] [CrossRef] [PubMed]
  23. Isinkaye, F.O. Matrix factorization in recommender systems: Algorithms, applications, and peculiar challenges. IETE J. Res. 2021, 1–14. [Google Scholar] [CrossRef]
  24. Schafer, J.B.; Konstan, J.A.; Riedl, J. E-commerce recommendation applications. Data Min. Knowl. Discov. 2001, 5, 115–153. [Google Scholar] [CrossRef]
  25. Hallinan, B.; Striphas, T. Recommended for you: The Netflix Prize and the production of algorithmic culture. New Media Soc. 2016, 18, 117–137. [Google Scholar] [CrossRef]
  26. Paatero, P.; Tapper, U. Positive matrix factorization: A nonnegative factor model with optimal utilization of error estimates of data values. Environmetrics 1994, 5, 111–126. [Google Scholar] [CrossRef]
  27. Mnih, A.; Salakhutdinov, R.R. Probabilistic matrix factorization. Advances in Neural Information Processing Systems. 2007, Volume 20. Available online: https://proceedings.neurips.cc/paper_files/paper/2007/file/d7322ed717dedf1eb4e6e52a37ea7bcd-Paper.pdf (accessed on 31 July 2023).
  28. Mastorakis, N.E. The singular value decomposition (svd) in tensors (multidimensional arrays) as an optimization problem. solution via genetic algorithms and method of nelder-mead. WSEAS Trans. Syst. 2007, 6, 17–23. [Google Scholar]
  29. Hu, Y.; Koren, Y.; Volinsky, C. Collaborative filtering for implicit feedback datasets. In Proceedings of the 2008 Eighth IEEE International Conference on Data Mining, Pisa, Italy, 15–19 December 2008; pp. 263–272. [Google Scholar]
  30. Shi, X.; Lu, H.; He, Y.; He, S. Community detection in social network with pairwisely constrained symmetric nonnegative matrix factorization. In Proceedings of the Proceedings of the 2015 IEEE/ACM International Conference on Advances in Social Networks Analysis and Mining 2015, Paris, France, 25–28 August 2015; pp. 541–546. [Google Scholar]
  31. Xue, H.J.; Dai, X.; Zhang, J.; Huang, S.; Chen, J. Deep matrix factorization models for recommender systems. In Proceedings of the IJCAI, Melbourne, Australia, 19–25 August 2017; pp. 3203–3209. [Google Scholar]
  32. Schlichtkrull, M.; Kipf, T.N.; Bloem, P.; Van Den Berg, R.; Titov, I.; Welling, M. Modeling relational data with graph convolutional networks. In Proceedings of the The Semantic Web: 15th International Conference, ESWC 2018, Heraklion, Crete, Greece, 3–7 June 2018; Proceedings 15. Springer: Cham, Switzerland, 2018; pp. 593–607. [Google Scholar]
  33. Tosyali, A.; Kim, J.; Choi, J.; Jeong, M.K. Regularized asymmetric nonnegative matrix factorization for clustering in directed networks. Pattern Recognit. Lett. 2019, 125, 750–757. [Google Scholar] [CrossRef]
  34. Girvan, M.; Newman, M.E. Community structure in social and biological networks. Proc. Natl. Acad. Sci. USA 2002, 99, 7821–7826. [Google Scholar] [CrossRef]
  35. Raghavan, U.N.; Albert, R.; Kumara, S. Near linear time algorithm to detect community structures in large-scale networks. Phys. Rev. E 2007, 76, 036106. [Google Scholar] [CrossRef]
  36. Blondel, V.D.; Guillaume, J.L.; Lambiotte, R.; Lefebvre, E. Fast unfolding of communities in large networks. J. Stat. Mech. Theory Exp. 2008, 2008, P10008. [Google Scholar] [CrossRef]
  37. Pons, P.; Latapy, M. Post-processing hierarchical community structures: Quality improvements and multi-scale view. Theor. Comput. Sci. 2011, 412, 892–900. [Google Scholar] [CrossRef]
  38. Traag, V.A.; Waltman, L.; Van Eck, N.J. From Louvain to Leiden: Guaranteeing well-connected communities. Sci. Rep. 2019, 9, 5233. [Google Scholar] [CrossRef] [PubMed]
  39. Kumar, R.; Verma, B.; Rastogi, S.S. Social popularity based SVD++ recommender system. Int. J. Comput. Appl. 2014, 87. [Google Scholar] [CrossRef]
  40. Rendle, S. Factorization machines. In Proceedings of the 2010 IEEE International Conference on Data Mining, Sydney, Australia, 13–17 December 2010; pp. 995–1000. [Google Scholar]
  41. Hosseinzadeh Aghdam, M.; Daryaie Zanjani, M. A novel regularized asymmetric nonnegative matrix factorization for text clustering. Inf. Process. Manag. 2021, 58, 102694. [Google Scholar] [CrossRef]
  42. Lee, D.D.; Seung, H.S. Learning the parts of objects by non-negative matrix factorization. Nature 1999, 401, 788–791. [Google Scholar] [CrossRef] [PubMed]
  43. Alzahrani, T.; Horadam, K.J. Community detection in bipartite networks: Algorithms and case studies. In Complex Systems and Networks: Dynamics, Controls and Applications; Springer: Berlin/Heidelberg, Germany, 2015; pp. 25–50. [Google Scholar]
  44. Gupta, S.; Singh, D.P. Recent trends on community detection algorithms: A survey. Mod. Phys. Lett. B 2020, 34, 2050408. [Google Scholar] [CrossRef]
  45. He, X.; Zhang, H.; Kan, M.Y.; Chua, T.S. Fast matrix factorization for online recommendation with implicit feedback. In Proceedings of the 39th International ACM SIGIR conference on Research and Development in Information Retrieval, Pisa, Italy, 17–21 July 2016; pp. 549–558. [Google Scholar]
  46. Seifikar, M.; Farzi, S.; Barati, M. C-blondel: An efficient Louvain-based dynamic community detection algorithm. IEEE Trans. Comput. Soc. Syst. 2020, 7, 308–318. [Google Scholar] [CrossRef]
  47. Kaggle. Available online: https://www.kaggle.com/datasets/prajitdatta/movielens-100k-dataset (accessed on 31 July 2023).
  48. Konect. Available online: https://www.kaggle.com/datasets/tranhungnghiep/goodbooks6m (accessed on 31 July 2023).
  49. Konect. Available online: https://www.kaggle.com/datasets/meirnizri/cellphones-recommendations (accessed on 31 July 2023).
  50. Konect. Available online: http://konect.cc/networks/librec-filmtrust-ratings/ (accessed on 31 July 2023).
  51. Konect. Available online: http://konect.cc/networks/Jester2/ (accessed on 31 July 2023).
  52. Konect. Available online: http://konect.cc/networks/Wikilens-ratings/ (accessed on 31 July 2023).
Figure 1. An example of simple rating matrix.
Figure 1. An example of simple rating matrix.
Entropy 25 01360 g001
Figure 2. An example of a user-item bipartite network with edges denoting a rating given by the user for the item.
Figure 2. An example of a user-item bipartite network with edges denoting a rating given by the user for the item.
Entropy 25 01360 g002
Figure 3. Community-Based Matrix Factorization (CBMF) framework.
Figure 3. Community-Based Matrix Factorization (CBMF) framework.
Entropy 25 01360 g003
Figure 4. Rating distribution plots for six different datasets, namely MovieLens 100K, Film Trust, Jester, Wikilens, Good Books, and Cell Phone Recommendation.
Figure 4. Rating distribution plots for six different datasets, namely MovieLens 100K, Film Trust, Jester, Wikilens, Good Books, and Cell Phone Recommendation.
Entropy 25 01360 g004
Figure 5. Modularity performance for six different datasets, namely MovieLens 100K, Film Trust, Jester, Wikilens, Good Books, and Cell Phone Recommendation for the Louvain algorithm.
Figure 5. Modularity performance for six different datasets, namely MovieLens 100K, Film Trust, Jester, Wikilens, Good Books, and Cell Phone Recommendation for the Louvain algorithm.
Entropy 25 01360 g005
Figure 6. RMSE plots using MF method for six different datasets, namely MovieLens 100K, Film Trust, Jester, Wikilens, Good Books, and Cell Phone Recommendation with 25 communities and different k (latent features) values.
Figure 6. RMSE plots using MF method for six different datasets, namely MovieLens 100K, Film Trust, Jester, Wikilens, Good Books, and Cell Phone Recommendation with 25 communities and different k (latent features) values.
Entropy 25 01360 g006
Figure 7. Time taken to calculate MF method for six different datasets, namely MovieLens 100K, Film Trust, Jester, Wikilens, Good Books, and Cell Phone Recommendation with 25 communities.
Figure 7. Time taken to calculate MF method for six different datasets, namely MovieLens 100K, Film Trust, Jester, Wikilens, Good Books, and Cell Phone Recommendation with 25 communities.
Entropy 25 01360 g007
Figure 8. (a) Time taken to calculate RMSE for community 1 using non-CBMF approach for MF method in MovieLens 100K dataset. (b) Time taken to calculate RMSE for community 2 using CBMF approach for MF method in MovieLens 100K dataset.
Figure 8. (a) Time taken to calculate RMSE for community 1 using non-CBMF approach for MF method in MovieLens 100K dataset. (b) Time taken to calculate RMSE for community 2 using CBMF approach for MF method in MovieLens 100K dataset.
Entropy 25 01360 g008
Figure 9. RMSE plots using SVD++ method for six different datasets, namely MovieLens 100K, Film Trust, Jester, Wikilens, Good Books, and Cell Phone Recommendation with 25 communities.
Figure 9. RMSE plots using SVD++ method for six different datasets, namely MovieLens 100K, Film Trust, Jester, Wikilens, Good Books, and Cell Phone Recommendation with 25 communities.
Entropy 25 01360 g009
Figure 10. Time taken to calculate SVD++ method for six different datasets, namely MovieLens 100K, Film Trust, Jester, Wikilens, Good Books, and Cell Phone Recommendation with 25 communities.
Figure 10. Time taken to calculate SVD++ method for six different datasets, namely MovieLens 100K, Film Trust, Jester, Wikilens, Good Books, and Cell Phone Recommendation with 25 communities.
Entropy 25 01360 g010
Figure 11. RMSE plots using FANMF method for six different datasets, namely MovieLens 100K, Film Trust, Jester, Wikilens, Good Books, and Cell Phone Recommendation with 25 communities.
Figure 11. RMSE plots using FANMF method for six different datasets, namely MovieLens 100K, Film Trust, Jester, Wikilens, Good Books, and Cell Phone Recommendation with 25 communities.
Entropy 25 01360 g011
Figure 12. Time taken to calculate FANMF method for six different datasets, namely MovieLens 100K, Film Trust, Jester, Wikilens, Good Books, and Cell Phone Recommendation with 25 communities.
Figure 12. Time taken to calculate FANMF method for six different datasets, namely MovieLens 100K, Film Trust, Jester, Wikilens, Good Books, and Cell Phone Recommendation with 25 communities.
Entropy 25 01360 g012
Table 1. Dataset statistics for six different datasets, namely MovieLens 100K, Film Trust, Jester, Wikilens, Good Books, and Cell Phone Recommendation datasets.
Table 1. Dataset statistics for six different datasets, namely MovieLens 100K, Film Trust, Jester, Wikilens, Good Books, and Cell Phone Recommendation datasets.
DatasetNumber of UsersNumber of ItemsNumber of RatingsRating ScaleAverage RatingSparsity
MovieLens
100K
9431682100,0001–53.5290.937
Film Trust1508207135,4940.5–43.0020.988
Jester31,9581401,048,575−10–+100.9550.839
Wikilens3265111269360.5–53.4680.983
Good
Books
13,12377741,048,5751–53.8060.989
Cell Phone
Recommendation
99339901–106.6890.708
Table 2. Time taken (in seconds) to calculate the community modularity score for six different datasets, namely MovieLens 100K, Film Trust, Jester, Wikilens, Good Books, and Cell Phone Recommendation.
Table 2. Time taken (in seconds) to calculate the community modularity score for six different datasets, namely MovieLens 100K, Film Trust, Jester, Wikilens, Good Books, and Cell Phone Recommendation.
DatasetTime (s)
MovieLens 100K1473.21
Film Trust936.00
Jester11,714.99
Wikilens1068.41
Good Books43,125.11
Cell Phone Recommendation15.58
Table 3. RMSE results of CBMF vs. non-CBMF.
Table 3. RMSE results of CBMF vs. non-CBMF.
MF method (→)/
Dataset (↓)
Basic MFSVD++FANMF
Non-CBMFCBMFNon-CBMFCBMFNon-CBMFCBMF
MovieLens 100K3.80.371.260.210.750.2
Film Trust7.690.0020.40.0030.20.0001
Jester3.10.551.90.470.840.005
Wikilens5.110.0040.630.0050.440.002
Good Books3.90.110.550.020.180.009
Cell Phone
Recommendation
5.80.424.70.72.90.003
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Tokala, S.; Enduri, M.K.; Lakshmi, T.J.; Sharma, H. Community-Based Matrix Factorization (CBMF) Approach for Enhancing Quality of Recommendations. Entropy 2023, 25, 1360. https://doi.org/10.3390/e25091360

AMA Style

Tokala S, Enduri MK, Lakshmi TJ, Sharma H. Community-Based Matrix Factorization (CBMF) Approach for Enhancing Quality of Recommendations. Entropy. 2023; 25(9):1360. https://doi.org/10.3390/e25091360

Chicago/Turabian Style

Tokala, Srilatha, Murali Krishna Enduri, T. Jaya Lakshmi, and Hemlata Sharma. 2023. "Community-Based Matrix Factorization (CBMF) Approach for Enhancing Quality of Recommendations" Entropy 25, no. 9: 1360. https://doi.org/10.3390/e25091360

APA Style

Tokala, S., Enduri, M. K., Lakshmi, T. J., & Sharma, H. (2023). Community-Based Matrix Factorization (CBMF) Approach for Enhancing Quality of Recommendations. Entropy, 25(9), 1360. https://doi.org/10.3390/e25091360

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop