Next Article in Journal
The Shape Operator of Real Hypersurfaces in S6(1)
Previous Article in Journal
On the Strong Secure Domination Number of a Graph
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

SSRES: A Student Academic Paper Social Recommendation Model Based on a Heterogeneous Graph Approach

Shanghai Engineering Research Center of Intelligent and Big Data, Shanghai Normal University, Shanghai 201418, China
*
Author to whom correspondence should be addressed.
Mathematics 2024, 12(11), 1667; https://doi.org/10.3390/math12111667
Submission received: 25 April 2024 / Revised: 20 May 2024 / Accepted: 23 May 2024 / Published: 27 May 2024
(This article belongs to the Section Mathematics and Computer Science)

Abstract

:
In an era overwhelmed by academic big data, students grapple with identifying academic papers that resonate with their learning objectives and research interests, due to the sheer volume and complexity of available information. This study addresses the challenge by proposing a novel academic paper recommendation system designed to enhance personalized learning through the nuanced understanding of academic social networks. Utilizing the theory of social homogeneity, the research first constructs a sophisticated academic social network, capturing high-order social relationships, such as co-authorship and advisor–advisee connections, through hypergraph modeling and advanced network representation learning techniques. The methodology encompasses the development and integration of a hypergraph convolutional neural network and a contrastive learning framework to accurately model and recommend academic papers, focusing on aligning with students’ unique preferences and reducing reliance on sparse interaction data. The findings, validated across multiple real-world datasets, demonstrate a significant improvement in recommendation accuracy, particularly in addressing the cold-start problem and effectively mapping advisor–advisee relationships. The study concludes that leveraging complex academic social networks can substantially enhance the personalization and precision of academic paper recommendations, offering a promising avenue for addressing the challenges of academic information overload and fostering more effective personalized learning environments.

1. Introduction

In an era defined by rapid advancements in information science and technology, the academic world has undergone a significant transformation, marked by an exponential growth in electronic literature. This surge in scholarly publications has heralded the onset of the “academic big data age”, bringing forth challenges such as information overload and learning disorientation. These challenges have made it increasingly difficult for students and researchers to navigate the vast sea of academic data and to find literature relevant to their research interests [1,2,3,4,5,6,7]. In response, academic paper recommendation systems have emerged as crucial tools. By leveraging sophisticated algorithms, these systems offer personalized recommendations that closely align with individuals’ research interests and directions, thus significantly enhancing the efficiency of academic research.
However, the utility of these systems is challenged by the vast scope of academic data and the complexity of pinpointing users’ interests. The essential task is to refine these systems for better navigation through the academic social fabric, thus elevating recommendation accuracy. There is a pressing need for advanced systems that can adeptly handle academic nuances and social network dynamics, bridging the resource–needs gap within the scholarly community more effectively.
The academic paper recommendation systems employ diverse strategies like content-based filtering, collaborative filtering, graph-based approaches, and hybrid models to link researchers with pertinent literature [8,9,10,11,12,13,14,15]. However, issues like data sparsity and the cold-start problem persist, hindering their efficacy by veiling users’ specific preferences, especially for new or seldom-interacted items. Our research proposes integrating social relationships into the recommendation framework, drawing on social science insights regarding the impact of social networks on individual preferences and decisions. This novel approach aims to improve recommendation accuracy and personalization by leveraging academic social networks, addressing key challenges and marking a methodological advancement in the field.
In the rapidly advancing fields of academic research and artificial intelligence, exploring complex academic networks, spanning co-authorships, citations and advisor–advisee relationships, presents both significant potential and challenges. These networks are vital for understanding scholars’ academic paths but are difficult to analyze due to their dynamic and intricate nature. Current academic paper recommendation systems often simplify these relationships, neglecting the nuanced dynamics of higher-order social interactions and the full spectrum of academic collaborations. This leads to an incomplete modeling of academic social networks, thereby hindering system performance. Additionally, despite the success of graph-based models, the utilization of hypergraph structures for capturing complex social relationships remains underexplored. Coupled with the challenges of acquiring high-quality labeled data and the scarcity of interaction data, the effectiveness of these models is limited. Moreover, the role of academic papers as central nodes rich in semantic information and intricate interactions is often overlooked in representation learning. Addressing these gaps and recognizing the multifaceted interactions within academic networks could significantly enhance recommendation accuracy and personalization.
Addressing these gaps, our study proposes a comprehensive approach that minimizes the dependency on student–paper interaction data and thoroughly captures the constructed complex multiple academic social relationships and the intricate interactions within heterogeneous academic literature networks. By learning the node representations of students and academic papers comprehensively, this work aims to optimally provide students with the academic paper resources that best meet their learning needs, marking a significant advancement in the field of academic paper recommendation systems.
In this study, we introduce the Student Academic Paper Social Recommendation Model (SSRES), a novel framework developed to mitigate the sparse interaction dilemma between students and academic papers in recommendation systems. Initially, SSRES meticulously constructs a complex academic social network by modeling high-order social relationships, such as co-authorship and co-citation, which forms the basis for a more relevant recommendation output. Progressing further, the model enriches the multi-relational network through the extraction of advisor–advisee relationship dynamics and academic team collaboration insights. Employing a contrastive learning approach, SSRES effectively captures and represents the nuanced interaction patterns between students and academic papers, decreasing the traditional reliance on extensive interaction data. The framework integrates self-supervised learning techniques inspired by social theories and a heterogeneous graph approach, fine-tuning user embeddings to enhance the personalization of the recommendation process. Each step in the development of SSRES represents a strategic move towards a comprehensive solution for personalized academic paper recommendations that cater specifically to the individualized learning requirements of students.
The major contributions of this work are summarized as follows:
  • The development and implementation of SSRES, a novel framework that integrates high-order social relationships into the fabric of academic social networks through the adept utilization of hypergraphs. This framework combines transition mechanisms with advanced network representation learning to encode nuanced co-authorial interactions, enabling the automated and precise delineation of advisor–advisee relationship dynamics and academic collaboration networks.
  • The introduction of a dual-structured enhancement to the recommendation process, including a ‘cross-social relation contrastive’ learning framework for refining student representations and a parallel contrastive learning strategy for academic papers. This approach enriches the representational depth of the student academic social network and provides a comprehensive framework for academic paper representation.
  • A synergistic integration of the core recommendation task with advanced self-supervised learning frameworks, facilitating a holistic optimization approach that uncovers students’ latent preferences. This strategy not only advances the precision of academic paper recommendations but also demonstrates the effectiveness of SSRES through empirical evaluations on real-world datasets.

2. Related Work

2.1. Advisor–Advisee Relationship Identification

The exploration of advisor–advisee relationships within academic networks has garnered significant attention, reflecting the importance of accurately mapping the intricate web of scholarly interactions. Among the various relationships forming the foundation of academic networks, the advisor–advisee relationship stands out as pivotal, influencing the formation of collaborative teams and the modeling of academic entities [16,17,18,19,20,21,22]. This pursuit falls under the category of relationship mining, a critical aspect of social network analysis, with the potential to benefit academic applications, such as double-blind peer reviews and the generation of academic genealogies. In addressing the challenges associated with extracting advisor–advisee relationships from dynamic and complex academic collaboration networks, several researchers have proposed innovative solutions. Notably, Wang et al. [23] introduced the TPFG model, designed to extract advisor–advisee relationships by leveraging the network’s inherent structure and the rich semantic information of nodes and edges. Zhao et al. [24] proposed the deep learning-based tARMM model, which recognizes advisor–advisee relationships through a sophisticated analysis of academic collaboration networks. Further advancing the field, Wang et al. [25] developed the Shifu model, a deep learning approach that not only identifies advisor–advisee relationships but also contributes a large-scale mentorship relationship dataset for future research endeavors. These models underscore the field’s movement towards more nuanced and effective methods of identifying advisor–advisee relationships, acknowledging the limitations posed by the lack of high-quality, real-world advisor–advisee relationship data.
Despite these advances, the current methodologies for advisor–advisee relationship identification face limitations, particularly in fully utilizing the network structure information and the semantic richness of the academic collaboration networks. The challenge lies in automating the extraction of advisor–advisee relationships from academic collaboration networks, a task made difficult by the networks’ dynamic nature and the scarcity of quality advisor–advisee relationship data. This gap highlights a critical area for future research, aiming to enhance our understanding of academic communities from the perspectives of team formation and academic entity modeling.

2.2. Social Recommendation

In the realm of information filtering technology, recommendation systems have become integral to enhancing user experience across various platforms, from Amazon’s product suggestions to Douban’s book recommendations. However, traditional algorithms face challenges like sparsity and the cold-start problem. Inspired by social science theories that suggest a user’s preferences and decisions are often influenced by their social circle, researchers have integrated social network data into recommendation systems for more personalized suggestions [26,27].
Early explorations primarily focused on matrix factorization techniques. For instance, “Sorec” proposed by Ma et al. [28] was among the first to incorporate social network information to address data sparsity. Ma et al. [29] further developed this concept through “social regularization”, enhancing the model’s ability to leverage social connections for improved recommendations. Yang et al. [30] introduced a trust-based social collaborative filtering approach, emphasizing the significance of trust in refining recommendation quality. Yin et al. [31] and Yin et al. [32] explored group dynamics, showcasing how understanding social influence within groups could lead to more accurate group recommendations.
With the advent of deep learning, graph neural networks (GNNs) have broadened the scope of social recommendations. Zhou et al. [33] provided a comprehensive review of GNN methods and applications, laying the groundwork for their use in capturing complex user interactions. “GraphRec”, introduced by Fan et al. [34], marked the first application of GNNs to social recommendation, demonstrating the networks’ capability to model the intricate web of user and item interactions effectively. Subsequent studies by Wu et al. [35], Song et al. [36] and Wei et al. [37] further expanded on this foundation, introducing models like “Dual Graph Attention Networks” and “Dynamic Graph Attention Networks” to represent multifaceted social effects in recommender systems.
Extensive research has focused on integrating social data and developing new algorithms to improve the performance of recommendation systems. However, in the field of academic recommendation systems, the comprehensive integration of complex academic social relationships to provide personalized and precise recommendations remains relatively neglected. The existing literature primarily concentrates on utilizing basic social relationships or trust information to enhance recommendations. Nonetheless, there is a noticeable gap in research on effectively capturing and leveraging the multi-dimensional features of student social relationships, as well as improving recommendation accuracy in cold-start scenarios.

2.3. Contrastive Learning in Self-Supervised Learning

In recent years, self-supervised learning has achieved remarkable success across various domains, including image processing and natural language processing. Among the techniques under the umbrella of self-supervised learning, contrastive learning has particularly drawn widespread attention from researchers due to its effective learning paradigm [38,39]. This approach learns low-dimensional, dense embeddings by contrasting positive and negative samples within the data, enabling the model to learn distinctive embeddings even in the absence of labels. The application of contrastive learning has been extended to the field of graph representation learning, showing considerable potential when integrated with recommendation systems, although research in this area remains relatively nascent.
Velickovic et al. [40] introduced the Deep Graph Infomax (DGI) model, which constructs local and global representations in the form of sample pairs and utilizes the Infomax principle for contrastive learning. Peng et al. [41] proposed the Graph Mutual Information (GMI) model, which contrasts the central node and its local representations from both node features and topological structures perspectives. Hassani et al. [42] developed the Multi-View Graph Representation Learning (MVGRL) model, focusing on contrasting multiple different views. Lastly, Park et al. [43] presented the Deep Multi-Graph Infomax (DMGI) model, which contrasts local and global representations of graphs on attributed multiplex networks and designs a consensus regularization framework to integrate embeddings from multiple graphs.
However, these existing models primarily focus on general graph representation tasks and do not specifically address the unique challenges posed by recommendation systems. For academic recommendation systems, there is a lot of information on advisor–advisee relationships and social science theories that needs to be specially addressed to improve model performance. For self-supervised learning, work that takes into account multiple aspects of information is still relatively rare and warrants further exploration.

3. Preliminaries

This section introduces the foundational concepts and definitions upon which our study builds. The task of this paper is to provide personalized academic paper recommendations to students. In the recommendation process, let A = { a 1 , a 2 , a 3 , , a m } and P = { p 1 , p 2 , p 3 , , p n } represent the set of scholars and the set of papers in the recommendation system, respectively, where m and n denote the number of scholars and papers, i.e., | A | = m and | P | = n . A subset of the scholar set, A t = { a t 1 , a t 2 , a t 3 , , a t o } , represents the set of students. The binary rating matrix R R m × n stores the interaction data between scholars and papers, where r a p = 1 indicates an interaction between the scholar and the paper, and r a p = 0 indicates no interaction. Moreover, Q R m × d and S R n × d represent the comprehensive embeddings of scholars and papers, respectively.

3.1. Definition 2-1 Heterogeneous Academic Literature Network

A heterogeneous academic literature network is a type of heterogeneous information network that contains multiple types of objects and interaction relationships. Let G = { V , E , B , D } represent the heterogeneous academic literature network, where V and E are the sets of nodes and edges, respectively, and B and D are the sets of node types and edge types, respectively. In this study, the heterogeneous academic literature network G mainly includes three types of nodes, identified as scholars, papers, and citations, denoted by A , P , and C , respectively; and two types of edges identified as writing and citing, denoted by D 1 and D 2 , respectively.

3.2. Definition 2-2 Metapath

A metapath π is defined on a heterogeneous information network G as a path B 1 D 1 B 2 D 2 D l B l , often abbreviated as B 1 B 2 B l , which defines a composite relationship D = D 1 D 2 D l between the node types B 1 and B l , where ∘ represents the composition operator. Therefore, metapaths can depict high-order structures containing rich and complex semantic information.

3.3. Definition 2-3 Hypergraph

A hypergraph is a generalization of a traditional graph where an edge can connect any number of nodes, providing a natural method for modeling complex high-order relationships. Let G h = ( V h , E h , W h ) represent the hypergraph, where V h and E h are the sets of vertices and hyperedges, respectively, and each hyperedge can connect any number of vertices. The diagonal matrix W h is the weight set of hyperedges, which, in this study, are uniformly assigned a weight of 1. The hypergraph can be represented by an incidence matrix H { 0 , 1 } | V h | × | E h | , where for each vertex v V h and hyperedge e E h ,
H v e = 1 if v e , 0 if v e .
Let D and L represent the diagonal degree matrices of vertices and hyperedges, respectively, where
D v v = e E h H v e , L e e = v V h H v e .

3.4. Definition 2-4 Advisor–Advisee Relationship

The advisor–advisee relationship represents the direct knowledge inheritance between scholars, playing a significant role in the learning process of students. This study employs the mentorship relationship R a a = { r i j } , i , j A to denote whether a mentorship relationship exists between scholars i and j. If r i j = 1 , it indicates a mentorship relationship where scholar i is the mentor of scholar j; otherwise, if r i j = 0 , it indicates no mentorship relationship exists.

4. Construction of Students’ Multiple Academic Social Relationships Based on Heterogeneous Academic Networks

In the landscape of academic big data, the imperative to navigate through information overload and enhance learning paths has led to the development of sophisticated academic paper recommendation systems. Central to these systems is the integration of scholarly social relationships, drawing upon the principle of social homophily to surmount the challenge of sparse interactions. This section unveils a methodological framework for constructing a nuanced scholarly social network, emphasizing the extraction of multifaceted student social relationship features. By harnessing high-order social relationships and the distinct academic backgrounds of scholars, the approach aims to elevate the recommendation system’s performance. A detailed exposition of this methodology, including the identification of mentorship relationships and the articulation of a comprehensive academic social network, is provided. The structural diagram illustrating this framework is presented in Figure 1, guiding the reader through the proposed enhancements.

4.1. High-Order Social Relationship Extraction through Co-Authorship and Co-Citation among Students

Conventional methodologies in academic social recommendation have predominantly modeled scholarly interactions as binary networks, simplifying complex academic relationships to mere pairwise connections. This approach overlooks the multifaceted nature of scholarly networks, particularly the nuanced higher-order relationships, such as co-authorship and co-citation, which are pivotal in understanding the academic landscape.
Addressing this limitation, the present research introduces a novel framework by constructing a hypergraph on a diverse academic literature network, denoted as G. This framework aims to capture and articulate higher-order social dynamics among scholars, focusing on the intricate patterns of co-authorship and co-citation. These high-order interactions are defined as the joint authorship of academic papers and the mutual citation of works, respectively. By leveraging the heterogeneous nature of the academic literature network G, the study delineates these sophisticated social connections through specific meta-paths: A D 1 P for co-authorship and A D 1 P D 2 C for co-citation, leading to the establishment of dedicated networks for these relationships.
In traditional graph theory, an edge represents a connection between two nodes, which proves insufficient for encapsulating complex scholarly interactions. In contrast, the hypergraph model extends this concept, allowing a hyperedge to encompass multiple nodes. This extension is critical for accurately representing the elaborate web of high-order academic relationships. Utilizing the constructed networks for co-authorship and co-citation, and adhering to definitions outlined in the study, the research constructs two hypergraphs to model these relationships. These relationships are quantified through incidence matrices H a and H c , with each matrix column representing a unique hyperedge, thereby offering a comprehensive depiction of the academic social fabric.

4.2. Algorithm for Advisor–Advisee Relationship Mining Based on Network Representation Learning and Transfer Idea

To accurately identify advisor–advisee relationships within academic networks, this section introduces an algorithmic framework consisting of four main modules: Scholar Node Representation Learning, Collaborative Edge Representation Learning, Transfer Mechanism, and Advisor–Advisee Relationship Identification. Initially, the Scholar Node Representation Learning module captures characteristic information related to each scholar node, establishing a foundation for understanding collaborative relationships among scholars. Subsequently, the Collaborative Edge Representation Learning module refines these node representations, encoding the collaboration information between scholars into the edge representations. These node and edge representations are then utilized by the Transfer Mechanism, which employs transfer learning techniques to further precisely adjust and capture the complex information embedded within the network. Finally, the Advisor–Advisee Relationship Identification module integrates the prior relationships as low-dimensional embeddings, employing a supervised classifier to determine advisor–advisee connections.

4.2.1. Scholar Node Representation Learning

The academic collaboration network G c = ( V , E c ) is established on the basis of the “writing” interaction among scholars within the academic literature network G, where each node v V symbolizes a scholar, and each edge e E c signifies a co-authorship link between scholars. This study employs the adjacency matrix A c R | V | × | V | to depict the structure of the academic collaboration network, alongside the matrix A n R | V | × f to encapsulate the attribute data of all scholar nodes within the network, where for any scholars v i V and v j V , A i j c = 1 if ( v i , v j ) E c , otherwise A i j c = 0 , with f representing the attribute types of scholar nodes. Additionally, to encompass the nodes’ inherent information along with their neighborhood data in the representation generation process, this study integrates self-loops within the adjacency matrix A c , that is, setting all diagonal elements of A c to 1.
In the realm of the academic collaboration network, node attributes delineate the intrinsic qualities of scholar nodes, which, coupled with the network’s topological characteristics, serve as vital indicators of the network’s attributes. To adeptly model the collaboration dynamics among scholars, our advisor–advisee relationship identification framework initially assimilates both the academic and structural information of scholar nodes within the network to derive their vector representations. The node attribute matrix A n encompasses scholar attribute features as delineated in Table 1, primarily consisting of feature labels and their descriptions.
The paper elaborates on several critical features as follows:
  • Scholar Academic Tenure. Defined by a a i = y l y f + 1 , where y l and y f represent the years of the latest and first publications by scholar i, respectively, academic tenure ( a a i ) underscores the duration of scholar i’s academic contributions.
  • Scholar’s h-index. A metric reflecting academic impact, a scholar’s h-index is illustrated through the scenario where a scholar with an h-index of 6 has published at least six papers, each cited no fewer than six times. This metric is crucial in our advisor–advisee relationship identification model, which is adapted to different academic fields to account for variations in h-index norms.
  • Institutional Affiliation. Advisor–advisee relationship often extends within the same institutional bounds. For scholars with multiple affiliations, a binary vector S a = { s j } j = 1 | L a | is utilized, where s j = 1 signifies the scholar’s association with institution j, and s j = 0 otherwise.
  • Research Focus. The alignment of research interests between a mentor and their mentee is pivotal. Scholars’ research domains are represented by a binary vector S r = { s j } j = 1 | L r | , with s j = 1 indicating the scholar’s engagement in research direction j, and s j = 0 otherwise.
To assimilate neighborhood and attribute information of scholar nodes within the network, this paper utilizes graph convolutional networks, deriving the scholar node embedding matrix A n e as follows:
A n e = σ D c 1 / 2 A c D c 1 / 2 A n W n e
where D c i i = j A i j c is the diagonal degree matrix, W n e is a learnable weight matrix for node embeddings, and σ represents the ReLU activation function.

4.2.2. Collaborative Edge Representation Learning

In addition to the attributes of scholar nodes and the topological characteristics of the network, the collaborative attributes encapsulated within the edges of the academic collaboration network G c = ( V , E c ) are pivotal for comprehensively reflecting the network’s dynamics. The collaboration pattern, especially between a mentor and a mentee, is uniquely profound, often exhibiting a closer connection than with other scholars. Thus, to adeptly model these scholarly collaboration relationships, our advisor–advisee relationship identification model extends its focus to encompass the collaborative attributes present on the network’s edges, aiming to derive vector representations for these collaborative edges. For each edge e E c , we utilize a collaborative edge attribute matrix A e R | E c | × g to document its attribute features, where g denotes the count of collaborative edge attribute types (Table 2).
Significant edge attributes are detailed as follows:
  • Academic Age of Collaboration. Defined as a c i j = y c y f + 1 , where y c and y f denote the years of first co-authorship and initial publication by scholar i, respectively, a c i j quantifies the academic tenure of i at the onset of collaboration with j.
  • Collaborative Similarity between Co-authors. The metric k u l c i j computes the collaboration affinity between scholars i and j as:
    k u l c i j = p n i j 2 1 p n i + 1 p n j
    where p n i j denotes the aggregate co-authored publications, with p n i and p n j reflecting individual scholarly contributions.
To holistically capture edge-centric semantic details within the academic collaboration network, a deep autoencoder is employed for learning edge embeddings. Specifically, the collaborative edge attribute matrix A e is processed through the autoencoder, bifurcated into encoder and decoder segments, each comprising multiple nonlinear transformation layers. The transformation is articulated as:
A 1 e m = f ( W 1 e m A e + b 1 e m )
A j e m = f ( W j e m A j 1 e m + b j e m ) , for j = 2 , , k
where k represents the encoder–decoder layer count, f signifies the activation function, employing tanh for edge embedding acquisition A e e = A k / 2 e m , and sigmoid for reconstructed edge representation A r e . W j e m and b j e m denote the linear transformation and bias for layer j, respectively. Dropout is applied to curb overfitting. The autoencoder’s objective is to minimize the discrepancy between A r e and A e , with loss L e = A e A r e F 2 .
During the edge representation learning’s pre-training phase, the Adaptive Moment Estimation (Adam) optimization method is utilized for fine-tuning the model parameters of the deep autoencoder. Subsequently, the model, with pre-trained weights, is further refined within our advisor–advisee relationship identification framework to enhance its predictive performance.

4.2.3. Transfer Mechanism

In the endeavor to unravel advisor–advisee relationship dynamics within academic collaboration networks, this study integrates network representation learning techniques to dissect node attributes, network architecture, and edge characteristics, thereby capturing the intricate information embedded within. Nevertheless, the interactions among academic nodes often harbor diverse and rich semantic nuances, necessitating their incorporation into the holistic modeling of academic collaborations. Drawing inspiration from transfer mechanisms employed in graph embedding algorithms for knowledge graphs, this investigation adopts a similar approach in the semantic realm of advisor–advisee relationships to encapsulate interactions between scholars, thereby refining the embeddings of both nodes and edges. Specifically, it envisions advisor–advisee relationship transfer, facilitating the translation of student node embeddings towards their mentor counterparts within the advisor–advisee relationship semantic space, leveraged through the collaborative edge embeddings to refine the node and edge representations.
In knowledge graph embedding algorithms, the transfer concept interprets a relation in a triplet (head entity, relation, tail entity) as a vector translation from the head entity to the tail entity, aiming for proximity in the representation space post-transfer. Analogously, in mentorship contexts, the student (advisee) representation augmented by a mentorship relationship vector should approximate the mentor (advisor) representation, i.e., h a d v i s e e + r t a d v i s o r .
However, advisor–advisee relationship models face the challenge of over-convergence due to the often one-to-many nature of mentor–student relationships. This simplistic transfer constraint may result in excessively proximal representations for students sharing a mentor, blurring semantic distinctiveness.
To circumvent this, inspired by advancements in transfer mechanisms within knowledge graph embeddings, this paper initiates with a feature transformation on scholar node representations prior to transfer operations. By mapping these representations to the advisor–advisee relationship’s semantic vector space before the transfer, the study addresses disparities in the feature space between node and edge embeddings and mitigates the over-convergence quandary in advisor–advisee relationship learning.
Echoing strategies from related research on transfer mechanisms in knowledge graph embeddings, this study posits that interactions among academic nodes can be construed as vector translations within the advisor–advisee representation space.
For any scholarly pair ( i , j ) within the academic network exhibiting a collaborative link, the paper first transforms their node representations into the advisor–advisee relationship’s semantic vector space:
A i t r n = σ ( W t r · A i n e + b t r )
A j t r n = σ ( W t r · A j n e + b t r )
where σ denotes the activation function, W t r the feature transformation matrix, and b t r the bias vector.
Acknowledging the advisor–advisee relationship’s inherent directionality, the model adjusts scholar node and edge embeddings by leveraging the transfer mechanism, predicated on collaborative edge embeddings. For scholarly pairs in advisor–advisee relationship linkage, their node embeddings should converge based on the edge embeddings; for those without such linkage, divergence is expected. The optimization criterion of this transfer mechanism, hence, seeks to minimize the loss function L t r :
L t r = max ( γ + d ( A i t r n + A i j e e , A j t r n ) d ( A i t r n + A i j e e , A k t r n ) , 0 )
Here, γ is a margin hyperparameter exceeding zero ( γ > 0 ), and ( i , j ) signifies a negative sample derived via negative sampling for the mentor–mentee dyad ( i , j ) , employing random substitution of either node to generate a collaboratively linked but non-advisor–advisee relationship pair. The distance function d employs Euclidean metrics to gauge proximity.

4.2.4. Advisor–Advisee Relationship Identification

Upon mastering the node representations A t r n and edge representations A e e , this study proceeds to model the intricate tapestry of academic collaborations within the network G c = ( V , E c ) , aiming to distill these relationships into low-dimensional embeddings. Subsequently, a supervised classifier discerns the presence of advisor–advisee connections between collaborating scholars v i V and v j V , represented by an edge e E c .
To refine the identification of mentor–student dyads amidst these scholarly collaborations, the network undergoes a filtration process, excluding pairs less likely to embody advisor–advisee relationship roles. This selective sieve, inspired by the academic tenure hypothesis posited by Wang et al. [23], posits that mentors generally boast a longer academic tenure than their mentees. Utilizing this criterion, the paper meticulously sifts through the network, excising improbable mentor–mentee linkages to forge a refined set of candidate pairs E c c earmarked for advisor–advisee relationship scrutiny. Upon successful classification, the specific mentor–student roles within these dyads are also ascertained.
Employing logistic regression, the paper crafts a supervised classification model, articulated through the loss function:
L l r = ( i , j ) E c c ( L l i j σ ( A i j c e · W l r + B l r ) ) 2
Here, σ signifies the sigmoid activation function, L l i j denotes the authentic advisor–advisee relationship label between scholars i and j, and A c e encapsulates the amalgamation of node and edge representations, i.e., A i j c e = A i t r n A i j e e A j t r n . The parameters W l r and B l r represent the logistic regression’s weight matrix and bias vector, respectively.
Converging towards an optimal model fidelity, the study introduces a composite loss function L a a i , formulated as:
L a a i = α L e + β L t r + L l r + γ L r e g
with α , β , and γ serving as hyperparameters to balance the contribution of each model component. The regularization term L r e g , aimed at mitigating overfitting, is defined as:
L r e g = ( W n e 2 + W e m 2 + b e m 2 + W t r 2 + b t r 2 + W l r 2 + B l r 2 )
to ensure model robustness. The overarching ambition of the advisor–advisee relationship identification model, fueled by the transfer mechanism, pivots on minimizing L a a i through the backpropagation algorithm, leveraging stochastic gradient descent (SGD) for optimization.

4.3. Identification of Academic Collaboration Teams via Cooperation Strength

In the evolving landscape of the natural sciences, characterized by increasing diversification and intersectionality, team-based research has become an indispensable paradigm in educational innovation. This paradigm shift, aiming to enhance research quality and efficiency, pivots from solitary scholarly endeavors to collaborative academic research teams. Such teams facilitate rapid knowledge dissemination and problem-solving through cohesive cooperation, often comprising several core units centered around mentor–mentee dynamics, fostering intimate collaborative bonds.
Leveraging insights from previously delineated mentor–mentee social networks, this investigation further delineates the structure of academic teams, mirroring the distinctive academic trajectories of students within the broader academic collaboration network. This methodology entails constructing a team collaboration network, recalibrating it based on collaboration intensity, and thereby identifying distinct academic teams.
Initiating this process, a team collaboration network G T C = ( V T , E T C ) is established upon the academic collaboration network G c . This network, G T C , emerges from co-authorship dynamics among core team constituents, aimed at discerning academic teams. Herein, each node v V T represents a core team, encapsulated around a mentor–mentee nexus with student memberships, while each edge e E T C delineates collaborative interconnections between these core units.
To intricately assess the collaboration strength between core teams, an index is utilized to assign weights to the edges e E T C within G T C , transforming it into a weighted network G T C = ( V T , E T C , W T C ) . Among various indices to evaluate collaborative vigor, such as CF and CI, this study adopts the Collaboration Intensity Index (CII) [21] for quantifying team collaboration strength within the network. The collaboration intensity index C I I i j between two core teams i and j is computed as:
C I I i j = Δ t i j 2 Δ t i × Δ t j
where Δ t i and Δ t j denote the publication outputs of teams i and j within the network’s temporal bounds, respectively, and Δ t i j signifies the shared publications by teams i and j over the same timeframe. A heightened C I I suggests a denser collaborative link, while a diminished C I I indicates a more tenuous connection. By evaluating the C I I across all edges e E T C , the paper recalibrates G T C ’s edges, rendering a network reweighted by C I I .
Conclusively, recognizing that extensive academic collaboration teams typically revolve around multiple core teams bound by stringent cooperative links, this study amalgamates closely knit core teams by filtering edges within G T C , thus pinpointing the academic teams students affiliate with. A distinct academic relationship constraint coefficient ω filters through G T C , delineating academic teams. For each edge e E T C within G T C , edges below ω denote loose collaborations, whereas those surpassing ω signify tight collaborations. This filtration, premised on the notion that 20% of relationships in social networks are strong and 80% are weak [44], sets ω at the minimum edge weight within the top 20% by weight in G T C . This method merges core teams exhibiting close collaborative ties within G T C , thus definitively identifying student-associated academic teams. Additionally, this exploration harnesses hypergraphs for encoding these academic teams, culminating in the portrayal of a multifaceted academic social network of students.

4.4. Augmenting Student Representation Learning via Social Interactions

Addressing the challenges of the “Academic Big Data Era”, we introduce a self-supervised social recommendation model for academic papers, designed to tackle information overload. This model enhances recommendation quality by leveraging advanced embeddings for students and papers, thus minimizing reliance on sparse interactions. It focuses on three main aspects: improved student representation, comprehensive paper embedding, and model optimization. This approach aims for precise preference identification and enhanced recommendation performance, facilitating a deeper understanding of its components and interactions, as depicted in Figure 2.

4.4.1. Hypergraph-Based Learning of High-Order Social Interactions

Leveraging the intricate multi-relational academic social network outlined in Section 4, this segment endeavors to harness the wealth of social interactions therein to derive comprehensive student embeddings. As elucidated in Section 4.1, for encapsulating the complex social interplays amongst scholars within the heterogeneous academic literature network G, this study erects hypergraphs centered around two pivotal higher-order academic social dynamics: co-authorship and co-citation. These dynamics are quantitatively captured through the incidence matrices H a and H c , respectively, laying the groundwork for two distinct hypergraphs to represent these sophisticated social interactions.
This discourse further explores the application of hypergraph convolutional neural networks (HCNNs) to encode the high-order structural data embodied within student representations corresponding to the aforementioned hypergraphs. Mirroring the encoding stratagem of graph convolutional networks (GCNs) yet tailored for hypergraph structures, HCNNs employ a dual-phase transformation mechanism—node to hyperedge and hyperedge back to node—to aptly encapsulate the original hypergraph formation and its high-order neighborhood essence. The operational formulae are delineated as follows:
Q a l + 1 = σ D a 1 2 H a L a 1 H a T Q a l W a
Q c l + 1 = σ D c 1 2 H c L c 1 H c T Q c l W c
where Q a l + 1 and Q c l + 1 epitomize the student embeddings at the l + 1 layer for co-authorship and co-citation hypergraphs, respectively, with σ representing the ReLU activation function. D a , D c , L a , L c , along with W a , W c , signify the vertex diagonal matrices, hyperedge diagonal matrices, and the trainable parameter matrices corresponding to the co-authorship and co-citation hypergraphs, respectively.
Consequent to L layers of information propagation and refinement, the student embeddings from each layer are aggregated to yield the final embeddings tailored to the high-order social constructs of co-authorship and co-citation, denoted as Q a and Q c , respectively. The amalgamation is executed as follows:
Q a = 1 L + 1 l = 0 L Q a l
Q c = 1 L + 1 l = 0 L Q c l
initiating with Q a 0 = Q c 0 through a process of random initialization.

4.4.2. Cross-Relationship Contrast for Enhanced Student Representation Learning

Veering from conventional self-supervised methodologies that juxtapose original and altered networks for contrastive learning, this segment introduces a novel “cross-relationship contrast” learning paradigm. This innovative framework exploits the symbiotic supervision inherent between two distinct high-order social structures, “co-authorship” and “co-citation”. It is further augmented by a self-supervised signal, rooted in an original sample selection strategy meticulously crafted for this study. Such an approach facilitates the autonomous derivation of student embeddings that are both comprehensive and reflective of the nuanced, multifaceted academic social network landscape.
The pivotal role of positive and negative sample construction in augmenting representation quality within contrastive self-supervised learning is well documented. Prior to delving into the “cross-relationship contrast” framework, an exploration of the mechanism for delineating positive and negative samples is paramount. Traditional methods frequently enlist random sampling or subgraph alterations for sample generation, potentially underutilizing the contrastive learning paradigm’s full potential, especially in recommendation contexts. This predicament prompts the introduction of a novel sample selection strategy that, diverging from conventional methodologies, pivots on indirect social relationships derived from academic collaboration teams. This approach notably diminishes the impact of noise on representation fidelity by integrating context from authentic academic team collaborations.
This strategy succinctly compacts the interaction network spanning team members and academic publications within academic teams into a focused internal student collaboration network G a c = ( V a c , E a c , W a c ) . Here, each node v V a c A t epitomizes a team member (mentors excluded due to their expansive research domains potentially diluting indirect social relationship accuracy and student representation learning quality). An edge e E a c signifies at least one joint paper interaction, with W a c quantifying the shared publications. The adjacency matrix A a c R | V a c | × | V a c | embodies the network’s graphical structure.
Utilizing the Louvain method, a renowned community detection algorithm based on topological configurations, the internal student collaboration network G a c facilitates the revelation of students within academic teams sharing analogous learning dispositions and preferences, thereby uncovering latent social interconnections among team members.
Leveraging these unearthed indirect social ties within academic teams, a “cross-relationship contrast” sample selection modality is formulated, defining positive P and negative N sample cohorts for the contrastive learning schema. For any student node i, team members linked via indirect social channels are deemed positive samples, encapsulated as P i , whereas all remaining team members are categorized as negative samples, N i . Furthermore, for student nodes i devoid of historical academic paper interactions, core team members are ascribed as positive samples P i , with all other students allocated to negative samples N i , thus enabling recommendations in scenarios typified by sparse data.
With the delineation of positive P i and negative N i samples for student node i, their embeddings, specifically tailored to the “co-authorship” and “co-citation” high-order social structures, Q i a and Q i c , respectively, are assimilated into the contrastive learning schema. The contrast loss for the “co-authorship” perspective is articulated as:
L i a = log j P i exp ( sim ( Q i a , Q j c ) / τ ) k { P i N i } exp ( sim ( Q i a , Q k c ) / τ )
Here, sim ( ) computes the cosine similarity between vector pairs, with τ denoting a temperature hyperparameter. This framework, diverging from the norm, encompasses multiple positive samples. For any sample duo, the target embedding within our “co-authorship” purview originates from the “co-authorship” domain, whereas the comparative positive and negative samples derive from the “co-citation” sphere.
Analogously, the contrast loss from the “co-citation” standpoint is evaluated as:
L i c = log j P i exp ( sim ( Q i c , Q j a ) / τ ) k { P i N i } exp ( sim ( Q i c , Q k a ) / τ )
In this scenario, for a given sample pair, the target embedding in our contrastive learning framework originates from the “co-citation” perspective, with the embeddings for positive and negative samples derived from the “co-authorship” perspective.
The aggregate loss for our “cross-relationship contrast” learning framework is formalized as:
L s t = 1 | A t | i A t [ ϵ L i a + ( 1 ϵ ) L i c ]
where ϵ is a hyperparameter that balances the contribution of the two perspectives. In our experiments, the embedding Q a from the “co-authorship” perspective is utilized for the recommendation task as it is presumed to preserve social information pertinent to the academic environment more effectively than the “co-citation” perspective.
Through this method, the study innovatively leverages cross-relationship contrastive learning to enrich the learning of student representations by intertwining the embeddings of student nodes from both “co-authorship” and “co-citation” high-order social structures. This original “cross-relationship contrast” strategy not only fosters a deeper comprehension of student social connections within the intricate academic social network but also introduces a novel approach to addressing the cold-start problem in recommendation tasks.

4.5. Self-Supervised Academic Paper Representation Learning

4.5.1. Learning Local Citation Relationships

Within the multifaceted realm of the heterogeneous academic literature network, a plethora of entity types and multi-faceted interaction relationships coexist, positioning papers as a pivotal explicit node type instrumental in delineating student preferences. This segment endeavors to encapsulate the comprehensive information and intricate interaction dynamics of paper nodes. It aims to autonomously derive detailed academic paper embeddings for recommendation objectives. Central to the paper recommendation schema, the paper citation network, denoted as G C I = { P , E C I } , is a derivative subgraph of the broader heterogeneous academic literature network G , predicated on paper nodes and citation linkages. In this construct, each edge e E C I embodies a citation linkage between pairs of paper nodes. Given the non-discriminatory nature of citing versus cited dynamics, G C I is modeled as an undirected graph.
To distill local neighborhood connections of papers, a single-layer graph convolutional neural network (GCN) is harnessed, capitalizing on the topological framework of the paper citation network. This approach facilitates the acquisition of local embeddings, articulated as follows:
S L = σ D C I 1 2 A C I D C I 1 2 S I W L
Herein, S L delineates the resultant local embedding matrix for papers, while A C I and D C I refer to the adjacency and degree matrices of the paper citation network, respectively. S I , procured through a process of random initialization, and W L , a designated linear transformation matrix, together with σ , the ReLU non-linear activation function, converge to formulate the foundation for learning paper embeddings within the local citation network context.
This methodology underscores a focused effort to mine the latent semantic and structural nuances resident within academic papers, thus contributing to the nuanced recommendation framework that aligns with student academic interests and research trajectories.

4.5.2. Global and Higher-Order Paper Relationship Learning

Expanding beyond the confines of local neighborhood dynamics within the paper citation network, this segment aims to encapsulate the expansive realm of global citation relationships. This endeavor supplements the paper representation learning framework with a comprehensive view of the academic landscape. Hypergraph structures, by transcending traditional graph limitations—where edges link multiple nodes—emerge as a potent tool for modeling intricate higher-order interactions. This characteristic is pivotal for representing global neighborhood affiliations amongst papers, offering a robust mechanism to capture complex interaction patterns.
A hypergraph, constructed from shared citation dynamics within the paper citation network, serves as the foundation for disseminating global insights through hyperedges. This dissemination leverages a “node-hyperedge-node” dual-phase feature transformation strategy, detailed as follows:
S G l + 1 = σ D G 1 2 H G L G 1 H G T S G l W G
Here, S G l + 1 delineates the paper embedding at iteration l + 1 , refined through shared citation interactions, with σ representing the ReLU activation function. D G , H G , L G , and W G correspond to the hypergraph’s vertex diagonal matrix, incidence matrix, hyperedge diagonal matrix, and the trainable parameter matrix, respectively.
Post L iterations of global information circulation and assimilation, the ultimate global paper embedding, S G , materializes through an averaging process across all layers:
S G = 1 L + 1 l = 0 L S G l
Furthermore, this discourse extends to incorporate latent higher-order interaction insights amongst academic papers. Specifically, higher-order interactions, distilled through metapaths that elucidate complex patterns, such as papers “co-authored by the same scholar” and “co-cited by the same scholar”, are adeptly modeled. The encapsulation of these higher-order interactions employs hypergraphs and their neural network counterparts, facilitating the following learning mechanism:
S a d l + 1 = σ D a d 1 2 H a d L a d 1 H a d T S a d l W a d
S c d l + 1 = σ D c d 1 2 H c d L c d 1 H c d T S c d l W c d
S a d l + 1 and S c d l + 1 encapsulate the embeddings for distinct high-order interactions at iteration l + 1 , with H a d , H c d , D a d , D c d , L a d , L c d , W a d and W c d representing the respective incidence matrices, vertex diagonal matrices, hyperedge diagonal matrices, and trainable parameter matrices for the two hypergraphs.
Concluding after L propagation cycles, the final representations for higher-order interactions, S a d and S c d , are ascertained as:
S a d = 1 L + 1 l = 0 L S a d l
S c d = 1 L + 1 l = 0 L S c d l

4.5.3. Local High-Order Contrastive Learning for Papers

This final segment introduces a contrastive learning architecture specifically tailored for academic papers, utilizing supervision signals from both global and local perspectives. This novel approach orchestrates a self-contrast between local representations and their global, higher-order counterparts, aiming to distill comprehensive paper representations. These representations encapsulate the dense web of complex interactions within the heterogeneous academic literature network. In this framework, positive sample pairs are constituted by identical paper nodes viewed through different lenses, while disparate paper nodes within the same views act as negative samples. The formula for calculating the overall loss in this paper contrastive learning architecture is articulated as:
L s p = 1 | P | i P log exp ( sim ( S i L , S i G ) / τ ) + exp ( sim ( S i L , S i a d ) / τ ) + exp ( sim ( S i L , S i c d ) / τ ) k P [ exp ( sim ( S i L , S k G ) / τ ) + exp ( sim ( S i L , S k a d ) / τ ) + exp ( sim ( S i L , S k c d ) / τ ) ]
where sim ( ) calculates the cosine similarity between vector pairs, and τ is a temperature parameter moderating the similarity measure.

4.6. Model Optimization

The preceding discussion has centered around the application of unsupervised, self-supervised signals for capturing both semantic and structural nuances within the network. This facilitated the learning of enriched representations for students and academic papers alike. Nonetheless, the essence of recommendation inherently intertwines with the interaction data between students and papers. Thus, a pairwise learning approach is adopted for optimizing the model within the academic paper top-k recommendation task. This approach not only steers the parameter learning but also refines the embedding learning trajectory.
To align student and paper representations within a unified vector space, the following mappings are employed:
q i = σ ( W Q · Q i a + b Q )
s j = σ ( W S · S j L + b S )
where σ signifies the ReLU activation function. W Q , W S , b Q , and b S represent the transformation matrices and bias vectors responsible for transmuting student and paper feature representations, respectively.
The model’s optimization leverages a pairwise learning strategy, predicated on the axiom that observed interactions ought to rank above non-observed ones:
L O = ( i , j , k ) T S ( r ^ i j r ^ i k 1 ) 2
Here, T S is the training dataset, with each triplet ( i , j , k ) T S incorporating a student a t i A t , an interacted academic paper p j P as a positive example, and a non-interacted academic paper p k P serving as a negative instance. r ^ i j = q i T s j delineates the predicted affinity score between student a t i and paper p j .
Subsequently, the self-supervised frameworks for both students and papers are amalgamated with the main recommendation task’s objective for holistic optimization. The comprehensive objective function is formulated as:
L = L O + δ ( L s t + L s p ) + λ Φ 2 2
In this equation, δ is a balancing hyperparameter for the self-supervised learning component, Φ encompasses the parameters engaged in the recommendation mechanism, and λ is the regularization term.
Building on the CBCS methodology [45] to streamline model computation and elevate recommendation efficacy, this study initially narrows down a candidate paper set for recommendations. Instead of evaluating the entire paper corpus, the recommendation process focuses on a select set of candidate papers, identified through indirect social ties and shared preferences within academic teams, enhancing recommendation diversity. This approach utilizes the CBCS method’s principle, where if a user has interacted with any item in another user’s item set, a weak tie is inferred between these two users. Utilizing these weak ties, the CBCS method narrows down a candidate item set for recommendations. Similarly, this study extends the concept of weak ties to include indirect social relationships within academic teams, identifying candidate scholar sets based on similar preferences discovered in previous sections. Consequently, papers associated with the candidate scholar set form the basis for generating a candidate set for student recommendations. The final top-k academic paper recommendations are then derived from this candidate set, optimizing both the model’s efficiency and the diversity of its recommendations.

5. Experiment

5.1. Experimental Datasets and Preprocessing

5.1.1. Experimental Datasets

This study employs the Aminer dataset [46], a comprehensive repository of academic publications meticulously compiled by Professor Tang and his team at Tsinghua University. Encompassing a wide array of academic disciplines, the Aminer dataset features 2,092,356 academic papers, 8,024,869 citation links, 1,712,433 authors, and 4,258,615 co-authorship instances. The dataset is publicly accessible via https://www.aminer.cn/aminernetwork (accessed on February 2024).
Due to the extensive scope of the original Aminer dataset, a curated subset was selected for the experimental evaluation and analyses conducted within this study. Furthermore, to ascertain the efficacy of the proposed mentor–mentee relationship identification model and to facilitate subsequent academic paper recommendation experiments for students, authentic mentor–mentee relationship pairs are necessitated. To this end, the most recent version of the Academic Family Tree (AFT) dataset [47], harvested from https://academictree.org/ (accessed on February 2024), serves as a vital resource. The AFT project, a dynamic, crowd-sourced endeavor, chronicles veritable academic genealogies spanning 73 disciplines. Table 3 delineates the statistical particulars of select research domains within the AFT dataset, distinguished by their data richness.

5.1.2. Data Preprocessing

Prior to the deployment of the Aminer and AFT datasets in our experimental framework, two pivotal challenges necessitate resolution: dataset alignment and the heterogeneity across disciplines. Aminer and AFT, while independent datasets, must be cohesively aligned for our analysis, given our requirement for an integrated academic publication dataset that encompasses mentor–mentee relationships from AFT for downstream experimentation. Moreover, [48,49] suggests that collaboration patterns among students and mentors vary distinctly across fields, potentially influencing model efficacy. Consequently, our experimental design entails discipline-specific evaluations to gauge both the effectiveness and universality of the proposed model.
In the preprocessing phase, real mentor–mentee relationship pairs from the AFT dataset, spanning the years 2000 to 2015, are meticulously extracted and subsequently aligned with the corresponding entities in the Aminer dataset. Notably, the Aminer dataset is preprocessed to address author name disambiguation, thereby facilitating the straightforward identification of scholars and their real-world identities.
The alignment process unfolds in two stages: initially, scholar names constituting real mentor–mentee pairs from 2000 to 2015 are harvested from the AFT dataset, specifically targeting the research domains outlined in Table 3. Concurrently, author names are extracted from co-authorship instances within the Aminer dataset, with the scholar names then rendered through regular expressions for standardization. Subsequently, scholar pairs from AFT are paired with their counterparts in Aminer. During this pairing process, a co-author duo from Aminer is recognized as a bona fide mentor–mentee pair solely if an exact match of scholar names is established across both datasets, thereby affirming the integrity of the alignment outcomes.
This rigorous preprocessing and alignment exercise across various disciplines culminates in the acquisition of authentic mentor–mentee relationship pairs for our experimental endeavors, culminating in a rich dataset of academic publications replete with verifiable mentor–mentee dynamics. Table 4 delineates the statistical specifics of the leading fields, as determined by the volume of successfully matched mentor–mentee pairs.

5.2. Experiments on Mentor–Mentee Relationship Identification Model

This section presents experiments and analyses of the mentor–mentee relationship identification model proposed in Section 4.2, aiming to demonstrate its effectiveness and universality. It encompasses evaluation metrics, comparison methods, and results analysis.

5.2.1. Evaluation Metrics and Comparison Methods

The advisor–advisee relationship mining issue is considered a binary classification problem, focusing on predicting whether a scholar i is the mentor of scholar j. To evaluate the performance of our mentor–mentee relationship identification model (NTARM), four widely used metrics are selected: precision, recall, accuracy, and F1-score, calculated as follows:
Precision = T P T P + F P
Recall = T P T P + F N
Accuracy = T P + T N T P + F P + T N + F N
F 1 - s c o r e = 2 × Precision × Recall Precision + Recall
where T P (True Positive) is the count of correctly predicted positive relationships; T N (True Negative), the count of correctly predicted negative relationships; F P (False Positive), the count of negative relationships predicted as positive; and F N (False Negative), the count of positive relationships predicted as negative.
Furthermore, our model is compared against the following network representation learning and existing mentor–mentee relationship identification methods:
  • DeepWalk [50], which uses random walks to obtain network locality information and skip-gram for vertex latent representation learning.
  • Node2vec [51], a semi-supervised algorithm that balances homophily and structural equivalence in embeddings through controllable search bias.
  • TransNet [52], a network representation learning model based on transition mechanisms.
  • Shifu [25], a deep learning-based advisor–advisee relationship identification method considering both local attributes and network structural features of scholars.
  • Shifu2 [53], which, besides considering network structural features, also accounts for semantic information of scholar nodes and collaboration edges.

5.2.2. Results and Analysis

This section describes a series of specific experiments and provides analyses to evaluate the effectiveness and applicability of the proposed mentorship relationship identification model NTARM, in comparison with several network representation learning methods and existing mentorship relationship identification models. Firstly, we compare the proposed NTARM model with the network representation learning methods mentioned in Section 5.2.1 across various research domains, including computer science, neuroscience, mathematics and chemistry. The detailed comparison results are presented in Figure 3.
This study divided the original dataset into training and testing sets in an 8:2 ratio based on the collaboration time between mentors and students. Experimental results demonstrate that the proposed NTARM model significantly enhances performance in mentor–mentee relationship identification tasks across various research domains, such as computer science. This indicates the effectiveness of the NTARM model. Specifically, among network representation learning methods, the Transnet model outperforms both DeepWalk and Node2Vec methods across all domains, underscoring the unique advantages of knowledge graph domain transfer concepts in advisor–advisee relationship identification tasks.
Subsequently, experiments and comparisons were conducted on the NTARM model and other state-of-the-art mentor–mentee relationship identification models mentioned in Section 5.2.1, across research domains including computer science, neuroscience, mathematics and chemistry. Detailed comparison results are presented in Table 5.
Additionally, NTARM demonstrates superior performance compared to advanced mentor–mentee relationship identification models across all research domains. The incorporation of semantic information and attribute features on collaboration edges in Shifu2 results in a noticeable enhancement in its performance compared to Shifu. Furthermore, the comprehensive comparisons presented in Figure 3 and Table 5 indicate that the proposed NTARM model consistently performs well in computer science, neuroscience, mathematics and chemistry, highlighting its universality and its ability to overcome disciplinary disparities effectively.

5.2.3. Parameter Sensitivity

This subsection delves into the sensitivity analysis of two critical hyperparameters, α and β , on the performance of the NTARM model, elucidating their impacts and optimizing their values for enhanced model efficacy.

Sensitivity to α

Hyperparameter α modulates the significance of the edge deep autoencoder in the combined loss function of the model. In the realm of chemistry, an exploratory search for α ’s optimal value was conducted within the range [0.01, 0.05, 0.1, 0.5], keeping other hyperparameters constant. The variations in the performance of the NTARM model with different α settings are illustrated in Figure 4a.
The investigation revealed a performance increment in the NTARM model with an increase in α . Notably, the model’s performance begins to plateau upon reaching an α value of 0.05, beyond which further increments in α cease to significantly enhance performance, indicating a threshold of sensitivity to α .

Sensitivity to β

Hyperparameter β influences the weight of the transition module within the joint loss function. With α set to 0.05, the search for the ideal β spanned the range [0.01, 0.05, 0.1, 0.5]. Figure 4b captures the NTARM model’s performance dynamics across varied β settings.
The results indicate a continuous improvement in the NTARM model’s performance with increasing β values, reaching an apex at β = 0.1, after which the performance begins to wane. This pattern underscores the auxiliary role of the transition module in mentor–mentee relationship identification. Smaller β values prove insufficient for augmenting model performance, whereas excessively large β values may disrupt the task-specific learning process during gradient descent.
The parameter sensitivity analysis underscores the importance of judicious hyperparameter tuning for the optimal performance of the NTARM model, illustrating the critical balance required in weighting the model components’ contributions to the overall learning objective.

5.3. Experiments on the Student Academic Paper Social Recommendation Model

This section evaluates the Student Social Recommendation Model for Academic Papers (SSRES) introduced in Section 4.4, to verify its recommendation performance and the significant improvement brought by constructing academic social relationships among students as discussed in Section 4 through ablation studies. As previously mentioned, the experiments utilize the Aminer and AFT datasets, which have been aligned and preprocessed for the experiments. Due to the large size of the original datasets, a subset containing a portion of the original data was used for the experiments in this section, including student nodes, paper nodes, ratings of papers by students, and social information among students, as detailed in Table 6.

5.3.1. Evaluation Metrics and Comparison Methods

The experiments employ commonly used metrics for Top-N recommendation tasks, namely Precision@N, Recall@N, and NDCG@N, with N set to 10. NDCG is one of the most commonly used ranking evaluation metrics, measuring the quality of the ranking of the recommended list against the true items, calculated as follows:
N D C G = D C G I D C G
D C G = r e l 1 + i = 2 N r e l i log 2 i
where D C G increases the impact of items ranked higher in the sequence and decreases the impact of items ranked lower. Specifically, if the i-th item in the sequence is relevant, then r e l i = 1 , otherwise r e l i = 0 , and I D C G is the ideal D C G in the sequence’s optimal state.
To validate the effectiveness of the SSRES model, it is compared against traditional recommendation methods, such as BPR, BPRMF and NeuMF, paper-hybrid recommendation methods, like CTR and UAGMT, and graph neural network-based recommendation methods, including LightGCN, DHCF, and DiffNet++:
  • BPR [54] is a method that derives the maximum posterior probability based on Bayesian analysis and optimizes the model through stochastic gradient descent for personalized ranking recommendations.
  • SBPR [55] improves upon BPR by estimating users’ relative preferences in the form of rankings based on BPR and social relationships.
  • NeuMF [56] uses a neural network architecture to learn the interactions between user and item features for recommendations.
  • CTR [57] combines the advantages of traditional matrix factorization-based collaborative filtering and probabilistic topic models to recommend existing and newly published academic papers to users.
  • UAGMT [58] is an efficient and straightforward dual-relational graph model for recommending newly published academic papers by integrating valuable information, such as readers, tags, content and citations, into the graph.
  • LightGCN [59] simplifies the design of graph convolutional networks (GCN) by using only neighborhood aggregation for collaborative filtering.
  • DHCF [60] introduces a dual-channel hypergraph collaborative filtering framework that incorporates a dual-channel learning strategy and models users and items using a hypergraph structure.
  • DiffNet++ [61] is an improved version of the DiffNet model, modeling neural influence diffusion and interest diffusion within a unified framework.

5.3.2. Results and Analysis

The SSRES model is compared against existing recommendation methods in a series of experiments to evaluate its recommendation performance. The experimental comparison results are presented in Table 7. From these results, the SSRES model demonstrates outstanding performance in the academic paper recommendation task across all evaluation metrics, proving its effectiveness in recommending academic papers to students.
Compared to traditional recommendation methods, paper-hybrid recommendation methods, and graph neural network-based recommendation methods, traditional recommendation methods perform the most modestly. Even the least performing paper-hybrid and graph neural network-based methods outperform traditional methods, with the UAGMT model and the DiffNet++ model showing considerable competitiveness in the student academic paper recommendation task.

5.3.3. Parameter Sensitivity

In this section, we explore the influence of several key hyperparameters on the performance of the SSRES model, specifically δ , the learning rate, and the embedding dimension.

Sensitivity to δ

Hyperparameter δ governs the weight of the self-supervised learning module for student and paper representations within the model’s joint loss. We searched for the optimal value of δ within the range [0.005, 0.01, 0.05, 0.1], and the performance of the SSRES model with various δ values is shown in Figure 5a.
As Figure 5a illustrates, the SSRES model’s performance varies with changes in δ , peaking at a δ value of 0.05 before starting to decline. Therefore, in our experiments, we set δ to 0.05.

Influence of Embedding Dimension

The embedding dimension, a manually set parameter, is crucial in the learning process of comprehensive embeddings for students and academic papers. We investigated the impact of embedding dimension on the SSRES model’s performance across the range [16, 32, 64, 96], with the results depicted in Figure 5b.
As shown in Figure 5b, the SSRES model’s performance varies under different embedding dimensions. With Precision@10 as an example, the model’s performance continually improves with increasing embedding dimension, peaking at 64 before beginning to decrease. Consequently, we set the embedding dimension to 64 for our experiments.

Impact of Learning Rate

The learning rate determines the update speed of the model. A too-high learning rate can cause the loss function to oscillate around the minimum value without converging, while a too-low rate makes the learning process excessively slow. We explored the impact of the learning rate on the SSRES model within the range [0.0001, 0.0005, 0.001, 0.005, 0.01], as depicted in Figure 5c.
As demonstrated in Figure 5c, the model’s performance peaks at a learning rate of 0.001. Therefore, we set the learning rate to 0.001 for our experiments.
This sensitivity analysis illustrates the significant role that the hyperparameters δ , embedding dimension, and learning rate play in optimizing the SSRES model’s performance.

5.4. Ablation Study

This section presents an ablation study to verify the contribution of various components within the proposed SSRES model to the final recommendation performance. Three variants of the model are introduced for this purpose: SSRES-HE, SSRES-AR, and SSRES-SP.
  • SSRES-HE is a variant where the high-order social relationships “co-authorship” and “co-citation” are modeled and learned as binary relations within the SSRES framework.
  • SSRES-AR modifies the sample selection strategy in the cross-relationship contrastive learning framework of SSRES, treating the student node itself as a positive sample and all other student nodes as negative samples.
  • SSRES-SP is the version of the SSRES model without the paper contrastive learning architecture.
The results of the ablation study are shown in Table 8.
From the results in Table 8, it is clear that the SSRES model outperforms the SSRES-HE and SSRES-AR variants in all aspects, which thoroughly demonstrates the significant role of the academic social relationships constructed in Section 4 in enhancing the recommendation effects of the SSRES model. Compared to SSRES-HE, which models high-order social relationships as binary relations and uses graph neural networks for learning, the performance of SSRES-AR is even less satisfactory. This highlights the importance of high-quality self-supervised signals for the self-supervised student representation learning process. Furthermore, the superior performance of SSRES over SSRES-SP underscores that capturing additional information on papers, a crucial explicit representation in academic networks, and refining and learning paper embeddings are indispensable components of the recommendation process. Relying solely on student–paper interaction data for learning paper representations is insufficient.

6. Discussion and Conclusions

The SSRES model represents a significant advancement in academic paper recommendation systems, effectively addressing the challenges of academic information overload and learning disorientation faced by students. The SSRES model integrates multiple information to reflect the complex relationships of academic paper recommendation systems, combining hypergraphs and networks to express high-order social relationships, a dual-structured enhancement to connect students’ social information and academic papers, and self-supervised learning frameworks to complete the recommendation task.
Through the innovative integration of high-order social relationships within the academic social network, SSRES not only refines the recommendation process but also enhances the model’s understanding of students’ diverse academic backgrounds and preferences. This study underscores the model’s effectiveness in leveraging social homophily principles, thereby providing tailored academic resources that foster personalized learning experiences.
Critically, the construction of a multifaceted academic social network emerges as a cornerstone of the SSRES model, enabling a nuanced understanding of complex student interactions and academic interests. The employment of self-supervised learning techniques further distinguishes SSRES, optimizing the representation of both students and academic papers through a contrastive learning framework that captures the essence of academic social relationships and interactions.
Empirical evaluations on real-world datasets reveal the SSRES model’s superior performance, particularly in mining mentor–mentee relationships within academic networks. This validation not only attests to the model’s robustness but also highlights its potential in transforming academic paper recommendation systems.
Looking forward, the exploration of more comprehensive datasets and the refinement of mentor–mentee relationship mining techniques present promising avenues for future research. Addressing the cold-start problem remains a pivotal challenge, inviting innovative approaches to enhance the SSRES model’s applicability across diverse academic settings. This study’s insights pave the way for the development of more effective, efficient, and personalized academic recommendation systems, marking a significant contribution to the field of educational technology.

Author Contributions

Conceptualization, Y.G. and Z.Z.; methodology, Y.G.; software, Z.Z.; validation, Y.G. and Z.Z.; formal analysis, Z.Z.; investigation, Z.Z.; resources, Y.G.; data curation, Z.Z.; writing—original draft preparation, Y.G.; writing—review and editing, Y.G.; visualization, Z.Z.; supervision, Y.G.; project administration, Y.G. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Data Availability Statement

Data are contained within the article.

Acknowledgments

This work was supported by the Shanghai Engineering Research Center of Intelligent Education and Big Data. This work was also supported by the Shanghai Normal University Student Innovation and Entrepreneurship Training Program. We would like to express our gratitude to Qin Zhou from Shanghai Engineering Research Center of Intelligent Education and Big Data for his invaluable support and guidance significantly contributed to the conceptualization of the initial idea and paper writing.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Berkovsky, S.; Freyne, J. Group-based recipe recommendations: Analysis of data aggregation strategies. In Proceedings of the Fourth ACM Conference on Recommender Systems, New York, NY, USA, 26 September 2010; pp. 111–118. [Google Scholar]
  2. Pujahari, A.; Padmanabhan, V. Group recommender systems: Combining user-user and item-item collaborative filtering techniques. In Proceedings of the 2015 International Conference on Information Technology (ICIT), Bhubaneswar, India, 21–23 December 2015; IEEE: Piscataway, NJ, USA, 2015; pp. 148–152. [Google Scholar]
  3. Dwivedi, P.; Bharadwaj, K.K. e-Learning recommender system for a group of learners based on the unified learner profile approach. Expert Syst. 2015, 32, 264–276. [Google Scholar] [CrossRef]
  4. Papamitsiou, Z.; Economides, A.A. Recommendation of educational resources to groups: A game-theoretic approach. In Proceedings of the 2018 IEEE Global Engineering Education Conference (EDUCON), Santa Cruz de Tenerife, Spain, 17–20 April 2018; IEEE: Piscataway, NJ, USA, 2018; pp. 760–767. [Google Scholar]
  5. Kompan, M.; Bielikova, M. Enhancing existing e-learning systems by single and group recommendations. Int. J. Contin. Eng. Educ. Life Long Learn. 2016, 26, 386–404. [Google Scholar] [CrossRef]
  6. Zapata, A.; Menéndez, V.H.; Prieto, M.E.; Romero, C. Evaluation and selection of group recommendation strategies for collaborative searching of learning objects. Int. J. Hum.-Comput. Stud. 2015, 76, 22–39. [Google Scholar] [CrossRef]
  7. Papamitsiou, Z.; Economides, A.A. Motivating students in collaborative activities with game-theoretic group recommendations. IEEE Trans. Learn. Technol. 2018, 13, 374–386. [Google Scholar] [CrossRef]
  8. Beel, J.; Gipp, B.; Langer, S.; Breitinger, C. Paper recommender systems: A literature survey. Int. J. Digit. Libr. 2016, 17, 305–338. [Google Scholar] [CrossRef]
  9. Bai, X.; Wang, M.; Lee, I.; Yang, Z.; Kong, X.; Xia, F. Scientific paper recommendation: A survey. IEEE Access 2019, 7, 9324–9339. [Google Scholar] [CrossRef]
  10. Bauer, J.; Nanopoulos, A. Recommender systems based on quantitative implicit customer feedback. Decis. Support Syst. 2014, 68, 77–88. [Google Scholar] [CrossRef]
  11. Langseth, H.; Nielsen, T.D. Scalable learning of probabilistic latent models for collaborative filtering. Decis. Support Syst. 2015, 74, 1–11. [Google Scholar] [CrossRef]
  12. Barragáns-Martínez, A.B.; Costa-Montenegro, E.; Burguillo, J.C.; Rey-López, M.; Mikic-Fonte, F.A.; Peleteiro, A. A hybrid content-based and item-based collaborative filtering approach to recommend TV programs enhanced with singular value decomposition. Inf. Sci. 2010, 180, 4290–4311. [Google Scholar] [CrossRef]
  13. Park, D.H.; Kim, H.K.; Choi, I.Y.; Kim, J.K. A literature review and classification of recommender systems research. Expert Syst. Appl. 2012, 39, 10059–10072. [Google Scholar] [CrossRef]
  14. Son, J.; Kim, S.B. Academic paper recommender system using multilevel simultaneous citation networks. Decis. Support Syst. 2018, 105, 24–33. [Google Scholar] [CrossRef]
  15. Ferrara, F.; Pudota, N.; Tasso, C. A keyphrase-based paper recommender system. In Proceedings of the Digital Libraries and Archives: 7th Italian Research Conference, IRCDL 2011, Pisa, Italy, 20–21 January 2011; Revised Papers 7. Springer: Berlin/Heidelberg, Germany, 2011; pp. 14–25. [Google Scholar]
  16. Steurer, M.; Trattner, C. Acquaintance or partner? Predicting partnership in online and location-based social networks. In Proceedings of the 2013 IEEE/ACM International Conference on Advances in Social Networks Analysis and Mining, Niagara Falls, ON, Canada, 25–28 August 2013; pp. 372–379. [Google Scholar]
  17. Liu, J.; Kong, X.; Xia, F.; Bai, X.; Wang, L.; Qing, Q.; Lee, I. Artificial intelligence in the 21st century. IEEE Access 2018, 6, 34403–34421. [Google Scholar] [CrossRef]
  18. Xia, F.; Wang, W.; Bekele, T.M.; Liu, H. Big scholarly data: A survey. IEEE Trans. Big Data 2017, 3, 18–35. [Google Scholar] [CrossRef]
  19. Fortunato, S.; Bergstrom, C.T.; Börner, K.; Evans, J.A.; Helbing, D.; Milojević, S.; Petersen, A.M.; Radicchi, F.; Sinatra, R.; Uzzi, B.; et al. Science of science. Science 2018, 359, eaao0185. [Google Scholar] [CrossRef] [PubMed]
  20. Wang, W.; Cui, Z.; Gao, T.; Yu, S.; Kong, X.; Xia, F. Is scientific collaboration sustainability predictable? In Proceedings of the 26th International Conference on World Wide Web Companion, Perth, Australia, 3–7 April 2017; pp. 853–854. [Google Scholar]
  21. Yu, S.; Xia, F.; Zhang, K.; Ning, Z.; Zhong, J.; Liu, C. Team recognition in big scholarly data: Exploring collaboration intensity. In Proceedings of the 2017 IEEE 15th Intl Conf on Dependable, Autonomic and Secure Computing, 15th Intl Conf on Pervasive Intelligence and Computing, 3rd Intl Conf on Big Data Intelligence and Computing and Cyber Science and Technology Congress (DASC/PiCom/DataCom/CyberSciTech), Orlando, FL, USA, 6–10 November 2017; IEEE: Piscataway, NJ, USA, 2017; pp. 925–932. [Google Scholar]
  22. Tang, W.; Zhuang, H.; Tang, J. Learning to infer social ties in large networks. In Proceedings of the Joint European Conference on Machine Learning and Knowledge Discovery in Databases, Bilbao, Spain, 13–17 September 2011; Springer: Berlin/Heidelberg, Germany, 2011; pp. 381–397. [Google Scholar]
  23. Wang, C.; Han, J.; Jia, Y.; Tang, J.; Zhang, D.; Yu, Y.; Guo, J. Mining advisor-advisee relationships from research publication networks. In Proceedings of the 16th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, Washington, DC, USA, 25–28 July 2010; pp. 203–212. [Google Scholar]
  24. Zhao, Z.; Liu, W.; Qian, Y.; Nie, L.; Yin, Y.; Zhang, Y. Identifying advisor-advisee relationships from co-author networks via a novel deep model. Inf. Sci. 2018, 466, 258–269. [Google Scholar] [CrossRef]
  25. Wang, W.; Liu, J.; Xia, F.; King, I.; Tong, H. Shifu: Deep learning based advisor-advisee relationship mining in scholarly big data. In Proceedings of the 26th International Conference on World Wide Web Companion, Perth, Australia, 3–7 April 2017; pp. 303–310. [Google Scholar]
  26. Cialdini, R.B.; Goldstein, N.J. Social influence: Compliance and conformity. Annu. Rev. Psychol. 2004, 55, 591–621. [Google Scholar] [CrossRef] [PubMed]
  27. McPherson, M.; Smith-Lovin, L.; Cook, J.M. Birds of a feather: Homophily in social networks. Annu. Rev. Sociol. 2001, 27, 415–444. [Google Scholar] [CrossRef]
  28. Ma, H.; Yang, H.; Lyu, M.R.; King, I. Sorec: Social recommendation using probabilistic matrix factorization. In Proceedings of the 17th ACM Conference on Information and Knowledge Management, Napa Valley, CA, USA, 26–30 October 2008; pp. 931–940. [Google Scholar]
  29. Ma, H.; Zhou, D.; Liu, C.; Lyu, M.R.; King, I. Recommender systems with social regularization. In Proceedings of the Fourth ACM International Conference on Web Search and Data Mining, Hong Kong, China, 9–12 February 2011; pp. 287–296. [Google Scholar]
  30. Yang, B.; Lei, Y.; Liu, J.; Li, W. Social collaborative filtering by trust. IEEE Trans. Pattern Anal. Mach. Intell. 2016, 39, 1633–1647. [Google Scholar] [CrossRef]
  31. Yin, H.; Wang, Q.; Zheng, K.; Li, Z.; Yang, J.; Zhou, X. Social influence-based group representation learning for group recommendation. In Proceedings of the 2019 IEEE 35th International Conference on Data Engineering (ICDE), Macao, China, 8–11 April 2019; IEEE: Piscataway, NJ, USA, 2019; pp. 566–577. [Google Scholar]
  32. Yin, H.; Wang, Q.; Zheng, K.; Li, Z.; Zhou, X. Overcoming data sparsity in group recommendation. IEEE Trans. Knowl. Data Eng. 2020, 34, 3447–3460. [Google Scholar] [CrossRef]
  33. Zhou, J.; Cui, G.; Hu, S.; Zhang, Z.; Yang, C.; Liu, Z.; Wang, L.; Li, C.; Sun, M. Graph neural networks: A review of methods and applications. AI Open 2020, 1, 57–81. [Google Scholar] [CrossRef]
  34. Fan, W.; Ma, Y.; Li, Q.; He, Y.; Zhao, E.; Tang, J.; Yin, D. Graph neural networks for social recommendation. In Proceedings of the The World Wide Web Conference, San Francisco, CA, USA, 13–17 May 2019; pp. 417–426. [Google Scholar]
  35. Wu, Q.; Zhang, H.; Gao, X.; He, P.; Weng, P.; Gao, H.; Chen, G. Dual graph attention networks for deep latent representation of multifaceted social effects in recommender systems. In Proceedings of the The World Wide Web Conference, San Francisco, CA, USA, 13–17 May 2019; pp. 2091–2102. [Google Scholar]
  36. Song, W.; Xiao, Z.; Wang, Y.; Charlin, L.; Zhang, M.; Tang, J. Session-based social recommendation via dynamic graph attention networks. In Proceedings of the Twelfth ACM International Conference on Web Search and Data Mining, Melbourne, Australia, 11–15 February 2019; pp. 555–563. [Google Scholar]
  37. Wang, W.; Liu, J.; Yu, S.; Zhang, C.; Xu, Z.; Xia, F. Mining advisor-advisee relationships in scholarly big data: A deep learning approach. In Proceedings of the 16th ACM/IEEE-CS on Joint Conference on Digital Libraries, Newark, NJ, USA, 19–23 June 2016; pp. 209–210. [Google Scholar]
  38. Chen, T.; Kornblith, S.; Norouzi, M.; Hinton, G. A simple framework for contrastive learning of visual representations. In Proceedings of the International Conference on Machine Learning, PMLR, Virtual Event, 13–18 July 2020; pp. 1597–1607. [Google Scholar]
  39. He, K.; Fan, H.; Wu, Y.; Xie, S.; Girshick, R. Momentum contrast for unsupervised visual representation learning. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 14–19 June 2020; pp. 9729–9738. [Google Scholar]
  40. Veličković, P.; Fedus, W.; Hamilton, W.L.; Liò, P.; Bengio, Y.; Hjelm, R.D. Deep graph infomax. arXiv 2018, arXiv:1809.10341. [Google Scholar]
  41. Peng, Z.; Huang, W.; Luo, M.; Zheng, Q.; Rong, Y.; Xu, T.; Huang, J. Graph representation learning via graphical mutual information maximization. In Proceedings of the Web Conference 2020, Taipei, Taiwan, 20–24 April 2020; pp. 259–270. [Google Scholar]
  42. Hassani, K.; Khasahmadi, A.H. Contrastive multi-view representation learning on graphs. In Proceedings of the International Conference on Machine Learning, PMLR, Virtual, 13–18 July 2020; pp. 4116–4126. [Google Scholar]
  43. Park, C.; Kim, D.; Han, J.; Yu, H. Unsupervised attributed multiplex network embedding. Aaai Conf. Artif. Intell. 2020, 34, 5371–5378. [Google Scholar] [CrossRef]
  44. Dunbar, R. How Many Friends Does One Person Need? Dunbar’S Number and Other Evolutionary Quirks; Harvard University Press: Cambridge, MA, USA, 2010. [Google Scholar]
  45. Qin, D.; Zhou, X.; Chen, L.; Huang, G.; Zhang, Y. Dynamic connection-based social group recommendation. IEEE Trans. Knowl. Data Eng. 2018, 32, 453–467. [Google Scholar] [CrossRef]
  46. Tang, J.; Zhang, J.; Yao, L.; Li, J.; Zhang, L.; Su, Z. Arnetminer: Extraction and mining of academic social networks. In Proceedings of the 14th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, Las Vegas, NV, USA, 24–27 August 2008; pp. 990–998. [Google Scholar]
  47. Liénard, J.F.; Achakulvisut, T.; Acuna, D.E.; David, S.V. Intellectual synthesis in mentorship determines success in academic careers. Nat. Commun. 2018, 9, 4840. [Google Scholar] [CrossRef] [PubMed]
  48. Iglič, H.; Doreian, P.; Kronegger, L.; Ferligoj, A. With whom do researchers collaborate and why? Scientometrics 2017, 112, 153–174. [Google Scholar] [CrossRef] [PubMed]
  49. Montoya, F.G.; Alcayde, A.; Baños, R.; Manzano-Agugliaro, F. A fast method for identifying worldwide scientific collaborations using the Scopus database. Telemat. Inform. 2018, 35, 168–185. [Google Scholar] [CrossRef]
  50. Perozzi, B.; Al-Rfou, R.; Skiena, S. Deepwalk: Online learning of social representations. In Proceedings of the 20th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, New York, NY, USA, 24–27 August 2014; pp. 701–710. [Google Scholar]
  51. Grover, A.; Leskovec, J. node2vec: Scalable feature learning for networks. In Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, San Francisco, CA, USA, 13–17 August 2016; pp. 855–864. [Google Scholar]
  52. Tu, C.; Zhang, Z.; Liu, Z.; Sun, M. TransNet: Translation-Based Network Representation Learning for Social Relation Extraction. In Proceedings of the IJCAI, Melbourne, Australia, 19–25 August 2017; pp. 2864–2870. [Google Scholar]
  53. Liu, J.; Xia, F.; Wang, L.; Xu, B.; Kong, X.; Tong, H.; King, I. Shifu2: A network representation learning based model for advisor-advisee relationship mining. IEEE Trans. Knowl. Data Eng. 2019, 33, 1763–1777. [Google Scholar] [CrossRef]
  54. Rendle, S.; Freudenthaler, C.; Gantner, Z.; Schmidt-Thieme, L. BPR: Bayesian personalized ranking from implicit feedback. arXiv 2012, arXiv:1205.2618. [Google Scholar]
  55. Zhao, T.; McAuley, J.; King, I. Leveraging social connections to improve personalized ranking for collaborative filtering. In Proceedings of the 23rd ACM International Conference on Conference on Information and Knowledge Management, Shanghai, China, 3–7 November 2014; pp. 261–270. [Google Scholar]
  56. He, X.; Liao, L.; Zhang, H.; Nie, L.; Hu, X.; Chua, T.S. Neural collaborative filtering. In Proceedings of the 26th International Conference on World Wide Web, Perth, Australia, 3–7 April 2017; pp. 173–182. [Google Scholar]
  57. Wang, C.; Blei, D.M. Collaborative topic modeling for recommending scientific articles. In Proceedings of the 17th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, San Diego, CA, USA, 21–24 August 2011; pp. 448–456. [Google Scholar]
  58. Cai, T.; Cheng, H.; Luo, J.; Zhou, S. An efficient and simple graph model for scientific article cold start recommendation. In Proceedings of the Conceptual Modeling: 35th International Conference, ER 2016, Gifu, Japan, 14–17 November 2016; Proceedings 35. Springer: Berlin/Heidelberg, Germany, 2016; pp. 248–259. [Google Scholar]
  59. He, X.; Deng, K.; Wang, X.; Li, Y.; Zhang, Y.; Wang, M. Lightgcn: Simplifying and powering graph convolution network for recommendation. In Proceedings of the 43rd International ACM SIGIR Conference on Research and Development in Information Retrieval, Xi’an, China, 25–30 July 2020; pp. 639–648. [Google Scholar]
  60. Ji, S.; Feng, Y.; Ji, R.; Zhao, X.; Tang, W.; Gao, Y. Dual channel hypergraph collaborative filtering. In Proceedings of the 26th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining, Virtual Event, CA, USA, 6–10 July 2020; pp. 2020–2029. [Google Scholar]
  61. Wu, L.; Li, J.; Sun, P.; Hong, R.; Ge, Y.; Wang, M. Diffnet++: A neural influence and interest diffusion network for social recommendation. IEEE Trans. Knowl. Data Eng. 2020, 34, 4753–4766. [Google Scholar] [CrossRef]
Figure 1. The overall framework for constructing students’ multiple academic social relationship networks.
Figure 1. The overall framework for constructing students’ multiple academic social relationship networks.
Mathematics 12 01667 g001
Figure 2. The self-supervised academic paper social recommendation framework for students: an overview of constructing multifaceted academic social relationships among students.
Figure 2. The self-supervised academic paper social recommendation framework for students: an overview of constructing multifaceted academic social relationships among students.
Mathematics 12 01667 g002
Figure 3. Mentorship relationship identification performance comparison (compared with network representation learning methods).
Figure 3. Mentorship relationship identification performance comparison (compared with network representation learning methods).
Mathematics 12 01667 g003aMathematics 12 01667 g003b
Figure 4. Parameter sensitivity experiment of advisor–advisee relationship recognition models.
Figure 4. Parameter sensitivity experiment of advisor–advisee relationship recognition models.
Mathematics 12 01667 g004
Figure 5. Mentorship relationship identification performance comparison (compared with network representation learning methods): (a) Parameter sensitivity to δ ; (b) Effect of embedding dimensionality on SSRES model performance; (c) Effect of learning rate on SSRES model performance.
Figure 5. Mentorship relationship identification performance comparison (compared with network representation learning methods): (a) Parameter sensitivity to δ ; (b) Effect of embedding dimensionality on SSRES model performance; (c) Effect of learning rate on SSRES model performance.
Mathematics 12 01667 g005
Table 1. Node attribute feature descriptions.
Table 1. Node attribute feature descriptions.
Feature LabelDescription
a a i Academic tenure of scholar i
p n i Total publications of scholar i
c n i Cumulative citation count of scholar i
h i i H-index of scholar i
s a i Affiliated institution of scholar i
s r i Research focus of scholar i
Table 2. Description of collaborative edge attribute features.
Table 2. Description of collaborative edge attribute features.
Feature LabelDescription
a c i j Scholar i’s academic age during collaboration with co-author j
a g i j Age disparity in academia between scholar i and co-author j
t b i j Total scholarly outputs by scholar i prior to collaboration with j
c t i j Frequency of collaborative works between scholar i and j
d t i j Temporal span of collaboration between scholar i and co-author j
t f s i j Incidences scholar i and j have occupied first and second author positions
t f l i j Incidences scholar i and j have been the first and last authors
k u l c i j Collaborative similarity index between scholar i and j
Table 3. Statistical data for select domains in the AFT dataset.
Table 3. Statistical data for select domains in the AFT dataset.
Research FieldNumber of ScholarsNumber of Advisor–Advisee Relationship Relations
Physics54,91658,875
Chemistry117,000136,536
Neuroscience142,531170,879
Education56,96649,925
Sociology23,42918,413
Economics22,08818,414
Anthropology11,68810,280
Microbiology17,29714,861
Nursing13,13810,764
Political Science16,42813,539
Literature26,01219,713
Computer Science23,67818,866
Theology15,44312,754
Mathematics35,55129,767
Evolutionary Biology14,57418,417
Table 4. Statistical data for the Aminer and AFT dataset matching results in select domains.
Table 4. Statistical data for the Aminer and AFT dataset matching results in select domains.
Research FieldNumber of Advisor–Advisee Relationship PairsTime Range
Computer Science10,6522000–2015
Neuroscience50282000–2015
Mathematics47732000–2015
Chemistry29572000–2015
Physics14522000–2015
Education10292000–2015
Table 5. Comparison of advisor–advisee relationship identification performance.
Table 5. Comparison of advisor–advisee relationship identification performance.
Computer ScienceNeuroscience
MethodsNTARMShifuShifu2MethodsNTARMShifuShifu2
MetricsMetrics
Precision0.8550.7930.814Precision0.8330.7570.786
Recall0.8930.8240.880Recall0.8770.7970.863
Accuracy0.8710.8040.839Accuracy0.8510.7710.823
F1-score0.8740.8080.845F1-score0.8550.7760.814
MathematicsChemistry
MethodsNTARMShifuShifu2MethodsNTARMShifuShifu2
MetricsMetrics
Precision0.8450.7750.817Precision0.8630.7860.814
Recall0.8900.8190.882Recall0.8950.8460.867
Accuracy0.8630.7910.842Accuracy0.8760.8080.834
F1-score0.8670.7970.849F1-score0.8790.8150.839
Table 6. Statistical data of the experimental sub-dataset.
Table 6. Statistical data of the experimental sub-dataset.
Data CategoryAmount
Number of Students10,929
Number of Papers89,013
Number of Ratings281,319
Number of Social Relations90,686
Rating Density0.0289%
Social Density0.15240%
Table 7. Comparison of paper recommendation performance.
Table 7. Comparison of paper recommendation performance.
MethodsPrecision@10Recall@10NDCG@10
Metrics
BPR0.01010.01640.0134
SBPR0.03650.08770.0779
NeuMF0.02720.07750.0676
CTR0.04250.11480.1021
UAGMT0.05650.12170.1162
LightGCN0.05440.10650.1056
DHCF0.05290.10800.1062
DiffNet++0.05670.12820.1181
SSRES0.05860.13870.1251
Table 8. Ablation study results.
Table 8. Ablation study results.
ModelPrecision@10Recall@10NDCG@10
SSRES-HE0.05490.12610.1183
SSRES-AR0.05190.12550.1136
SSRES-SP0.04870.11940.1088
SSRES0.05860.13870.1251
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Guo, Y.; Zhou, Z. SSRES: A Student Academic Paper Social Recommendation Model Based on a Heterogeneous Graph Approach. Mathematics 2024, 12, 1667. https://doi.org/10.3390/math12111667

AMA Style

Guo Y, Zhou Z. SSRES: A Student Academic Paper Social Recommendation Model Based on a Heterogeneous Graph Approach. Mathematics. 2024; 12(11):1667. https://doi.org/10.3390/math12111667

Chicago/Turabian Style

Guo, Yiyang, and Zheyu Zhou. 2024. "SSRES: A Student Academic Paper Social Recommendation Model Based on a Heterogeneous Graph Approach" Mathematics 12, no. 11: 1667. https://doi.org/10.3390/math12111667

APA Style

Guo, Y., & Zhou, Z. (2024). SSRES: A Student Academic Paper Social Recommendation Model Based on a Heterogeneous Graph Approach. Mathematics, 12(11), 1667. https://doi.org/10.3390/math12111667

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop