In this section, we provide a detailed overview of the design of our blockchain-based system for sharing machine learning models. We first discuss our approach for incentivizing users to provide honest feedback on models, as opposed to maliciously leaving negative comments. Next, we propose solutions for motivating model owners to share high-quality models.
4.1. Honest Evaluation Incentive Model
The honest evaluation incentive stage can be divided into three steps. The first step is to collect user evaluations, and when the number of evaluations reaches a threshold, the system proceeds to the second step. In the second step, the collected evaluations are classified into honest and malicious evaluations, and the model quality is calculated using the honest evaluations. In the third step, the system calculates the user’s reputation value based on the deviation between the model quality and the user’s evaluation. In our proposed system, model users may either provide honest or malicious feedback. Due to the anonymity and lack of trust in blockchain-based systems, the possibility of collusion between model owners and model users is low, thereby reducing the likelihood of malicious praise. However, model users may also be potential model owners themselves, and therefore, there is a high probability of providing poor evaluations for other models in order to improve the competitiveness of their own models. Therefore, while considering the evaluation behaviors of malicious model users, we only take into account malicious negative comments and we do not consider malicious favorable comments for the time being. In general, the evaluation data are expected to follow a normal distribution [
44]. For the model quality evaluation phase, we consider two possible states of the model users—honest or malicious—and treat the evaluation data as a combination of data that follow a normal distribution with different parameters. We use the Gaussian mixture model to fit the evaluation data. The problem of determining the model quality is then transformed into finding the expectation of the normal distribution that fits the honest evaluation data, which can be abstracted as a parameter estimation problem of the Gaussian mixture model. We use the EM algorithm to classify honest evaluations and malicious negative comments, and we use the mean of the honest evaluations as the model quality evaluation result. In the honest evaluation incentive phase, we use reputation as the incentive carrier and we measure the honesty of the evaluation behavior by calculating the deviation between the model user’s evaluation and the model evaluation result. We design an evaluation incentive function based on the sigmoid-prime function to provide feedback and targeted incentives for the model user’s evaluation behavior. The sigmoid function, also known as the logistic function, is a mathematical function that maps input values to a range of between 0 and 1. The sigmoid function has the property where it has the maximum slope at x = 0, meaning that the rate of change of the y-value is higher. As x increases or decreases, the rate of change of the y-value gradually decreases. Based on this characteristic, the reputation value of users undergoes significant changes in the initial state, but as the system stabilizes, the rate of change of the reputation value decreases. The more honest the model user is, and the closer the evaluation is to the evaluation result, the more reputation value they can obtain.
The proposed system estimates model quality based on honest evaluations, but it needs to determine whether the user’s evaluation is honest, since the system cannot identify it directly. Assuming that there are more honest model users than malicious users, the distribution with a higher mixing coefficient in the Gaussian mixture model can be considered as the distribution followed by the evaluation data of honest model users. Its parameters can represent the quality of models. The problem of estimating model quality based on comments can be transformed into a parameter problem of solving the Gaussian mixture model. The EM algorithm is a commonly used algorithm for solving the Gaussian mixture model. In the following section, we will describe how the EM algorithm is used to estimate the quality of models.
Let
be the set of models,
be the
i-th model, and
be the comments set of
. The
can be divided into two groups; the first group of comments are from honest model users and the second group of comments are from malicious model users. We denote the first group of comments as
; these comments fit the normal distribution
, whose mean value is
, and the variance is
. We denote the first group of comments as
; these comments fit the normal distribution
, whose mean value is
, and the variance is
and
. The length of
equals the number of
users, that is, each user can only evaluate a mode once, and must evaluate once. We denote the coefficients of
and
in the Gaussian mixture model as
,
, representing the proportion of data from
and
in the total data; we can obtain
. Generally, honest users are more common than malicious users, i.e.,
. The probability distribution function of the Gaussian mixture model can be expressed as:
In Equation (
1),
is a model parameter, and
.
is the
k-th normal distribution
.
We define hidden variables as variables that cannot be directly observed but can be derived from other variables. We do not know whether the user is honest; we can calculate the probability that the user belongs to each state. So, the state of the model user is a hidden variable. We define the state of a model user as
In order to reduce the computational complexity and to eliminate irrelevant variables, take the logarithm of the left and right sides of Equation (
3) at the same time and obtain the log-likelihood function of the complete data as follows:
The EM algorithm is an iterative algorithm, and the input of each iteration is the output of the previous iteration. Assuming that we are at the
th iteration and need to solve for the parameter values, we can use
to represent them. Based on the existing evaluation data
and parameter estimation value
, the conditional expectation of the complete data log-likelihood function
can be constructed as follows:
E represents the expectation of the user’s state under all data. The user’s state can only be honest or malicious, and the expression for expectation can be expanded as Equation (
6).
The expected value of the latent variable is equivalent to the conditional probability when the latent variable takes a value of 1, which is the probability of the user’s state, given that the parameters of the Gaussian mixture model and the evaluation data are known, and belongs to the posterior probability. According to Bayes’ theorem, the posterior probability can be expressed as Equation (
7).
where
represents the expected value of parameters after the m-th iteration. Substituting Equation (
7) into Equation (
5), the expected complete-data log-likelihood function
Q can be obtained.
Next, we move on to the M-step of the EM algorithm, where we calculate the parameter values
that corresponds to the maximum of the expected log-likelihood function. We calculate the partial derivatives of
with respect to the parameters
,
, and
, set them to zero, and solve for each parameter. Letting
, the new parameter
,
,
can be expressed as follows:
If the expectation of the log-likelihood function does not converge, we update the parameters
, and
. We iterate the E and M steps to update the expectation of the log-likelihood function
until it converges or reaches the maximum number of iterations, and output the current parameter value
. Based on the assumption that there are more honest learners than malicious ones,
corresponding to
is the evaluation result of the model quality, denoted as
.
If users cannot obtain any reward after evaluating the model, their willingness to evaluate will decrease, and malicious or arbitrary evaluation behaviors will emerge, which will increase the difficulty and error of quality evaluation, reduce the credibility of the system, and be unfavorable for the sustainable development of the sharing ecosystem. Therefore, designing reasonable and effective incentive functions is of great significance. As an attribute of users, reputation is related to the cost of purchasing models, and it can also be used as the weight of user evaluation data to measure the credibility of evaluation. Therefore, using reputation to provide targeted incentives to users, and designing an honest evaluation incentive function is necessary. The function should satisfy two conditions: motivate users to actively participate in evaluations, and motivate users to provide honest evaluations.
Firstly, we evaluate the honesty level of the user using the evaluation bias. The evaluation bias of is defined as the distance between the evaluation and the quality evaluation result , denoted as . The smaller the evaluation bias, the more honest the user is, and the more rewards they will receive. Evaluations that are higher or lower than by the same amount are considered to have the same level of honesty. We assume that the user’s evaluation of models is a real number within the range of [0, 5], and that the evaluation bias is also a real number within the range of [0, 5].
Secondly, the incentive function related to the honesty level is defined to measure the reputation value obtained by users participating in the evaluation in a quality evaluation process. The user’s total reputation value is the cumulative result of the incentive received from a single evaluation. When reputation is used to exchange system models, the total reputation value will decrease correspondingly. To incentivize users to actively and honestly evaluate, the function should satisfy the following two properties.
Property 1. For any participating , always holds. According to the law of large numbers, when a large number of users participates in the evaluation, the expected evaluation is closer to the model quality. However, evaluations incur costs for users. In order to encourage users to actively participate in the model evaluation process, the evaluation incentive function needs to satisfy the participation constraint. The participation constraint, also known as the individual rationality constraint, in the context of teaching quality evaluation, refers to the expected benefit of participating in the evaluation being not less than the expected benefit of not participating in the evaluation. When users do not participate in the evaluation, their reputation gain is 0, and so satisfying the participation constraint is equivalent to the evaluation incentive function .
Property 2. For any two users and , who evaluate , if , then . The incentive function for evaluation needs to satisfy the incentive compatibility constraint, which means that the behaviors of individuals who pursue the maximization of their own interests should be consistent with the goal of maximizing the collective interests. Incentivizing users to provide honest evaluations is a goal that maximizes collective interests, as it helps to reduce the error of model quality evaluation, and improves the authenticity and accuracy of quality evaluation results. To satisfy the incentive compatibility principle, under the induction of the evaluation incentive function, honest evaluation should be the optimal strategy for users. For the same model, the more honest the user is, the smaller the evaluation deviation, and the more rewards they receive.
Based on the above two properties, a sigmoid-prime function is introduced as the incentive function, with the evaluation deviation Dev as the independent variable. According to Skinner’s reinforcement theory, the ratio of reputation value to evaluation deviation is not a fixed value, which is conducive to delaying the trend of the decay of the honest evaluation behaviors of users. The sigmoid-prime function is the derivative of the sigmoid function, and the gradient increases and then decreases. Using it as the evaluation incentive function is conducive to long-term incentives for users to make honest evaluations. The evaluation incentive function
can be expressed as:
where k and b are parameters of the incentive function that affect the range and strength of incentives, as detailed in
Section 5.2 and set by the educational institution deploying the EIC. The function passes through the point (0, b/4), and when the user’s evaluation value is equal to the model quality assessment result, the user receives the maximum reputation value of b/4. The incentive function varies with changes in the user’s evaluation and evaluation bias as follows.
Let denote the reputation score of , and let be the initial reputation score of the user. The change in the reputation score depends on the user’s behavior. Actively participating in evaluation will increase the reputation score, while using reputation score and points to exchange for other models will decrease the reputation score. In addition, as a user’s attribute, reputation can be used to indicate the credibility of the user’s evaluation. The higher the reputation score, the more honest the user is in the history of evaluations, and the more likely the user will remain honest in future evaluations, indicating a higher evaluation credibility. Therefore, reputation can be used as the weight of user evaluations to calculate the weighted expectation of evaluations, assigned to the EM algorithm as the initial model parameters. In summary, a reputation-based honest evaluation incentive function is proposed based on the sigmoid-prime function as a prototype, and the user’s evaluation bias as the independent variable. On the one hand, it provides feedback to users on their evaluations, and on the other hand, it encourages them to actively and honestly evaluate, constraining the evaluation behavior and promoting a good evaluation atmosphere.
4.2. Model Sharing Incentive Model
Model sharing can improve the utilization rate of models, reduce the repeated training of models, and benefit more companies and individuals. However, training a high-precision model requires a lot of data and consumes a lot of computing models. In addition, there may be competition among them. In order to protect their own interests, many model owners are unwilling to share their models. It is particularly important to establish a good incentive mechanism to encourage model owners to share their own models.
A model owner can improve the accuracy of his own model by sharing the model with others, so the model owners will benefit from each other, in addition to competition. A large number of repetitive and inferior models will increase the difficulty of obtaining information for model users, and increase the storage burden of the system.
Inspired by the above, we propose a model sharing incentive model based on evolutionary game. We first introduce the overall model, and then we design an incentive function based on the sigmoid function by integrating quality and similarity. Finally, we construct an evolutionary game model among model owners, and we analyze the impact of incentive function on model sharers’ strategy selection.
First, the model owner views the incentive information of the shared model through the platform and decides whether to share his own model. If they choose to share the model and upload it to the platform, the platform will calculate the similarity between the model and the existing model based on the description information of the model. The model users can search, download, and evaluate the model. The platform will evaluate the quality of the model after receiving the evaluation. Finally, the platform calculates the results of the incentive function according to the model quality, similarity, and parameters of the incentive function, and rewards the model sharers.
4.3. Incentive Function Design
Improving the sense of value and identity of model owners and giving appropriate incentives is the key to creating a good sharing atmosphere. The use of an integral system can encourage model owners to share models with high quality and low similarity, and realize the long-term development of the model sharing system. Model owners can use points to download the other models in the platform.
Model quality is an important factor affecting the development of the model sharing system. However, the time cost and opportunity cost of high-quality models are higher in the production process. If the quality factor is ignored and the same degree of incentive is given to users who share different quality models, model owners will be guided to share low-quality models. When the quality of most models in the system cannot satisfy the requirements of the demanders, the attractiveness of the model sharing platform to users will be greatly reduced, and sharing will lose its value.
Model repeatability is another important factor affecting the development of the model sharing system. There is a common problem of the repeated construction of models in the existing sharing system, which on the one hand increases the burden for model demanders to retrieve and browse models, and on the other hand, increases the storage pressure of the sharing system.
In order to solve the problems of poor model quality and high repeatability in the shared system, we design an incentive function I (Q, sim) related to model quality Q and similarity sim, which should satisfy the following properties. Among them, the model quality Q is calculated using the quality evaluation function in
Section 4.1. The similarity sim is the ratio of the function of the repetitive part of the model to the total function of the model in the sharing system, which is calculated using the smart contract.
Theory 1. Function I is a monotone increasing function about Q and a monotone decreasing function about sim. When the similarity is the same, the higher the quality of the teaching models is, the more incentives will be obtained. When the quality is the same, the higher the similarity of the model, the more repetitive parts of the model and existing models in the system. Even if it is shared into the system, the less benefits it brings to the system as a whole and to the model demanders, and so the less incentive it obtains.
Theory 2. If all sharing behaviors can gain benefits, there will usually be malicious users who gain benefits without labor, such as sharing blank files, meaningless models, or copying existing models in the system. In order to prevent this phenomenon, users cannot obtain any incentive when the quality of the shared model is 0 or the similarity is 1.
Theory 3. Let
;
denotes the impact of model similarity on incentives. We can obtain the following conclusions. means the proportion of the incentives actually received by the model owner in the deserved incentives. The revenue from sharing a certain model should be less than or equal to the revenue from sharing the original model with the same accuracy and function. In extreme cases, when users download models from the system and re-upload them to the system, no matter what the quality of the models are, users will not benefit from them. When users share completely original models, the benefits obtained are only related to the accuracy of model. So, , and the closer is to 0, the closer is to 1.
Theory 4. With the increase in ,
the trend of is as follows: It is assumed that the tolerance of model users to model similarity is nonlinear. When the similarity is small, the users are more tolerant of models, and the change of the incentive coefficient is small. When the similarity is large, the information gain brought by models to users is small, and the change of the incentive coefficient is small. With the increase in similarity, the incentive coefficient decreases, and the decreasing rate first increases and then decreases.
If functions satisfy the above properties, they can be used as incentive functions. We choose the sigmoid function, which is often used in machine learning as the excitation function. Because the sigmoid function realizes the nonlinear mapping of input data to (0, 1), the greater the absolute value of input data, the smaller the gradient and the lower the sensitivity. This change trend is shown in
Figure 2.
We can obtain the
function by converting the sigmoid function.
The
is the threshold of similarity, and the tolerance of model demanders to model similarity is relatively sensitive near the threshold. A (a > 0) affects the decreasing speed of the excitation function. When the similarity is between 0 and 1, the trend of incentive coefficient changing with the similarity is shown in
Figure 3:
Then, we can obtain
as follows:
indicates the excitation parameter. For specific excitation function I, when , , and the incentive function is a monotonically increasing function of model quality; , the incentive function is a monotonically decreasing function of model similarity. The function satisfies Property 1. ; the function satisfies Theory 2. According to the figure, the function satisfies Theories 3 and 4. The function satisfies the above three theories and can be used to encourage model owners to share high-quality and low-homogeneity models.
4.4. Evolutionary Game Model
The model owner does not fully know other owners’ private information (such as the quality and similarity of the model owned), the selection strategy, the utility function, and other information. The computing power to derive the optimal strategy based on the collected information is limited. Therefore, model owners belong to individuals with limited rationality. In the multiple rounds of the game, model owners can choose strategies to maximize their own interests, through learning and correcting mistakes. From the perspective of the evolutionary game, this paper analyzes the strategic preferences of model owners and the state of the group when it finally reaches dynamic equilibrium.
Assumption 1. N is a game group composed of a large number of participants with models, which can be expressed as . Assuming that each model owner is bounded and independent, there is no conspiracy between model owners.
Assumption 2. S is the strategic space of the evolutionary game. is the set of strategies, and is the selected strategy in a certain round of the game. Each model owner has two strategies: participating in sharing models and not participating in sharing models, namely, . Define as the combination of strategies of other owner except the owner i. In each round of game, model owners interact with each other through repeated random matching. The evolution process conforms to natural laws, and the final group will reach a dynamic and stable state. If at a certain stage, the proportion of model owners who choose strategy si in the group is pi, the evolutionary game strategy can be expressed as , and the state of the group can be expressed as , where and .
Assumption 3. U is a utility function that represents the mapping from the strategy combination to the reward; . Given a strategy combination such as , the utility function represents the benefit of model owner i when i selects strategy and the strategy set selected by others other than i is .
Assumption 4. In the initial state, the model owner has a certain preference for strategy selection. Randomly select two model owners as game subjects, representing two groups. Assuming that the probabilities of both sides of the game choosing the sharing strategy are x and y, respectively, the probabilities of choosing not to participate in the sharing strategy are , , , , respectively. X is a function of time t; , . According to the assumption of inertia behavior, the proportions of people who choose to participate in the strategy in the two groups are x and y, respectively.
Assumption 5. Model accuracy and model quantity are the private information of the model owner. Before the collective makes a choice, this information is only known to the model owner. The guesses of other model owners about private information satisfy the common prior probability distribution. In addition, users have the same level of information about the model.
Assumption 6. The knowledge production function is introduced to measure the value of the model. The machine learning model not only has the function of prediction, but also has the value of inspiration for the re-creation of the model. Model users can gain knowledge by accessing the shared model, and combine the newly learned knowledge with the original model to create a new model. This process is called knowledge spillover.
There are similarities between knowledge spillover and the production process. Economics uses a production function to describe the relationship between the input factors and the output in the production process. Greitz introduced the production function into the research of knowledge sharing for the first time, and put forward the concept of the knowledge production function [
45]. Since then, the knowledge production function has been widely used to study the relationship between the knowledge output and input factors. Among them, the Cobb-Douglas function is widely used to predict production and to analyze the development path because of its simple model and convenient calculation in parameter estimation. Therefore, the Griliches-Jaffe knowledge production function based on the Cobb Douglas function is introduced to measure the knowledge outputs of teaching models in the sharing process. The calculation method is
, wherein,
Q represents the benefits brought about by knowledge output, namely, the knowledge value.
is a coefficient, which is related to technical level, teaching model type, field, productivity, and other factors.
K is the model stock, expressed by the product of the model scale and quality.
L is the number of personnel.
represents the elasticity coefficient of the knowledge stock and the number of inputs, and it represents the impact of the two factors on the knowledge output.
When no one else shares the model in the system, the model stock of model owner i, where is the data volume of the model and is the model quality. When other people share the model in the system, the model stock of model owner i. is the amount of data available in the shared model, that is, the amount of data complementary to the existing model, and is the quality of the shared model. Let the number of personnel invested be 1, and the additional benefits obtained by model owner i from the system model can be expressed as .
Other parameters used in the evolutionary game are shown in
Table 1:
Based on the above assumptions, the corresponding returns of game players 1 and 2 in different strategy choices are shown in
Table 2:
Each subject has two strategic choices. There are four possible outcomes of the game, which can be summarized into the following three situations:
Case 1: When the game subjects choose not to share the strategy, neither party can access the other party’s models, gain shared benefits, and pay no costs. At this time, the income is 0.
Case 2: When one party in the game entity shares models while the other party does not, it corresponds to or ) in the payment matrix. Most incentive mechanisms believe that “the party who chooses not to share cannot obtain the models of the sharing party” . However, realizing model sharing is the original intention of the construction of teaching the model sharing platform. Therefore, on the basis of the article, EduShare allows users to access models without sharing, but they need to pay a certain amount of points to the system. At this time, the model owners who choose to share cannot obtain the desired models from the other party, so they cannot obtain additional benefits, but they need to bear the losses caused by sharing. For the party who does not share the models, it is necessary to pay the system the credit c to access the models and to obtain additional benefits from the models. Therefore, when player 1 chooses P and player 2 chooses Non, the income of player 1 is , The income of entity 2 is . When player 1 chooses Non and player 2 chooses P, the income of player 1 is , and the income of player 2 is .
Case 3: When both sides of the game choose to share models, the model owner can not only visit the other side’s models, but also understand the evaluations of others on their own models, gain more recognition, and provide direction and motivation for improving models. At this time, the benefits of model owners from shared models can be expressed as . Among them, is the sharing income coefficient. In addition, the model owner will also receive the sharing incentive issued by the system and bear the losses caused by sharing. Therefore, when all players choose to share, the benefit of player 1 is . The income of entity 2 is .
4.5. Game Process Analysis
This section analyzes the change trend of the model owner’s choice preference by solving the equilibrium strategy of the evolutionary game model. The equilibrium strategy for solving the evolutionary game model can be transformed into solving the fixed point problem of the dynamic system. Since the evolution direction is related to the expected reward of the players in different strategy choices, the expected reward and the average expected reward when the players choose to share and not share are analyzed, and the dynamic system trajectory expression is constructed as follows.
When player 1 chooses the sharing strategy, the expected reward consists of two parts. If game player 2 chooses the shared strategy, then player 1 gains
. If player 2 chooses not to share the model, the income obtained by player 1 is
. According to the assumptions, the probabilities of player 2 choosing the two strategies are
y and
, respectively. For player 1, the expected income
corresponding to the strategy selected to participate in the sharing is:
In the same way, the expected income
of the non-shared strategy can be expressed as:
Then, the average expected income
can be expressed as:
For player 2, the strategies of sharing, not-sharing, and the average expected returns,
,
, and
are, respectively:
Assume that in the game group 1,
is the proportion of the individuals who choose the sharing strategy in the next round in the group. Use the finite difference form to express the dynamic change of the group strategy, then
can be expressed as
. The change of the proportion of individuals who choose this strategy in the population per unit time can be expressed as:
The trajectory and fixed point of the differential equation are equivalent to the following equation, and we can build the replication dynamic equation of the evolutionary game.
Similarly, the replication dynamic equation of game group 2 can be expressed as:
The replication dynamic equation reveals the evolution trend of the strategy of participation and sharing in the group. When the expected return of selecting shared models is greater than the average expected return, the probability of selecting this strategy in the next round of the game will increase, e.g.,
,
. On the contrary, the probability of selecting the sharing strategy will be reduced, e.g.,
,
. When R
and
, the evolutionary game reaches the equilibrium point. Therefore, there are five equilibrium points in the model:
Because the equilibrium of the evolutionary game is dynamic, the equilibrium point may not be stable. Next, we analyze the local stability of the five equilibrium points. According to evolutionary game theory, the linearization theorem is introduced to judge the stability of the fixed point of the dynamic system, which is mainly divided into the following three steps.
Step 1: According to the replication dynamic equation, the Jacobian matrix is constructed. Calculate the partial derivatives of
x and
y for
and
, respectively, to obtain the Jacobian matrix as follows:
Step 2: Calculate the determinant value of the Jacobian matrix and the trace of the matrix. The determinant Det.J of the Jacobian matrix can be expressed as:
The trace
of the Jacobian matrix can be expressed as:
Step 3: Judge the stability of the equilibrium point of the model according to the linearization theorem. According to the linearization theorem, the equilibrium point is stable if and only if the determinant of the Jacobian matrix is greater than 0 and the trace of the matrix is less than 0, otherwise the equilibrium point is not stable. As the number of games increases, the equilibrium point will eventually tend to evolve into a stable strategy. The determinant and trace of the Jacobi matrix corresponding to the five equilibrium points are as follows (
Table 3):
The status of the equilibrium point is discussed according to the parameter values, as follows:
Case 1: When
and
, the status of the equilibrium point is shown in
Table 4. At this point, (1,1) is an evolutionary stability strategy. Regardless of the initial willingness of model owners to share, the group will eventually tend to share teaching models. At this time, the incentive function plays a strong incentive role, and the strategy selection of individual benefit maximization is consistent with the strategy selection of group benefit maximization, satisfying the incentive compatibility constraint.
Case 2: When
and
, the status of the equilibrium point is shown in
Table 5. At this time (0,0) is an evolutionary stability strategy. Regardless of the initial willingness of model owners to share, the group will eventually tend to not share models. The incentive mechanism satisfying this condition is invalid.
Case 3: When
and
, the equilibrium point status is shown in
Table 6. At this time, (0,0) and (1,1) are evolutionary stability strategies, and the evolution direction is related to the equilibrium point. When the proportion of model owners selected to share in game subject 1 x > 1, game subject 2 evolves towards sharing. When
, game subject 2 evolves towards non-sharing. When the proportion of shared model owners selected by game player 2 is
, game subject 1 evolves towards sharing. When
, the game subject 1 evolves towards non-sharing.
4.6. Dynamic Incentive for Model Sharing
Based on the above analysis, this model has different evolutionary outcomes when the incentive function satisfies different conditions. When the incentive function satisfies the conditions in Case 1, the model owner’s preference for strategy selection is sharing. When the incentive function satisfies the condition in Case 2, the model owner’s preference for strategy selection is not sharing. When the incentive function satisfies the conditions in Case 3, the evolution direction of the population, and the current state and the equilibrium solution .
The system uses the integral incentive model owner to share the model. When the integral in the system reaches the saturation state, the attraction of the integral to the model owner decreases, and the increase in the share rate brought by the unit integral decreases, which is called the phenomenon of the diminishing marginal effect. If the incentive is set to a fixed value, if the setting is too low, the incentive effect will not be achieved. If the setting is too high, the incentive saturation speed will be too fast and the marginal effect will decrease. Therefore, the dynamic adjustment of the incentive function can achieve a long-term incentive for model owners.
Based on the results of evolutionary game analysis, the incentive function is dynamically adjusted. Realizing model sharing is the goal of maximizing social benefits. Therefore, evolutionary stability strategy (1, 1) is the expected result of the game. Since the evolutionary stability strategy corresponding to case 2 is (0, 0), which does not satisfy the incentive compatibility principle, only case 1 and case 3 are considered. Under the initial conditions, the proportion of the model owners selected to share in the system is small, and the incentive function satisfying the conditions in Case 1 is used for the incentive. After a period of time, the excitation function satisfying the conditions in Case 3 is used to continue the excitation.
In order to make the shared models in the system have high quality, the model access threshold is designed to realize the constraint on the model owner. Assume that the minimum quality of models that the model demander can accept in the sharing system is qmin, and that the maximum model similarity that can be tolerated is simmax. The threshold of model access is recorded as T, then T shall satisfy:
This means that the system allows for high-quality model sharing with high similarity, and that it allows for low-quality but original model sharing. For example, the minimum acceptable model quality for model demanders is 2, and the maximum tolerable similarity corresponds to
; the threshold value is max(2, 2.5) = 2.5. After the introduction of T, the incentive function parameters of game subject i are set as shown in Equation (
24), where
is the proportion of model owners who choose to share with another subject in the game.
When the quality or similarity of models does not satisfy the system requirements, not sharing models is the best strategy choice for rational individuals. With the increase in the proportion of sharing strategies selected in the system, the dependence of the system on incentives gradually decreases to 0. After that, the system can still maintain a good sharing atmosphere without incentives.
Model owners can know the value of incentive parameters through smart contracts, and then decide whether to share. If the quality of models is high and the innovation is strong, and the expected return of model owners to share models is greater than the average expected return, they will choose to share models. If the model has low quality, high repeatability, and does not satisfy the system requirements, the model owner will gain less than the average income when choosing to share the model. At this time, model owners can choose to share models after improving the quality of models and reducing the repetition rate, or give up sharing models. In this way, we encourage the sharing of high-quality models, punish the sharing of inferior models, and realize the quality control of the models shared in the system.
When the model owner decides to share the model, the model will be uploaded to the local database of the educational institution, and the address summary, digital signature, detailed introduction, and parameter information of the current incentive function of the model will be stored in the blockchain together. The model quality is calculated using the quality evaluation function in
Section 4.1. The incentive for model owners to share models can be calculated using Equation (
11). By adjusting the parameters of the incentive function through Equation (
24), the dynamic incentive for the model owner is realized, which not only reduces the burden of the shared system, but also maintains the good operation of the system and delays the incentive saturation and marginal effect decline caused by excessive incentive.