Next Article in Journal
Some Non-Obvious Consequences of Non-Extensiveness of Entropy
Next Article in Special Issue
Asymptotic Expansion and Weak Approximation for a Stochastic Control Problem on Path Space
Previous Article in Journal
Airborne Quantum Key Distribution Performance Analysis under Supersonic Boundary Layer
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Stochastic Expectation Maximization Algorithm for Linear Mixed-Effects Model with Interactions in the Presence of Incomplete Data

1
Laboratoire de Mathématiques et Applications, Université de Poitiers, 11 Boulevard Marie et Pierre Curie, 86962 Futuroscope Chasseneuil, CEDEX 9, 86073 Poitiers, France
2
CeRCA-CNRS UMR 7295, Université de Poitiers, 5 rue T. Lefebvre, MSHS, CEDEX 9, 86073 Poitiers, France
*
Author to whom correspondence should be addressed.
Entropy 2023, 25(3), 473; https://doi.org/10.3390/e25030473
Submission received: 9 February 2023 / Revised: 27 February 2023 / Accepted: 2 March 2023 / Published: 8 March 2023
(This article belongs to the Special Issue Monte Carlo Simulation in Statistical Physics)

Abstract

:
The purpose of this paper is to propose a new algorithm based on stochastic expectation maximization (SEM) to deal with the problem of unobserved values when multiple interactions in a linear mixed-effects model (LMEM) are present. We test the effectiveness of the proposed algorithm with the stochastic approximation expectation maximization (SAEM) and Monte Carlo Markov chain (MCMC) algorithms. This comparison is implemented to highlight the importance of including the maximum effects that can affect the model. The applications are made on both simulated psychological and real data. The findings demonstrate that our proposed SEM algorithm is highly preferable to the other competitor algorithms.

1. Introduction

The time between the presentation of a stimulus and a participant’s motor response is the oldest and most widely used measure for exploring the functioning of the human mind. In 1869, ref. [1] theorized this duration, called reaction time (RT), as involving three sets of activities: perceptual mechanisms, cognitive processing and motor preparation. Based on the assumption that the first and last sets of processing can be considered as having virtually identical duration for the same task, any change in RTs between two experimental conditions is then interpreted as indicating a change in the duration of cognitive processing. RT is then considered by psychologists as a tool to explore cognitive processing mechanisms ([2]).
Psycholinguistics research on the cognitive mechanisms involved in language recognition or production frequently uses RT as a measure of behavior. Like all scientific disciplines, psycholinguistics relies on hypothesis testing to support its theoretical propositions. Since the early 2000s, researchers have taken up linear mixed-effects models (LMEMs). As described by [3], a LMEM allows for the proper consideration of one of the characteristics of psycholinguistic experiments: the presence of two random-effect variables. Indeed, the experimental structure of the experiments conducted in this field involves having a group of participants process a set of stimuli (see hereafter). Thus, the statistical analysis must allow the inclusion of these two random-effect variables, i.e., participants and items, in the structure of the model. The introduction of LMEM was thus an important methodological advance for psycholinguistics.
There is, however, one point that has not received much attention from psycholinguists. Experimentation with human beings is often subject to many vagaries. Imagine researchers whose goal is to understand how an adult retrieves the spelling of a word from their memory ([4]). They ask a group of thirty participants to write down the names of 150 images presented on a computer screen. A device allows them to measure the time between the appearance of the image on the screen and the first writing gesture of the participant. These researchers can potentially collect 4500 RT values. However, between trials lost for technical reasons, spelling errors, and certain habits during data analysis (e.g., censoring of data greater than two standard deviations from the mean, right-censoring data), a non-negligible number of data are removed. For example, ref. [4] reported removing just over 20% of the data.
The issue of missing and censored data has received relatively little attention in psycholinguistics (see, however, [5,6]). This is especially true since the introduction of LMEMs. Psycholinguists assume that these models can be run on a sample of data with “holes”. The analysis strategy is then of the “keep the case empty” type, ignoring the bias that this introduces in the estimation of the parameters of the model and thus of the decisions taken. The objective of this work is thus to develop algorithms to manage the presence of missing and censored data in these psycholinguistic experiments. Two points are to be taken into account. On one hand, experimental designs in scientific psychology may involve assumptions about interactions between two or more fixed-effect variables. On the other hand, researchers suggest, for theoretical reasons, that all interactions between fixed-effect variables and random-effect variables made possible by the model should be included ([7]). The potential presence of these two types of interactions (fixed–fixed variables; random–fixed variables) are constraints on the development of the missing data procedure.
Let us first recall that the LMEM ([8]) is an extension of the simple linear model that allows both fixed and random effects to be represented. In 1861, the LMEM was introduced under the name of a one-way random-effects model ([9]), that is, a model with one random variable and without any fixed variable. From 1990 and onward, ref. [9] underlined that LMEMs became popular in many research applications including economics, sociology, insurance, agronomy, epidemiology, genetics, etc. They are used in longitudinal data analysis, multilevel modeling and small estimations. The analysis of this type of models is presented in [10]; for more literature about the methodology, theoretical results and software, see the books [11,12,13,14]. LMEMs are fitted and analyzed in R by using the package lme4 or lmertest ([15]).
In [16], the author presented the fixed interactions between two factors in an LMEM, using the maximum product interaction F-test. Ref. [17] was interested in the sample size of an LMEM. He proposed a formula to estimate the sample size based on testing a mixed model that contained fixed interactions. In 2019, ref. [18] proposed an estimation method to recover the principal effects and interactions, because the existing method did not allow the integration of these effects in a mixed data frame. They approved that the proposed method gave optimal results to their applications’ conditions. On the same side, ref. [19] presented the estimation of the fixed interactions. These interactions are normally introduced by the product of the variables, but the algebraic transformations revealed that this technique did not produce a within-unit estimator. A Monte Carlo method confirmed that the FE (fixed-effects) estimator of ( x , z ) was biased, if one of these variables was correlated with another one. In order to present the interaction between x and z, it is possible to use the current syntax x z . This consideration is called “double-demeaned”; it is less efficient than the standard FE and only works with T < 2 . For the application in R, the Markov chain Monte Carlo method is applied using the package mcmc. Ref. [20] introduced the MCMCpack package that contains functions to perform Bayesian inference using a posterior simulation for a number of statistical models (https://CRAN.R-project.org/package=MCMCpack (accessed on 14 Jun 2011)), while [21] presented the package MCMCglmm (https://cran.r-project.org/web/packages/MCMCglmm/index.html (accessed on 2 February 2010)) for the MCMC method to fit the generalized linear mixed-models for multiple response vectors. This method was also used by [22] to identify unknown parameters in the biological field, such as detecting the concentration of target molecules, because of the importance of this method for extracting the information. The MCMC method was applied in the work of [23] to solve mechanical problems by using a Bayesian method.
Ref. [24] presented five types of LMEMs by introducing two random variables (see Appendix A.1), where the models were compared to show the importance of including the maximum effects that could affect the model.
One of the solutions that handles the missing data can be the expectation–maximization (EM) algorithm. It is a very useful algorithm for the estimation of the maximum likelihood function. This method can be a solution when the only data available do not allow the estimation of the parameters ([25]), and/or the expression of the likelihood is analytically impossible to maximize.
In other words, it aims to provide an estimator when the problem comes from the presence of missing data. When the data are incomplete, the knowledge of these values would make it possible to estimate the parameters. The EM algorithm takes its name from the fact that at each iteration, it operates two distinct steps:
Step E: 
This step is called the expectation (E) step, where we are interested in finding the expected value of the unobserved or unknown variables given the observed data and the value of the parameters.
Step M: 
This step is called the maximization (M) step; in this step, we maximize the expected log-likelihood by using the estimation of the unknown data carried out in the previous step. These parameter estimates are then used to determine the distribution of the unknown variables in the next iteration.
At some points, the expectation or the maximization steps are impossible to apply directly ([26]); from there, the use of an extension form of EM is useful, such as MCEM, where the (E) step is replaced by a Monte Carlo simulation, or SAEM, where the (E) step is replaced by a stochastic approximation. SAEM was a solution of nonlinear mixed-effects models (NLMEM). Ref. [27] proposed a new methodology for maximum likelihood estimation in mixtures of nonlinear mixed-effects models. The resulting MSAEM (mixture SAEM) algorithm is now implemented in the Monolix software tool.
The aim of this paper was to perform the SEM algorithm, under an LMEM by including two types of incomplete data (the censored and the MAR types) and by taking into consideration for the first time the interactions, where our proposed model contains the interactions between fixed variables and fixed–random variables. This document is organized as follows: In Section 2, we present three algorithms based on the expectation–maximization (EM) method, the first one is called the SAEM algorithm, the second is our proposed SEM algorithm and the third one is based on the MCMC method. We also present the incomplete data types, divided into missing data and censored one. In Section 3, we define the proposed model with some specific cases. In Section 4, we introduce a method to achieve the convergence. In Section 5, we compare the results obtained from simulated psychological and real data. In Section 6, we conclude the proposed study with some perspectives and future directions.

2. Methodology

2.1. EM Algorithm

As previously said, the EM algorithm is a widely used algorithm in the case of incomplete data; in this situation, the maximum likelihood function is difficult or impossible to use to estimate the parameter vector θ of the considered model. We formalize directly an iteration from which we can understand clearly how this algorithm works:
Let y = ( y 1 , , y n ) be the independent and identically distributed (i.i.d.) observations of likelihood p ( y | θ ) .
The maximization of log p ( y | θ ) is impossible.
We consider hidden data z = ( z 1 , , z n )  which make the maximization of the likelihood of the complete data possible when known.
As we do not know these data z, we estimate the likelihood of the complete data by taking into account all the known information so the estimator is given as follows (E step):
Q ( θ | θ k 1 ) = E [ log p ( y , z ; θ | y ; θ k 1 ) ] ,
where θ k 1 is the vector of the parameters at iteration k 1 .
Finally, we maximize this estimated likelihood to determine the new value of the parameter (M step). Thus, the transition from iteration k 1 to iteration k in the algorithm consists in determining the parameters vector at iteration k, θ k :
θ k = arg max θ Q ( θ | θ k 1 ) .
where θ 0 is chosen arbitrarily. When one of these two steps are impossible to complete, we can consider an extension form of the EM algorithm such as the SAEM, SEM or MCEM algorithms. In the next subsection, we derive the SAEM algorithm into the formula of the EM algorithm.

2.1.1. SAEM Algorithm

The stochastic approximation expectation maximization (SAEM) algorithm was proposed by [26], in which the E step was replaced by a stochastic approximation. The stochastic approximation algorithm was first introduced by [28] and also used by [29], where the algorithm generates iterates of the form:
θ k = θ k 1 γ k [ h ( θ k 1 ) + ϵ k ] ,
where γ k is a sequence of positive step sizes, h is a function of θ , and ϵ k is a constant such that ϵ k = E [ h ( θ k ) ] . From this form, the SAEM algorithm is obtained where the E step of the EM algorithm is divided into two steps; consider the iteration k:
  • First, we sample a realization z k of the latent variable from the conditional distribution ( p θ k 1 ( z | y ) ) of z given y, using the value of the parameter θ k 1 at iteration k 1 .
  • Second, by using the realization z k from the first step, we update the value of Q k ( θ | θ k 1 ) (see, (1)) through a stochastic approximation procedure.
Then, the algorithm continues as follows:
  • Initialization step: Initialize θ 0 in a fixed compact set.
Then, for all k 1 , the kth iteration consists in three steps:
  • Simulation step: simulate z k from the conditional distribution p θ k 1 ( . | y ) .
  • Stochastic approximation step: compute the quantity
    Q k ( θ ) = Q k 1 ( θ ) + γ k 1 m ( k ) j = 1 m ( k ) log f ( z k ( j ) ; θ ) Q k 1 ( θ ) ,
    where m k is the number of simulations at each iteration.
  • Maximization step: update the parameter value according to θ k = arg max θ Q k ( θ ) .
The SAEM algorithm is more efficient compared to the MCEM, where, at each iteration, the simulation of the missing values is repeated and the ones obtained previously are not used. In the SAEM, all the simulated missing values contribute to the expectation step and that is the advantage of this algorithm where the maximization step is cheaper than the simulation one. Next, we consider the SEM algorithm to show how it differs from an SAEM.

2.1.2. SEM Algorithm

The SEM approach is a specific case of the SAEM algorithm. In the SEM algorithm, the step size is set to zero γ k = 0 and the number of simulations by iteration is constant (usually, m k = 1 ).
This method is an extension of the EM algorithm and is a solution when the maximum likelihood is impossible to complete. This algorithm was introduced by [30], where between the steps E and M, they used a stochastic step that consisted in generating a complete sample containing latent variables from the conditional density, based on the observed data. Starting with an initial position θ 0 , an iteration k of the SEM algorithm, where, to each θ k , a θ k + 1 is associated, is defined as follows:
Step E: 
Compute the conditional density f ( y | x ; θ k ) ;
Step S: 
Draw z k from the conditional distribution, then obtain the complete sample y k = ( x , z k ) ;
Step M: 
Update the parameters θ k + 1 by maximizing the likelihood function based on the complete vector y k .
In the next subsection, we consider the MCEM algorithm to show how a Monte Carlo simulation is used in the EM algorithm.

2.1.3. MCEM Algorithm

In 1990, ref. [31] proposed to replace the expectation step to compute Q ( θ | θ k 1 ) by a Monte Carlo integration, hence the name MCEM algorithm (Monte Carlo EM). At iteration k, the E step is replaced by the following procedure:
Simulation step (S-step): generate m ( k ) realizations z k ( j ) (with j = 1 , , m ( k ) ) of the missing data vector under the distribution function p ( z ; θ k )
Monte Carlo integration: compute the approximation of Q ( θ | θ k 1 ) according to:
Q k ( θ ) = 1 m ( k ) j = 1 m ( k ) log f ( z k ( j ) ; θ ) .
The maximization step remains unchanged.
As said before, these algorithms can be used in such a way as to handle the unobserved values of the data that are defined in this next subsection.

2.2. Incomplete Data

We can talk about incomplete data when the values in our response vector are not all observed for many reasons; we can also say it is a question without an answer.
These types of data are a common problem recognized by statisticians. Ref. [32] presented many solutions to handle this problem based on simple and multiple imputations. Ref. [33] proposed to handle the incomplete data in a hierarchical model by using the SEM algorithm.
In this work, we are interested in data that contain two types of incomplete values: censored data denoted by Y c and missed data Y m i s s . In the next subsection, the censored data are considered.

2.2.1. Censoring

Censored data are any data for which we do not know the exact event time.
There are two types of censored data: right-censored and left-censored.
Right-censored data: these are data where the event has not yet been achieved when the study is finished.
Left-censored data: these are data when the event is achieved before the study starts.
We consider in this paper both censoring types of (left- and right-censoring) data with four percentages ( 0 % , 5 % , 10 % and 20 % ). Let t be the censoring level and T the set of censored indices that can be written as T = { ( i 1 , , i p + k ) I , Y i 1 , , i p + k , j < t or Y i 1 , , i p + k , j > t } , and let Y = ( Y o , Y c ) , with  Y o = { Y i 1 , , i p + k , j , ( i 1 , , i p + k ) T } the vector of observed response variables and Y c = { Y i 1 , , i p + k , j , ( i 1 , , i p + k ) T } defining the vector of censored response variables. The second type is the missing data, defined in the next subsection.

2.2.2. Missing Data

In general, missing values are produced when a value in the data is not represented for a given variable, for many reasons that can be linked to the objective of the study (for example, participants do not answer the questions). This appears in many research studies, particularly when collecting data and where participants are studied over a period of time.
At first, studies were developed assuming no missing values. In the late 1980s, with the advancement in technology, that problem attracted the attention of many researchers who wished to study several techniques for handling it. Depending on the reasons for their absence, these values could be divided into three types: missing completely at random (MCAR), missing at random (MAR) and missing not at random (MNAR).
In Appendix A.2, we define the types of missing value and the methods to generate these types in a simulation study.
In this paper, we are interested in imputing the MAR type of missing data with four percentages ( 0 % , 5 % , 10 % and 20 % ), which are crossed with four other percentages of censoring, resulting in 16 cases of incomplete data. For the missing data, we denote by Y o the vector of observed response variable and Y m i s s the missed one, so Y = ( Y o , Y m i s s ) . In the next section, we present the main results of this work.

3. Main Results

3.1. The Proposed Model

Motivated by psychological data, in which we typically have two random variables, participants designed by S and items designed by I, we consider a model that contains p covariates { x 1 ; x 2 ; ; x p } and two random variables. The predicted variable, also named the variable of interest, is denoted by Y. Then, we consider the following linear mixed model:
Y = β 0 , s , i + β 1 , s , i x 1 + β 2 , s , i x 2 + + β p , s , i x p + r j , i , s ,
with r the residual of the model that follows a normal distribution. We give:
β 0 , s , i = δ 0 + S 0 , s + I 0 , i β 1 , s , i = δ 1 + S 1 , s + I 1 , i β 2 , s , i = δ 2 + S 2 , s + I 2 , i β p , s , i = δ p + S p , s + I p , i .
By a replacement of these considerations in Equation (2), our model can be rewritten as follows:
Y = δ 0 + δ 1 x 1 + δ 2 x 2 + + δ p x p + S 0 , s + S 1 , s x 1 + S 2 , s x 2 + + S p , s x p + I 0 , i + I 1 , i x 1 + I 2 , i x 2 + + I p , i x p + r .
By taking
X = 1 x 1 x 2 x p + k = 1 x j , δ = δ 0 δ j T
and denoting by T the transpose matrix,
Z = z j = x j , u = u j = S j , l , s + I j , l , i T ;
and e j l = e j .
This leads us to rewrite our model as follows:
Y = X δ + k = 1 K z k u k + e = X δ + Z u + e ,
where E ( Y ) = X δ and Var ( Y ) = V = Z G Z T + R , with u N ( 0 , G ) , e N ( 0 , R ) and Cov ( u , e T ) = 0 .
Now, by including all the interactions, the general model can be written as:
Y = F + I FF + R + I FR + e ,
where we take:
F = δ 0 + δ 1 x 1 + + δ p x p fixed variables , I FF = δ p + 1 x p + 1 + + δ p + k x p + k interaction between fixed fixed variables , R = ( S 1 , s x 1 + I 1 , i x 1 ) + + ( S p , s x p + I p , i x p ) random variables , I FR = ( S p + 1 , s x p + 1 + I p + 1 , i x p + 1 ) + + ( S p + k , s x p + k + I p + k , i x p + k ) interaction between fixed random variables , and e = ( S 0 , s + I 0 , i ) + r residuals .
We set k, the number of interactions, as  k = 2 p p 1 and let j , l [ 1 , p + k ] .
Our random participants variable is given as follows:
S j , l , s N ( 0 , Σ S ) ,
and the variance–covariance matrix of the participants is written as
Σ S = σ j , S 2 ; if j = l ρ S σ j , S σ l , S ; otherwise .
We consider that the items random variable also follows a normal distribution:
I j , l , I N ( 0 , Σ I ) ;
then, the variance–covariance matrix of the items is given by:
Σ I = σ j , I 2 ; if j = l ρ I σ j , I σ l , I ; otherwise .
For j , l [ 1 , p + k ] , the variance–covariance matrix G is equal to:
G = Σ u = σ j , S 2 + σ j , I 2 ; if j = l ρ S σ j , S σ l , S + ρ I σ j , I σ l , I ; otherwise .
The goal is to predict our model by determining the parameters δ and u; we propose to use Henderson’s linear system ([34]) based on the maximum likelihood function. This approach is presented in Appendix A.3. The matrix R has the form:
R = I ; if j = l n p k σ e j 2 ; if j = l > n p k 0 ; otherwise ,
where n represents the total number of observations. Therefore, with  Z = ( z 1 , , z p + k ) , the variance of Y is equal to:
V = ( j = 1 p + k z j Σ u 1 , u j ) z 1 + ( j = 1 p + k z j Σ u 2 , u j ) z 2 + + ( j = 1 p + k z j Σ u p , u j ) z p + + ( j = 1 p + k z j ) Σ u p + k , u j z p + k + R .
The linear system equations of Henderson applied in this study is also presented in Appendix A.4. In the next subsection, we present some specific cases.

3.2. Specific Cases

In this section, we show how the model can be simplified if we take the following two cases: first, only the fixed–fixed interaction part is presented and second, there is no interaction.

3.2.1. Case 1: Fixed–Fixed Interaction

In this case, we consider the interactions between the fixed variables. Then, in the random part, the u vector is equal to 0:
u = u 1 u 2 u p u p + 1 u p + k T = 0 0 0 0 0 T ;
therefore, the  V matrix is equal to:
V = R = I ; if j = l n p k σ e j 2 ; if j = l > n p k 0 ; otherwise .
For more mathematical development, see Appendix A.4.

3.2.2. Case 2: No Interactions

In this case, there are no interactions between the variables; our model is simplified, where the random part with the fixed–fixed interaction is ignored:
X = 1 x j ; δ = δ 0 δ j T ; Z = z j ; u = u j = 0 0 0 T ; e j l = e j .
The V matrix is reduced to:
V = R = I ; if j = l n p σ e j 2 ; if j = l > n p 0 ; otherwise .
See the simplified system in Appendix A.4. Next, we present how the SEM algorithm can be applied in this work.

3.3. SEM Algorithm

In this paper, we propose to handle missing data and censored problems in the presence of interactions by using the SEM algorithm, proposed by [33]. The stochastic expectation maximization (SEM) algorithm is a method used to estimate the parameters when it is complicated to use the EM algorithm; it is a particular case of multiple imputations (MI).
To understand MI ([35]), we need to know the idea behind a simple imputation (SI) in which the nonobserved data { Y n o } (missed or censored) are replaced by a value and then the parameters are estimated using known methods to maximize the likelihood function. To obtain a robust estimate, the simple imputation can be repeated several times with different values of { Y n o } and the results can be combined. This method is called multiple imputations.
We applied the algorithm after crossing the MAR values with the censored ones (see Algorithm 1). In the SEM algorithm ([30]), the values of { Y n o } are drawn from the conditional distribution of the nonobserved data given the observed ones using the current values of the parameters. We generated samples from P Θ ( Y n o | Y o ) where this distribution was calculated using Gibbs’s sampling; the procedure is underlined in Algorithm 1 (the SEM method using Gibbs’s sampling was developed in [33]). For more information about the expectation maximization algorithm, see [36]. In the next subsection, we present the extension form of the SEM that leads us to the SAEM algorithm.
Algorithm 1 SEM algorithm: N is the number of iterations of the SEM algorithm, M is the burn-in level, Y 1 , , Y n is the response vector, G is the number of iterations of Gibbs sampling, Y [ u i ] is the response vector with respect to the ith random variable, and fixed is the summation of the fixed effects with the fixed interaction part.
Input: N, M, Y 1 , , Y n , and G.
  1:  Random initialization of Θ
  2:  for  j = 1 , , N  do
  3:      for  g = 1 , , G  do
  4:          draw e ( g ) from N ( 0 , σ e 2 , lower = min ( e ) fixed u ( g 1 ) , upper = max ( e ) fixed u ( g 1 ) )
  5:          draw u 1 ( g ) from N ( 0 , Σ u 1 , lower = max ( min ( Y [ u 1 ] ) u 2 ( g 1 ) e ( g ) ) , upper = min ( max ( Y [ u 1 ] ) u 2 ( g 1 ) e ( g ) ) )
  6:          draw u 2 ( g ) from N ( 0 , Σ u 2 , lower = max ( min ( Y [ u 2 ] ) u 1 ( g ) e ( g ) ) , upper = min ( max ( Y [ u 2 ] ) u 1 ( g ) e ( g ) ) )
  7:      end for
output:      Y n o obtained from sampled ( u ( G ) , e ( G ) )
  8:       Θ j + 1 = argmax Θ L Θ | Y o , Y n o
  9:  end for
10:  Θ ^ = ( N M ) 1 j = M + 1 N Θ j
output:   Θ ^

3.4. SAEM Algorithm

In the SEM algorithm, Θ j + 1 (the parameter at state j + 1 ) depends only on Θ j and Y j . By taking T the operator of the EM algorithm and M, the operator which associates Θ j + 1 with Y j via step M, the updated parameter Θ j + 1 can be written as follows (see [30]):
Θ j + 1 = T ( Θ j ) + V ( Θ j , Y j ) ,
where V ( Θ j , Y j ) = M ( Y j ) T ( Θ j ) .
The SAEM algorithm (see also Algorithm 2) is an extension form of the SEM algorithm. Starting with an initial position Θ 0 , the iteration of the SAEM algorithm, to which for each Θ j , a Θ j + 1 is associated, is defined by the equation:
Θ j + 1 = T ( Θ j ) + γ j V ( Θ j , Y j ) ,
where γ j is the step size, which is a decreasing sequence of positive real numbers starting with γ 0 = 1 , γ j 0 when j . The theoretical and practical performance values of the SAEM algorithm are strongly dependent on the speed of convergence towards 0 and the regular decreasing of the sequence γ j , which explains the importance of choosing the step size.
Ref. [37] performed many tests to see which sequence was more efficient, and they obtained two types of sequences that were significant, one of a slow mode:
γ j = cos ( j × π 2 N ) ,
the other type was of a linear mode:
γ j = ( j + 1 ) N .
Ref. [38] chose the decreasing sequence from their previous experiments [39]; they took that sequence as
γ j = cos ( N × α ) 𝟙 { 0 j 20 } + c j 𝟙 { j > 20 } , where cos ( 20 × α ) = c 20 = 0.3 .
That sequence was also chosen by [40,41] in a nonlinear mixed model as 1 at the first iteration and ( j N ) 1 during the last iterations.
Algorithm 2 SAEM algorithm: N is the number of iterations of the SEM algorithm, M is the burn-in level, Y 1 , , Y n is the response vector, G is the number of iterations of Gibbs’s sampling, γ j is the decreasing sequence, Y [ u i ] is the response vector with respect to the ith random variable, and fixed is the summation of the fixed effects with the fixed interaction part.
Input: N, M, Y 1 , , Y n , and G.
  1:  Random initialization of Θ
  2:  for  j = 1 , , N  do
  3:      for  g = 1 , , G  do
  4:          draw e ( g ) from N ( 0 , σ e 2 , lower = min ( e ) fixed u ( g 1 ) , upper = max ( e ) fixed u ( g 1 ) )
  5:          draw u 1 ( g ) from N ( 0 , Σ u 1 , lower = max ( min ( Y [ u 1 ] ) u 2 ( g 1 ) e ( g ) ) , upper = min ( max ( Y [ u 1 ] ) u 2 ( g 1 ) e ( g ) ) )
  6:          draw u 2 ( g ) from N ( 0 , Σ u 2 , lower = max ( min ( Y [ u 2 ] ) u 1 ( g ) e ( g ) ) , upper = min ( max ( Y [ u 2 ] ) u 1 ( g ) e ( g ) ) )
  7:      end for
output:       Y n o obtained from sampled ( u ( G ) , e ( G ) )
  8:       Θ j + 1 = argmax Θ γ j × L Θ | Y o , Y n o
  9:  end for
10:   Θ ^ = ( N M ) 1 j = M + 1 N Θ j
output:   Θ ^
When applying these two algorithms, we were confronted with a problem of convergence caused by the presence of the interaction between the fixed and random variables. Therefore, in the next section, we present how to handle this problem by using the Hamiltonian Monte Carlo algorithm.

4. Convergence of Parameters

While introducing the fixed–random effect in the application section, we faced a convergence problem in the SEM and SAEM algorithms when using the standard form of Monte Carlo (MC). In order to solve this problem, we propose to use the hybrid MC algorithm ([42]), which uses the symmetric Metropolis–Hastings algorithm to accept or reject a proposal based on the Hamiltonian function, hence the name of the algorithm, the Hamiltonian Monte Carlo algorithm (HMC). In the next subsection, we present how to implement the HMC algorithm.

Implementation

The Hamiltonian function H ( p , q ) is written in term of the joint density π ( p , q ) :
H ( p , q ) = log π ( p , q ) ,
which decomposes into two terms:
H ( p , q ) = log π ( p | q ) log π ( q ) = K ( p , q ) + V ( p , q ) .
The first term corresponds to the density of the target distribution, the second one is determined by the target distribution when K ( p , q ) is unconstrained and must be specified by the implementation.
The acceptance probability of moving from state ( p 0 , q 0 ) to ( p i , q i ) is determined using the Metropolis–Hastings algorithm ([43,44]), with  Q ( p i , q i | p 0 , q 0 )  being the probability density defining each proposal:
a ( p i , q i | p 0 , q 0 ) = min 1 , Q ( p 0 , q 0 | p i , q i ) π ( p i , q i ) Q ( p i , q i | p 0 , q 0 ) π ( p 0 , q 0 ) .
Referring to the symmetric Metropolis ([45]), the acceptance probability is simplified to the form:
a ( p i , q i | p 0 , q 0 ) = min 1 , π ( p i , q i ) π ( p 0 , q 0 ) = min 1 , exp ( H ( p i , q i ) ) exp ( H ( ( p 0 , q 0 ) ) = min ( 1 , exp ( H ( p i , q i ) + H ( p 0 , q 0 ) ) ) .
In this study, we considered that we had evidence of the convergence towards stationarity due to the Markov Chain (see, [45,46]).
In the next section, we present some numerical applied results based on a psychological simulation and then real data.

5. Numerical Experiment

This section aims to compare the SEM algorithm proposed in this paper with other methods in the presence of incomplete data in a complete LMEM. The method was first applied to simulated data, then to real ones.

5.1. Simulated Data

The simulation data were created from an experimentally obtained database. In order to explore the cognitive process of retrieval in memory of the spelling of French words, 30 participants had to handwrite the label of 150 drawings of objects, constituting a database of 4500 RTs ([4]). After removing the RTs corresponding to technical errors, spelling errors, and censoring data greater than 2.5 standard deviations from the mean, 3434 values remained. Ref. [4] performed a linear regression analysis involving nine fixed-effects variables. In the present work, we retained two of them. The first was an estimate of the age at which the image name was acquired in childhood (age of acquisition, AoA hereafter). This factor is one of the most important predictors of picture-naming RTs ([47]. The second fixed-effects variable corresponded to the number of letters in a picture label (Lett_cat). We performed a median split to obtain a categorical variable. Thus, image labels with a number of letters lower than the sample average were considered as short words and those with a higher number of letters were considered as long words. The interaction between these two fixed-effects variables was included in the model. Finally, there were two random effect variables: items, i.e., the 150 image labels produced by the 30 participants, and the participants themselves. The fixed–random interaction between AoA and the participants was also included in the LMEM.
Therefore, the model could be written as follows:
RTs = b 0 + b 1 × AoA + b 2 × Lett _ cat + b 3 × AoA × Lett _ cat +   participants + items + AoA × participants + e .
In this section, we illustrate the proposed algorithm by comparing the results to the SAEM and MCMC algorithms using different proportions of incomplete data. The comparison was performed by computing the mean absolute error (MAE), which measured the errors between the reference and the predicted vectors (with Θ as the reference value and Θ ^ as the predicted one); the standard equation of the MAE is given by:
MAE = 1 n i = 1 n | Θ ^ i Θ i | .
We also computed the linear correlation between the parameters ( LCor   =   Cov ( Θ ^ i , Θ i ) σ ( Θ ^ i ) σ ( Θ i ) ) and the Spearman rank correlation (RCor =  1 6 d 2 n ( n 2 1 ) , where d 2 is the square of the difference in the ranks of the two coordinates).
In this numerical section, we provide the 16 cases to see the performance of the methods in the absence of one of these two considered types of unobserved values (MAR type/censored data) or in the presence of a low or high percentage. In addition, we used the simulated data of 4500 observations after testing our proposed methods on a larger number of observations. We obtained the same results as those for 4500 values.
After the implementation of our algorithm, we checked the convergence by illustrating the values of the estimated parameters obtained at each iteration; this is presented in Figure 1.
In Figure 2, we plotted the normalized residuals of each case. We observed a similar dispersion between the standard simulated data and the SEM algorithm, while we noticed a small but visible scattering by implementing the SAEM algorithm, and finally, we observed that when we used the MCMC method, the residuals were dispersed far from the normal line.
Our results are visualized quantitatively in Table 1. Based on the MAE value, we can observe that the SEM gave the lowest values in all cases except for the case of 5 % MAR and 20% censoring, where the SAEM ( 8.108 ) beat the SEM ( 8.118 ) by a very narrow margin. By observing the full vectors, the SEM gave the best parameter values in two cases: (0%, 10%) and (20%, 5%). In the other cases, we observed that with a high percentage of MAR (20%), b 2 and b 3 in the SAEM method gave the closest values while for the other parameters, it was the SEM algorithm. In all cases, we see that the MCMC was not a good choice either for the parameters values or for the other measurements. Moreover, we can see that for LCor (respectively, Rcor), the SEM and the SAEM algorithms gave approximately the same results 0.999 (respectively, 1) in all cases. This showed that these two methods had a small difference except for the case of (20%, 5%) where the RCor of the SAEM algorithm decreased from 1 to 0.9285 . The MCMC had an RCor ranging between a maximum value of 1 and a minimum value of 0.738 ; that maximum value was obtained with 5% of MAR and the absence of censored data.

5.2. Real Data

In this section, we applied the three algorithms to the database used to create the simulated data. Of the 4500 RTs recorded during the experiment (30 participants producing the names of 150 items), 115 were removed by right censoring ( 2.56 % of MNAR). Ref. [4] also removed 951 values, creating 21.13 % of missing MCAR/MAR data. By applying the three proposed algorithms with some missing data to solve the problem in the presence of fixed and fixed–random interactions, we obtained the results presented in Table 2. The results were compared with respect to the initial vector where the missed values were treated by keeping the cases empty (KE). Interestingly, this is how missing and censored data are handled in psycholinguistic studies.

6. Conclusions

In this study, we proposed an algorithm based on the stochastic expectation maximization (SEM) to estimate the parameters under a linear mixed-effects model (LMEM) that contained fixed–fixed and fixed–random interactions in the presence of different percentages of incomplete data.
The simulated and real data showed that the proposed SEM algorithm gave the best results compared to other competitors, where the bias of the parameters was smaller than the bias in the SAEM and MCMC algorithms.
The simulation results were obtained by using statistical software R (see Appendix A.5 for some source code).
We plan to extend this approach by considering a more complicated level of interaction such as random–random effects and the interactions between more than two variables. Another direction could be to consider an extension of the present work by considering a generalized model under the logistic regression by taking one for the NA values and zero for the observed ones.

Author Contributions

Conceptualization, C.P. and Y.S.; methodology, Y.S.; software, A.Z.; validation, A.Z., C.P. and Y.S.; formal analysis, A.Z.; investigation, C.P. and Y.S resources, C.P.; data curation, C.P.; writing—original draft preparation, A.Z.; writing—review and editing, Y.S.; visualization, C.P. and Y.S.; supervision, C.P. and Y.S; project administration, C.P and Y.S.; funding acquisition, Y.S. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Informed Consent Statement

Informed consent was obtained from all subjects involved in the study.

Data Availability Statement

Data is unavailable due to privacy.

Acknowledgments

We would like to thank the Editor-in-Chief of Entropy, the associate Editor and two anonymous reviewers for their helpful comments, which helped us to focus on improving the original version of the paper.

Conflicts of Interest

The authors declare no conflict of interest.

Appendix A

In the following section, we present how [24] defined the models by including the maximum effects. It is interesting to show the types of missing data with the implementation methods. Then, the approach of Henderson is introduced and our model matrices to estimate the parameters are defined. Finally, we present some source code of the SEM algorithm.

Appendix A.1. Models Extension

Ref. [24] compared five types of models in order to prove the importance of taking the maximum effects. They presented the two random variables subject, designed by S, and items, designed by I. By taking i as the items’ index, s as the subject’s index and c as the observation, the first model was the maximum model that contained the fixed and random intercepts with the interactions, including also ρ S and ρ I :
Y c , i , s = β 0 + S 0 , s + I 0 , i + ( β 1 + S 1 , s + I 1 , i ) X c + ϵ c , i , s ,
S 0 , s S 1 , s N 0 0 , τ 00 2 ρ S τ 00 τ 11 ρ S τ 00 τ 11 τ 11 2 ,
I 0 , i I 1 , i N 0 0 , w 00 2 ρ I w 00 w 11 ρ I w 00 w 11 w 11 2 ,
ϵ c , i , s ( 0 , σ 2 ) .
The second model differed by its correlation parameters, where S 0 , s and S 1 , s (respectively, I 0 , i and I 1 , i ) were independent. ρ S (respectively, ρ I ) was set to zero:
Y c , i , s = β 0 + S 0 , s + I 0 , i + ( β 1 + S 1 , s + I 1 , i ) X c + ϵ c , i , s ,
S 0 , s S 1 , s N 0 0 , τ 00 2 0 0 τ 11 2 ,
I 0 , i I 1 , i N 0 0 , w 00 2 0 0 w 11 2 ,
ϵ c , i , s ( 0 , σ 2 ) .
In the third model, they reduced Equation (A1) by ignoring the items’ random slopes (the interaction between the items’ fixed and random effects), where w 11 was set to 0:
Y c , i , s = β 0 + S 0 , s + I 0 , i + ( β 1 + S 1 , s ) X c + ϵ c , i , s ,
S 0 , s S 1 , s N 0 0 , τ 00 2 0 0 τ 11 2 ,
I 0 , i ( 0 , w 00 2 ) ,
ϵ c , i , s ( 0 , σ 2 ) .
The fourth model kept the random items’ slope and excluded that of the subject, where τ 11 = 0 :
Y c , i , s = β 0 + S 0 , s + I 0 , i + ( β 1 + I 1 , i ) X c + ϵ c , i , s ,
S 0 , s ( 0 , τ 00 2 ) ,
I 0 , i I 1 , i N 0 0 , w 00 2 0 0 w 11 2 ,
ϵ c , i , s ( 0 , σ 2 ) .
The fifth model excluded the two random slopes (for items and subject), where τ 11 = w 00 = 0 :
Y c , i , s = β 0 + S 0 , s + I 0 , i + β 1 X c + ϵ c , i , s ,
S 0 , s ( 0 , τ 00 2 ) ,
I 0 , i ( 0 , w 00 2 ) ,
ϵ c , i , s ( 0 , σ 2 ) .
In the next section, we present the different types of missing data and how to generate them using different methods.

Appendix A.2. Missing Data Types and Imputations

Let M be the missing data indicator that defines what is known and what is missing. The response variable is denoted by Y in the complete data, where Y = { Y o , Y n o } with Y o is the observed part and Y n o is the nonobserved one. Ref. [48] denoted the missing part by Y m i s s . Consider a dataset of p variables and n subjects, with  Y = { Y 1 , , Y p } . Y should look like a matrix of n rows and p columns. In the literature on missing data, see also [49,50].
We show in this section the three types of missing values and how to generate them in a simulation study based on a rule in each case ([51]).
We start with the first type named MCAR.
The missing completely at random type (MCAR) is a special case of the MAR type that is presented later. In this type, the value is missed due to chance and the absence is not related to the subject, so the distribution of M does not depend on Y o or on Y m i s s and is identical for all the observations:
P ( M | Y ) = P ( M ) ,
The statistical advantage of MCAR data is that the analysis remains unbiased, none of the variables is affected more than another.
Suppose that each subject has a probability π of being missed from a variable, the missing data rule is given by P ( M = 1 ) = π . From this rule, we can determine various properties associated with the MCAR data such as the expected percentage of missing values and the expected number of missing data patterns in the MCAR data.
If we are working with univariate patterns, let n be the number of subjects in the data and K the random variable indicating the number of subjects with missing values in the data that follows a binomial distribution given by K B ( n , π ) , where 0 π 1 . Since E ( K ) = n π and V ( K ) = n π ( 1 π ) , the expected percentage of missing values is:
E ( Π ) = 1 n E ( K ) = π ,
where Π = K n is the random variable that defines the estimated percentage of missing values in a sample. the variance of this estimated percentage is given by:
V ( Π ) = 1 n 2 V ( K ) = π ( 1 π ) n .
In this case, we have two missing data patterns, pattern 1 includes subjects with complete data; pattern 2 includes subjects with missing data. Let I j be the indicator variable of the event, where pattern j is present in at least one subject in the sample with j { 1 , 2 } , The probability that pattern 1 is present in at least one subject is P ( I 1 = 1 ) = E ( I 1 ) = 1 π n . The probability that pattern 2 is present in at least one subject is P ( I 2 = 1 ) = E ( I 2 ) = 1 ( 1 π ) n . Let D be the number of distinct missing data patterns, D = j = 1 2 I j ; we can determine the expected value:
E ( D ) = j = 1 2 E ( I j ) = ( 1 π n ) + ( 1 ( 1 π ) n ) = 2 π n ( 1 π ) n .
Therefore, in MCAR data that contain more than two missing data patterns, the expected number of distinct pattern in a sample is given by:
E ( D ) = m j = 1 m ( 1 η j ) n ,
with η 1 , , η m are the corresponding probabilities for patterns 1 , , m , and depends only on the probability of missing values in each variable i (i.e., π 1 , , π l ). m = 2 l is the number of patterns and l is the total number of variables. Generally, the most used method to implement the MCAR type is randomly deleting the desired percentage of missing values.
The second type is MAR.
Missing at random is the most classic case; Rubin and D.B (1976) defined missing data to be MAR if the distribution of missing data did not depend on Y m i s s :
P ( M | Y ) = P ( M | Y o ) .
MAR data occur when the absence is not random but can be explained by the variables for which complete information exists. More generally, we are in the MAR type if the cause of missing values is not related to the goal of the study. If  Y 1 has missing values, then it is regressed on other variables Y 2 to Y l . The missing values in Y 1 are then replaced by the obtained predictive values. Similarly, if  Y 2 has missing values, then the Y 1 , Y 3 to Y p variables are used in the prediction model as independent variables. Thus, the missing values are replaced with the predicted ones. The missing data rules for MAR data can be organized into several categories: single-cutoff method, multiple-cutoff method, percentile method and logistic regression method.
Here, we define these three rules.
Single-cutoff method: Consider Y 1 as the variable with missing value and Y 2 is the missing data predictor with a cutoff point a. If a subject has Y 2 a , then its probability of being missing from Y 1 is π 1 , and if Y 2 < a , then its probability of being missing from Y 1 is π 2 . Let U be an indicator that takes the value 1 ( U = 1 ) when Y 2 a and U = 0 when Y < a . The missing data rule is given as:
P ( M = 1 | U = 1 ) = π 1 ; P ( M = 1 | U = 0 ) = π 2 ,
where the parameters associated with this rule are π 1 and π 2 , and π 0 is the probability that Y 2 is equal to or greater than the cutoff: P ( Y 2 a ) = P ( U = 1 ) = π 0 . To find the expected percentage of missing values, we first calculate the unconditional probability of a subject being missing from Y 1 :
π m i s s = P ( M = 1 ) = P ( M = 1 | U = 1 ) P ( U = 1 ) + P ( M = 1 | U = 0 ) P ( U = 0 ) = π 1 π 0 + π 2 ( 1 π 0 ) .
The expected percentage of missing values in a sample is given by:
E ( Π ) = π m i s s = π 1 π 0 + π 2 ( 1 π 0 ) ,
and the variance of this estimated percentage of missing values is written as:
V ( Π ) = π m i s s ( 1 π m i s s ) n .
Multiple-cutoff method: One advantage of this method is that it can be used to create a nonlinear relationship between the missing data indicator and the missing data predictor. To create a nonlinear relationship between the indicator and the predictor, we need to fix an upper cutoff and a lower cutoff. Suppose we fix the two cutoff points a and a , U = 1 when Y 2 a or Y 2 a , and  U = 0 when a < Y 2 < a . The missing data rule can be written as:
P ( M = 1 | U = 1 ) = π 1 ; P ( M = 1 | U = 0 ) = π 2 ,
and we can see that this missing data rule is the same as the missing data rule in the single-cutoff method. In the case of a linear relationship, we need to specify two or more cutoff points in the missing data predictor. Suppose we take the quartile points ( Q 1 , Q 2 and Q 3 ) as a cutoff, let V be a discrete uniform random variable created based on the values of Y 2 and that takes the following values:
V = 1 if Y 2 < Q 1 2 if Q 1 Y 2 < Q 2 3 if Q 2 Y 2 < Q 3 4 if Y 2 Q 3 .
Then, for the linear relation case, the missing data rule is given by:
P ( M = 1 | V = 1 ) = π 1 ; P ( M = 1 | V = 2 ) = π 2 ; P ( M = 1 | V = 3 ) = π 3 ; P ( M = 1 | V = 4 ) = π 4 .
Let n be the total number of subjects and n j be the number of subjects in each quartile group. The variance of the estimated π 0 is:
V ( Π 0 ) = π 0 ( 1 π 0 ) n ,
and the variance of the estimated π j where j { 1 , 2 , 3 , 4 } is defined as follows:
V ( Π j ) = π j ( 1 π j ) n j .
We can calculate each subject’s probability of being missing by calculating the marginal probability of M = 1 :
π m i s s = P ( M = 1 ) = π 1 π 0 + π 2 π 0 + π 3 π 0 + π 4 π 0 ,
and the expected percentage of missing values can be written as:
E ( Π ) = π m i s s = π 1 π 0 + π 2 π 0 + π 3 π 0 + π 4 π 0 ;
thus, the variance of this estimated percentage is given by:
V ( Π ) = π m i s s ( 1 π m i s s ) n .
To implement the missing data rule in (A4), researchers typically delete π 1 subjects with Y 2 < Q 1 , π 2 subjects with Q 1 Y 2 < Q 2 , and so on.
Percentile method: This case is an extension of the multiple-cutoff method, where each subject has a probability of being missing that depends on its percentile rank in the missing data predictor.
If there is a direct relation between the missing data indicator and the predictor, then the missing data rule is defined by:
P ( M = 1 | Y 2 = q k ) = k 100 ,
where q k is the Y 2 value corresponding to its kth percentile. If there is an indirect relationship, then the missing data rule is defined as:
P ( M = 1 | Y 2 = q k ) = 1 k 100 .
Logistic regression method: When generating MAR data by using the logistic regression method, the logistic regression model is considered as the missing data rule and the population regression coefficients of the model are the parameters associated with the missing data rule. Then, the logistic regression model for subject i is written as:
log P ( M i = 1 | y 2 , i ) 1 P ( M i = 1 | y 2 , i ) = β 0 + β 1 y 2 , i ,
where y 2 , i is the subject i’s value on Y 2 . The probability of being missing for each subject is given by:
P ( M i = 1 | y 2 , i ) = 1 1 + e β 0 β 1 y 2 , i .
Because the above function is continuous, it means the probability of being missing for Y 1 gradually increases or decreases as the value of Y 2 increases. With the logistic regression, there is no simple formula for calculating the expected percentage of missing data, so we estimate the expected percentage of missing data by calculating the mean of the probabilities in a sample with a large sample size:
π m i s s = 1 n i = 1 n 1 1 + e β 0 β 1 y 2 , i .
The third and last type is the MNAR data:
In missing not at random data, the cause of missing data is related to the variable of interest. In this type, the probability that an observation is missing depends on the observed Y o and the missing data Y m i s s . If  Y 1 is neither MCAR nor MAR, then it is MNAR. Generating MNAR data is the same as generating MAR data; the only difference is that in MNAR data the probability of missing values depends on the variable’s own value and not on the observed values of the other variables. Therefore, we can change the missing data predictor to the variable with missing values, then we use one of the methods presented above to generate MAR data. For example, to change rule (A3) that generates MAR data to one that generates MNAR data, we only need to replace the variable Y 2 by Y 1 ; therefore, the corresponding missing data rule for generating MNAR data is when Y 1 value is above a cutoff point a. Y 1 has a π 1 probability of being missing, otherwise, Y 1 has a π 2 probability of being missing.
Next, we define the linear system of Henderson to deal with the parameters’ estimation.

Appendix A.3. Henderson’s Approach

In order to estimate the parameters δ ^ and u ^ , we need to maximize the log function ( δ ^ , u ^ ) = arg max δ , u ln f ( Y , u ) . We define f ( Y , u ) = f ( Y | u ) f ( u ) with Y | u N ( X δ + Z u , R ) and u N ( 0 , G ) , so we can write:
2 ln f ( Y | u ) = N ln ( 2 π ) + l n R + ( Y X δ Z u ) R 1 ( Y X δ Z u ) , 2 ln f ( u ) = q ln ( 2 π ) + ln G + u G 1 u .
Denote l ( δ , u ; Y ) = ln f ( Y , u ) ; to maximize this function, we need to cancel the first derivatives of ln f ( Y , u ) with respect to δ and u:
[ 2 l ( δ , u ; y ) ] δ = 2 X R 1 ( y X δ Z u ) = 0 , [ 2 l ( δ , u ; y ) ] u = 2 Z R 1 ( y X δ Z u ) + 2 G 1 u = 0 .
From the two equations in (A5), Henderson presented the linear mixed model under the system:
X R 1 X X R 1 Z Z R 1 X Z R 1 Z + G 1 δ ^ u ^ = X R 1 y Z R 1 y
In the next part, we present how we applied Henderson’s approach to our proposed model.

Appendix A.4. Model Simplification

Complete model: We consider the following matrices, where r determines the rows, c determines the columns, and l , h [ 1 , p + k ] :
( X R 1 X ) M × N = ( n p k ) + σ e 1 2 + + σ e p 2 + σ e p + 1 2 + + σ e p + k 2 ; if r = c = 1 j = 1 n p k x l , j + o = 1 p + k x l , n p k + o σ e o 2 ; if r = 1 ; c 1 j = 1 n p k x l , j x h , j + o = 1 p + k x l , n p k + o x h , n p k + o σ e o 2 ; otherwise .
( X R 1 y ) M × 1 = j = 1 n p k y j + o = 1 p + k y n p k + o σ e o 2 ; if r = 1 j = 1 n p k x l , j y j + o = 1 p + k x l , n p k + o y n p k + o σ e o 2 ; otherwise .
( X R 1 Z ) M × N = j = 1 n p k z l , j + o = 1 p + k z l , n p k + o σ e o 2 ; if r = 1 j = 1 n p k x l , j z h , j + o = 1 p + k x l , n p k + o z h , n p k + o σ e o 2 ; otherwise .
( Z R 1 Z ) M × N = j = 1 n p k z l , j 2 + o = 1 p + k z l , n p k + o 2 σ e o 2 ; if r = c j = 1 n p k z l , j z h , j + o = 1 p + k z l , n p k + o z h , n p k + o σ e o 2 ; otherwise .
( Z R 1 y ) M × 1 = j = 1 n p k z l , j y j + o = 1 p + k z l , n p k + o y n p k + o σ e o 2 .
Case of no interactions: In this case, all the matrices of the linear system are simplified to reach p instead of p + k . l , h [ 1 , p ] :
( X R 1 X ) M × N = ( n p ) + σ e 1 2 + + σ e p 2 ; if r = c = 1 j = 1 n p x l , j + o = 1 p x l , n p + o σ e o 2 ; if r = 1 , c 1 j = 1 n p x l , j x h , l + o = 1 p x l , n p + o x h , n p + o σ e o 2 ; otherwise .
( X R 1 y ) M × 1 = j = 1 n p y j + o = 1 p y n p + o σ e o 2 ; if r = 1 j = 1 n p x l , j y j + o = 1 p x l , n p + o y n p + o σ e o 2 ; otherwise .
( X R 1 Z ) M × N = j = 1 n p z l , j + o = 1 p z l , n p + o σ e o 2 ; if r = 1 j = 1 n p x l , j z h , j + o = 1 p x l , n p + o z h , n p + o σ e o 2 ; otherwise .
( Z R 1 Z ) M × N = j = 1 n p z l , j 2 + o = 1 p z l , n p + o 2 σ e o 2 ; if r = c j = 1 n p z l , j z h , j + o = 1 p z l , n p + o z h , n p + o σ e o 2 ; otherwise .
( Z R 1 y ) M × 1 = j = 1 n p z l , j y j + o = 1 p z l , n p + o y n p + o σ e o 2 .

Appendix A.5. R Source Code

Finally, we give the source code in R for the SEM algorithm by considering our real data.
    library(lme4); # for lmer
    library(msm); # for rtnorm
    # initial model
    model_initial=lmer(RT~AoA*Lett_cat+(1+AoA|participants)+(1|items), data,
    control = lmerControl(optimizer="bobyqa", calc.derivs = FALSE));
 
    # random initialization of parameters
    b0=fixef(model_initial)[1];
    b1=fixef(model_initial)[2];
    b2=fixef(model_initial)[3];
    b3=fixef(model_initial)[4];
    sig_items=attr(VarCorr(model_initial)$items,"stddev");
    sig_parts=attr(VarCorr(model_initial)$participants,"stddev")[1];
    sig_parts_conver<-rep(0.0,itermax);
    sig_parts.AoA=attr(VarCorr(model_initial)$participants,"stddev")[2];
    sig_Err=attr(VarCorr(model_initial),"sc");
 
    Err=rnorm(n, mean=0, sd=sig_Err);
    Rand_parts=rep(0,n);
    Rand_items=rep(0,n);
    Rand_parts.AoA=rep(0,n);
    Rand_parts=rep(rnorm(np,mean=0, sd=sig_parts),each=ni);
    Rand_items=rep(rnorm(ni,mean=0, sd=sig_items),each=np);
    Rand_parts.AoA=rep(rnorm(np,mean=0, sd=sig_parts.AoA),each=ni);
 
    Lett_cat_1=vector(length=n)
    for(i in 1:n){
     Lett_cat_1[i]<-ifelse(test=(Lett_cat[i]==Lett_cat[2]),1,0)
    }
 
    u=runif(1,min=0,max=1);
    # acceptance prob based on the Hamiltonian MC
    p<-function(x,newx){
     a=min(1,exp(x-newx));
     if(u<a){
      x=newx;
     }
     else{
      x=x;
     }
    }
 
    # SEM (Algorithm 1)
    itermax=50; # max number of iteration for SEM
    burnin=10; # burnin length
    GS=5; # number of iterations for Gibbs~sampling
 
    # main loop
    for(iter in 1:itermax){
     # fixed part
     fixed=b0+b1*AoA+b2*Lett_cat_1+b3*AoA*Lett_cat_1;
     # Gibbs sampling
     for(j in 1:GS) {
      # Err with contraints
      newErr=rtnorm(length(Err),sd=sig_Err,
      lower=min(Err)-fixed-Rand_parts-Rand_items-Rand_parts.AoA,
      upper=max(Err)-fixed-Rand_parts-Rand_items-Rand_parts.AoA);
      p(Err,newErr);
      # Rand_parts with contraints
      for (a in 1:length(factor(participants))){
       Subset1=(participants == participants[a]) & censored;
       newRand_parts=Rand_parts;
       if (sum(Subset1)>0) {
        low=max(min(!is.na(data[participants == participants[a],5]))-Rand_items[Subset1]
        -Rand_parts.AoA[Subset1]-Err[Subset1]-fixed[Subset1]);
        upp=min(max(!is.na(data[participants == participants[a],5]))-Rand_items[Subset1]
        -Rand_parts.AoA[Subset1]-Err[Subset1]-fixed[Subset1]);
       upp=ifelse(upp-low>0,upp,upp+abs(upp-low)+100);
       newRand_parts[Subset1]=rtnorm(1,sd=sig_parts, lower=low,upper=upp);
       }
      }
      p(Rand_parts,newRand_parts);
      # Rand_items with contraints
      for (a in 1:length(factor(items))){
       Subset2=(items == items[a]) & censored;
       newRand_items=Rand_items;
       if (sum(Subset2)>0){
        low=max(min(!is.na(data[items == items[a],5]))-Rand_parts[Subset2]
        -Rand_parts.AoA[Subset2]-Err[Subset2]-fixed[Subset2]);
        upp=min(max(!is.na(data[items == items[a],5]))-Rand_parts[Subset2]
        -Rand_parts.AoA[Subset2]-Err[Subset2]-fixed[Subset2]);
        upp=ifelse(upp-low>0,upp,upp+abs(upp-low)+100);
        newRand_items[Subset2]=rtnorm(1,sd=sig_items, lower=low,upper=upp);
       }
      }
      p(Rand_items,newRand_items);
      # Rand_parts.AoA with contraints
      for (a in 1:length(factor(participants))){
       Subset1=(participants == participants[a]) & censored;
       newRand_parts.AoA=Rand_parts.AoA;
       if (sum(Subset1)>0) {
        low=max(min(!is.na(data[participants == participants[a],5]))-Rand_items[Subset1]
        -Rand_parts[Subset1]-Err[Subset1]-fixed[Subset1]);
        upp=min(max(!is.na(data[participants == participants[a],5]))-Rand_items[Subset1]
        -Rand_parts[Subset1]-Err[Subset1]-fixed[Subset1]);
        upp=ifelse(upp-low>0,upp,upp+abs(upp-low)+100);
        newRand_parts.AoA[Subset1]=rtnorm(1,sd=sig_parts.AoA, lower=low,upper=upp);
       }
      }
      p(Rand_parts.AoA,newRand_parts.AoA);
      # end of Gibbs sampling algorithm
     }
     # compute the new response vector
     RT[censored]=round(fixed[censored]+Rand_parts[censored]+Rand_items[censored]
     +Rand_parts.AoA[censored]+Err[censored],0);
     # train the model with the complete data
     model_final<-lmer(RT~AoA*Lett_cat+(1+AoA|participants)+(1|items), data,
     control = lmerControl(optimizer="bobyqa",calc.derivs = FALSE));
     # end of SEM algorithm
    }

References

  1. Donders, F.C. On the speed of mental processes. Acta Psychol. 1969, 30, 412–431. [Google Scholar] [CrossRef] [PubMed]
  2. Roelofs, A. One hundred fifty years after Donders: Insignts form unpublished data, a replication, and modeling of his reaction times. Acta Psychol. 2018, 191, 228–233. [Google Scholar] [CrossRef] [PubMed]
  3. Baayen, R.H.; Davidson, D.J.; Bates, D.M. Mixed-effects modeling with crossed random effects for subjects and items. J. Mem. Lang. 2008, 29, 390–412. [Google Scholar] [CrossRef] [Green Version]
  4. Perret, C.; Solier, C. Application of a Bayesian approach for exploring the impact of syllable frequency in handwritten picture naming. J. Cogn. Psychol. 2022, 34, 622–635. [Google Scholar] [CrossRef]
  5. Ratcliff, R. Methods for dealing with reaction time outliers. Psychol. Bull. 1993, 114, 510–532. [Google Scholar] [CrossRef]
  6. Van Selst, M.; Jolicoeur, P. A solution to the effect of the size on outlier elimination. Q. J. Exp. Psychol. 1994, 47, 631–650. [Google Scholar] [CrossRef]
  7. Barr, D.J.; Levy, R.; Scheepers, C.; Tily, H.J. Random effects structure for confirmatory hypothesis testing: Keep it maximal. J. Mem. Lang. 2013, 68, 255–278. [Google Scholar] [CrossRef] [Green Version]
  8. Dempster, A.P.; Rubin, D.B.; Tsutakawa, R.K. Estimation in Covariance Components Models. J. Am. Stat. Assoc. 1981, 76, 341–353. [Google Scholar] [CrossRef]
  9. West, T.B.; Welch, B.K.; Galecki, T.A. Linear Mixed Models, A Practical Guide Using Statistical Software; Chapman & Hall: New York, NY, USA, 2006. [Google Scholar]
  10. Jiang, J.; Wand, M.P.; Bhaskaran, A. Usable and precise asymptotics for generalized linear mixed model analysis and design. J. R. Stat. Soc. Ser. B 2022, 84, 55–82. [Google Scholar] [CrossRef]
  11. Faraway, J.J. Extending the Linear Model with R, 2nd ed.; CRC Press: Boca Raton, FL, USA, 2016. [Google Scholar]
  12. Jiang, J. Asymptotic Analysis of Mixed Effects Models; CRC Press: Boca Raton, FL, USA, 2017. [Google Scholar]
  13. McCulloch, C.E.; Searle, S.R.; Neuhaus, J.M. Generalized, Linear, and Mixed Models, 2nd ed.; John Wiley & Sons: New York, NY, USA, 2008. [Google Scholar]
  14. Stroup, W.W. Generalized Linear Mixed Models; CRC Press: Boca Raton, FL, USA, 2013. [Google Scholar]
  15. Bates, D.; Mächler, M.; Bolker, M.B.; Walker, C.S. Fitting Linear Mixed-Effects Models Using lme4. J. Stat. Softw. 2015, 67, 1–48. [Google Scholar] [CrossRef]
  16. Boik, J.R. The Analysis of Two-Factor Interactions in Fixed Effects Linear Models. J. Educ. Stat. 1993, 18, 1–40. [Google Scholar] [CrossRef]
  17. Yi, Q.; Panzarella, T. Estimating Sample Size for Tests on Trends across Repeated Measurements with Missing Data Based on the Interaction Term in a Mixed Model; Elsevier: Amsterdam, The Netherlands, 2002; Volume 23, pp. 481–496. [Google Scholar]
  18. Robin, G.; Josse, J.; Moulines, E.; Tibshirani, R. Main effects and interactions in mixed and incomplete data frames. J. Am. Stat. Assoc. 2019, 115, 1292–1303. [Google Scholar] [CrossRef] [Green Version]
  19. Giesselmann, M.; Schmidt-Catran, A.W. Interactions in fixed effects regression models. Sociol. Methods Res. 2022, 51, 1100–1127. [Google Scholar] [CrossRef]
  20. Martin, D.A.; Quinn, M.K.; Park, H.J. MCMCpack: Markov chain Monte Carlo in R. J. Stat. Softw. 2011, 42, 22. [Google Scholar] [CrossRef] [Green Version]
  21. Hadfield, J.D. MCMC methods for multi-response generalized linear mixed models: The MCMCglmm R package. J. Stat. Softw. 2010, 33, 1–22. [Google Scholar] [CrossRef] [Green Version]
  22. Khodadadian, A.; Parvizi, M.; Teshnehlab, M.; Heitzinger, C. Rational Design of Field-Effect Sensors Using Partial Differential Equations, Bayesian Inversion, and Artificial Neural Networks. Sensors 2022, 22, 4785. [Google Scholar] [CrossRef]
  23. Noii, N.; Khodadadian, A.; Ulloa, J.; Aldakheel, F.; Wick, T.; Francois, S.; Wriggers, P. Bayesian Inversion with Open-Source Codes for Various One-Dimensional Model Problems in Computational Mechanics. Arch. Comput. Methods Eng. 2022, 29, 4285–4318. [Google Scholar] [CrossRef]
  24. Matuschek, H.; Kliegl, R.; Vasishth, S.; Baayen, H.; Bates, D. Balancing Type I error and power in linear mixed models. J. Mem. Lang. 2017, 94, 305–315. [Google Scholar] [CrossRef]
  25. Dempster, A.; Laird, N.; Rubin, D. Maximum-likelihood from incomplete data via the EM algorithm. J. R. Stat. Soc. Ser. B 1977, 39, 1–38. [Google Scholar]
  26. Delyon, B.; Lavielle, M.; Moulines, E. Convergence of a stochastic approximation version of the EM algorithm. Ann. Stat. 1999, 1, 94–128. [Google Scholar] [CrossRef]
  27. Lavielle, M.; Mbogning, C. An improved SAEM algorithm for maximum likelihood estimation in mixtures of non linear mixed effects models. Stat. Comput. 2014, 24, 693–707. [Google Scholar] [CrossRef]
  28. Robbins, H.; Monro, S. A Stochastic Approximation Method. Ann. Math. Stat. 1951, 22, 400–407. [Google Scholar] [CrossRef]
  29. Bhatnagar, S.; Prasad, H.; Prashanth, L. Stochastic Recursive Algorithms for Optimization; Springer: Heidelberg, Germay, 2013; Volume 434, pp. 17–28. [Google Scholar]
  30. Celeux, G.; Diebolt, J. The SEM algorithm: A probabilistic teacher algorithm derived from the EM algorithm for the mixture problem. Comput. Stat. Q. 1985, 2, 73–82. [Google Scholar]
  31. Wei, G.; Tanner, M. A Monte-Carlo implementation of the EM algorithm and the Poor’s Man’s data augmentation algorithm. J. Am. Stat. Assoc. 1990, 85, 699–704. [Google Scholar] [CrossRef]
  32. Bennett, D.A. How can I deal with missing data in my study. Aust. N. Z. J. Public Health 2001, 25, 464–469. [Google Scholar] [CrossRef] [PubMed]
  33. Slaoui, Y.; Nuel, G. Parameter Estimation in a Hierarchical Random Intercept Model with Censored Response: An Approach using a SEM Algorithm and Gibbs Sampling. Sankhya Indian J. Stat. 2014, 76, 210–233. [Google Scholar] [CrossRef]
  34. Foulley, J.L. Mixed Model Methodology. Part I: Linear Mixed Models. In Technical Report; Université de Montpellier: Montpellier, France, 2015. [Google Scholar]
  35. Grund, S.; Lüdtke, O.; Robitzsch, A. Multiple imputation of missing data in multilevel models with the R package mdmb: A flexible sequential modeling approach. Behav. Res. Meth. 2021, 53, 2631–2649. [Google Scholar] [CrossRef]
  36. Santos, F. L’algorithme EM: Une Courte Présentation; CNRS, UMR, 5199; PACEA: Talence, France, 2015. [Google Scholar]
  37. Celeux, G.; Diebolt, J. Une Version de Type Recuit Simule de L’algorithme EM. Ph.D. Thesis, INRIA, Le Chesnay, France, 1989. [Google Scholar]
  38. Celeux, G.; Chauveau, D.; Diebolt, J. On Stochastic Versions of the EM Algorithm. Ph.D. Thesis, INRIA, Le Chesnay, France, 1995. [Google Scholar]
  39. Celeux, G.; Diebolt, J. A stochastic approximation type EM algorithm for the mixture problem. Stoch. Int. J. Probab. Stoch. Process. 1992, 41, 119–134. [Google Scholar] [CrossRef] [Green Version]
  40. Richard, F.; Samson, A.; Cuénod, A.C. A SAEM algorithm for the estimation of template and deformation parameters in medical image sequences. Stat. Comput. 2009, 19, 465–478. [Google Scholar] [CrossRef]
  41. Panhart, X.; Samson, A. Extension of the SAEM algorithm for nonlinear mixed models with 2 levels of random effects. Biostatistics 2008, 10, 121–135. [Google Scholar] [CrossRef] [Green Version]
  42. Betancourt, M. A conceptual introduction to Hamiltonian Monte Carlo. arXiv 2017, arXiv:1701.02434. [Google Scholar]
  43. Metropolis, N.; Rosenbluth, A.W.; Rosenbluth, M.N.; Teller, A.H.; Teller, E. Equation of State Calculations by Fast Computing Machines. J. Chem. Phys. 1953, 21, 1087–1092. [Google Scholar] [CrossRef] [Green Version]
  44. Hastings, W.K. Monte Carlo sampling methods using Markov Chains and their applications. Biometrika 1970, 57, 97–109. [Google Scholar] [CrossRef]
  45. Roberts, O.G.; Rosenthal, S.J. General state space Markov chains and MCMC algorithms. Probab. Surv. 2004, 1, 20–71. [Google Scholar] [CrossRef] [Green Version]
  46. Andrieu, C.; De Freitas, N.; Doucet, A.; Jordan, M. An Introduction to MCMC for Machine Learning. KAP 2003, 50, 5–43. [Google Scholar]
  47. Perret, C.; Bonin, P. Which variables should be controlled for to investigate picture naming in adults? A Bayesian meta-analysis. Behav. Res. Methods 2019, 51, 2533–2545. [Google Scholar] [CrossRef]
  48. Schafer, L.J.; Graham, W.J. Missing Data: Our View of the State of the Art. Psychol. Methods 2002, 7, 147–177. [Google Scholar] [CrossRef]
  49. Enders, C.K. Applied Missing Data Analysis; Guilford Press: New York, NY, USA, 2010. [Google Scholar]
  50. Little, R.J.; Rubin, D.B. Statistical Analysis with Missing Data; John Wiley & Sons: Hoboken, NJ, USA, 2019; Volume 793. [Google Scholar]
  51. Zhang, X. Tutorial: How to Generate Missing Data For Simulation Studies. 2021. Available online: https://files.osf.io/v1/resources/rq6yb/providers/osfstorage/60e93adf600da501810a8ea4?format=pdf&action=download&direct&version=1 (accessed on 8 February 2023).
Figure 1. Value of the parameter estimates at each the SEM iteration with a burn-in period M = 10, we separate the parameters according to their values in (a), we plot b0, b1, sig1 and sig2 and in (b), we plot b2, b3, sig3 and sig4.
Figure 1. Value of the parameter estimates at each the SEM iteration with a burn-in period M = 10, we separate the parameters according to their values in (a), we plot b0, b1, sig1 and sig2 and in (b), we plot b2, b3, sig3 and sig4.
Entropy 25 00473 g001
Figure 2. Normalized residual presented; (a)-standard simulated data, (b)-SEM algorithm, (c)-SAEM algorithm, (d)-MCMC algorithm.
Figure 2. Normalized residual presented; (a)-standard simulated data, (b)-SEM algorithm, (c)-SAEM algorithm, (d)-MCMC algorithm.
Entropy 25 00473 g002
Table 1. Comparison of the SEM algorithm with the SAEM and MCMC algorithms in the complete model with the different percentages of incomplete data where the number of iterations was 50. sig 1 , sig 2 , sig 3 and sig 4 represent the standard deviations of the participants, items, the I FR part and the error, respectively. The percentages on the left present the MAR, and on the right are the percentages of censored data.
Table 1. Comparison of the SEM algorithm with the SAEM and MCMC algorithms in the complete model with the different percentages of incomplete data where the number of iterations was 50. sig 1 , sig 2 , sig 3 and sig 4 represent the standard deviations of the participants, items, the I FR part and the error, respectively. The percentages on the left present the MAR, and on the right are the percentages of censored data.
( 0 % , 0 % ) ( 0 % , 5 % ) ( 0 % , 10 % ) ( 0 % , 20 % )
Θ ^ i ReferenceSEMSAEMMCMCSEMSAEMMCMCSEMSAEMMCMC
b 0 104510441045106810441043101010471047822.8
b 1 93.1693.5488.5148.8293.9189.3741.2593.5188.8477.96
b 2 −52.28−51.30−42.33−36.75−52.50−47.3051.49−30.27−51.6618.26
b 3 2.803.17−0.186.924.435.52−14.44−1.795.02−18.20
sig 1 14.5714.6514.435.2914.6512.5412.5614.7012.1711.30
sig 2 230.8231.7225.9350.5231.1223.0430.5233.2219.0538.7
sig 3 9.799.819.842.809.808.138.619.819.128.94
sig 4 282.0282.0275.5283.1282.0269.4277.8282.0253.5253.9
MAE 00.4213.63728.070.5054.62851.843.9356.57583.64
LCor 10.9990.9990.9920.9990.9990.9710.9990.9990.915
RCor 1110.904110.738110.833
( 5 % , 0 % ) ( 5 % , 5 % ) ( 5 % , 10 % ) ( 5 % , 20 % )
Θ ^ i SEMSAEMMCMCSEMSAEMMCMCSEMSAEMMCMCSEMSAEMMCMC
b 0 1044103810041069103710191069103696310781038780
b 1 93.3094.3393.7987.5790.0452.8288.0790.9045.2986.5290.7580.88
b 2 −46.69−53.57−49.05−49.96−42.94−28.23−52.68−47.9455.58−32.18−50.2126.67
b 3 1.044.586.558.271.278.639.706.75−11.903.825.61−15.79
sig 1 15.1913.6516.5915.2517.028.3415.3613.9316.1915.7912.804.10
sig 2 231.0223.7213.6229.8218.7332.4228.9216.0412.7232.4211.9518.8
sig 3 9.947.9311.079.978.635.059.957.5210.6110.118.426.56
sig 4 282.4281.4347.6282.6275.0349.4282.6268.7342.2282.6253.2317.6
MAE1.1462.64616.784.9705.51634.435.0896.28762.038.1188.10888.97
LCor0.9990.9990.9960.9990.9990.9910.9990.9990.9710.9990.9990.911
RCor111110.928110.738110.809
( 10 % , 0 % ) ( 10 % , 5 % ) ( 10 % , 10 % ) ( 10 % , 20 % )
Θ ^ i SEMSAEMMCMCSEMSAEMMCMCSEMSAEMMCMCSEMSAEMMCMC
b 0 10701037981106910359991070103493810801037759
b 1 86.5994.0590.5287.5890.2750.90288.1090.8046.4386.2990.3381.31
b 2 −45.96−50.70−62.01−45.00−39.50−44.14−47.79−44.5745.21−28.34−48.0123.38
b 3 6.795.2612.256.691.4614.458.136.78−8.832.535.81−16.26
sig 1 14.8024.434.3614.9727.3525.0015.1522.0930.5115.9321.3918.64
sig 2 228.7220.1205.1230.0215.3323.1229.1212.5401.1232.5208.7506.5
sig 3 9.7911.465.949.8912.106.3369.8810.1613.8710.1110.508.73
sig 4 282.4279.6398.0282.5273.2389.6282.5267.2379.8282.5251.8354.3
MAE5.6084.69930.125.3878.27940.155.4248.19168.858.7479.63693.20
LCor0.9990.9990.9900.9990.9990.9870.9990.9990.9690.9990.9990.908
RCor110.904110.976110.833110.833
Θ ^ i SEMSAEMMCMCSEMSAEMMCMCSEMSAEMMCMCSEMSAEMMCMC
b 0 10451035896104510449081045103685710511039687
b 1 93.3494.05100.893.7080.7864.4394.0590.1457.3592.7889.4191.59
b 2 −35.09−38.362.54−34.54−128.8721.66−36.11−34.7997.19−19.91−39.2485.55
b 3 −1.821.90−16.60−1.5539.61−13.36−0.364.00−31.09−4.733.67−42.98
sig 1 12.1226.8829.4912.4411.7039.8912.6225.5833.2513.6226.5322.49
sig 2 230.9211.4204.7231.8118.2315.9231.3204.4387.9233.1200.9485.3
sig 3 8.8411.338.958.985.1511.489.0310.9110.939.4111.766.227
sig 4 282.3276.2465.9282.3265.3453.6282.3264.0442.2282.4249.2411.6
MAE3.2618.00257.073.37432.8367.382.99410.90593.016.34312.47117.2
LCor0.9990.9990.9690.9990.9910.9690.9990.9990.9440.9990.9980.873
RCor110.97610.92850.928110.761110.833
We marked in bold for each case the closest value to the reference among the three considered methods (the reference vector here was the simulated one that did not contain any missed value).
Table 2. Comparison of the initial parameters in the KE method versus the parameters estimated after the application of the SEM, SAEM and MCMC algorithms. The initial parameters presented here are the ones obtained from the model that contained 1066 NA (NA represent the incomplete data; here, we had 115 censored data and 951 missing data.
Table 2. Comparison of the initial parameters in the KE method versus the parameters estimated after the application of the SEM, SAEM and MCMC algorithms. The initial parameters presented here are the ones obtained from the model that contained 1066 NA (NA represent the incomplete data; here, we had 115 censored data and 951 missing data.
Θ ^ KESEMSAEMMCMC
b 0 852.5875.7892.3341298
b 1 146.6134.2120.4−161.4
b 2 158.6126.179.63−335.3
b 3 −58.12−42.08−26.72146.6
sig 1 184.3196.0203.5216.9
sig 2 133.0127.894.87221.05
sig 3 37.0528.6732.8466.13
sig 4 279.5302.5262.9516.6
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Zakkour, A.; Perret, C.; Slaoui, Y. Stochastic Expectation Maximization Algorithm for Linear Mixed-Effects Model with Interactions in the Presence of Incomplete Data. Entropy 2023, 25, 473. https://doi.org/10.3390/e25030473

AMA Style

Zakkour A, Perret C, Slaoui Y. Stochastic Expectation Maximization Algorithm for Linear Mixed-Effects Model with Interactions in the Presence of Incomplete Data. Entropy. 2023; 25(3):473. https://doi.org/10.3390/e25030473

Chicago/Turabian Style

Zakkour, Alandra, Cyril Perret, and Yousri Slaoui. 2023. "Stochastic Expectation Maximization Algorithm for Linear Mixed-Effects Model with Interactions in the Presence of Incomplete Data" Entropy 25, no. 3: 473. https://doi.org/10.3390/e25030473

APA Style

Zakkour, A., Perret, C., & Slaoui, Y. (2023). Stochastic Expectation Maximization Algorithm for Linear Mixed-Effects Model with Interactions in the Presence of Incomplete Data. Entropy, 25(3), 473. https://doi.org/10.3390/e25030473

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop