1. Introduction
As modern technology becomes more advanced, the study of the reliability of HMS has become more significant, developing into an essential part of the research on reliability theory. The HMS is a general term for systems where the human is the subject and various types of machines are controlled. Due to the development of automation and microelectronics, the allocation of human–machine functions is continually shifting to the machine side. The high precision and performance of machines increase the importance of human work responsibility, and there is the risk of major accidents caused by HR. Additionally, machines are affected by external environmental contingencies, such as earthquakes, floods, fires, and so on. When these destructive contingencies occur, numerous components in the overall system fail simultaneously, which is known as CCF.
By studying the history of reliability theory, it is easy to see that many reliability problems have been solved by establishing reliability models in which the SVT plays an essential role. The SVT was first proposed by Cox [
1] in 1955 and was first introduced into the reliability theory by Gaver [
2]. Following this research approach, researchers such as Abbas and Kuo [
3], Yang and Dhillon [
4], Sridharan and Mohanavadivu [
5], Asadzadeh and Azadeh [
6], and Wang [
7] studied various human–machine reparable systems.
In general, the mathematical model established by the SVT is described by a finite or infinite set of partial differential integral equations with integral boundary conditions. Therefore, obtaining exact solutions is quite challenging. Because of this, some papers on reliability models assume that TDS converges to SSS. However, they do not address whether or not this assumption is correct. In 2003, Gupur [
8] was the first to apply
-semigroup theory to investigate the TDS of an HMS consisting of an active and a standby component by the SVT. The strong [
9] and exponential [
10] convergence of the TDS to its SSS is then obtained separately. Wang and Xu [
11] examined the well-posedness and asymptotic behavior of the TDS of an HMS with two parallel working components and a standby component. As systems become more complex, these simple systems are no longer adequate for engineering needs. In addition, as equipment becomes more reliable, the contribution of human error to system problems is relatively more significant. HR and CCF are vital issues in system reliability. Therefore, Chung [
12], Narmada and Jacob [
13], Hajeeh [
14], Liu et al. [
15], Shneiderman [
16], and others (see the references therein) have examined the reliability of systems with HR and CCF. Yang and Dhillon [
4] established a complex HMS consisting of
active components and
standby components with HR and CCF. Xu et al. [
17] identified that the above system exists with a unique TDS, and it converges to the SSS. Other than the above results, there are no further results on this model’s dynamic analysis.
In this paper, first of all, we study the spectral properties of the underlying operator and demonstrate that, in a strip region on the left-half complex plane, it has a maximum number of finitely many eigenvalues and has an algebraic multiplicity of 1, of which 0 is strictly the dominant eigenvalue. Then, we show that the semigroup generated by the underlying operator of the model is quasi-compact and it is exponentially convergent to a projection operator. By studying the essential growth bound of the operator semigroup, it is shown that 0 is a pole of order 1. These results give an explicit expression for the projection operator by the residue theorem of the operator form. Finally, we provide the asymptotic expressions of the TDS of the system.
2. Mathematical Model of the System
The assumptions and symbols associated with our mathematical model are as follows.
In the system, there are n active components and m standby components.
When one of the operating components fails, the standby component switches into operation; when all components fail, the system fails.
A CCF or a HR can trigger system failure from any of the system operable states. represents the constant CCF rate from state i to state m + n + 1; denotes the constant critical HR rate from state i to state m + n + 2; for i = 0, 1, 2, …, m + n − 1.
common-cause and other failure rates are constant. denotes the constant hardware failure rate of a unit in state i; i = 0, 1, 2, …, m + n − 1.
The failed system repair times are arbitrarily distributed; represents the system’s time-dependent repair rate when the system is in state j and satisfies for j = m + n, m + n + 1, m + n + 2.
The repaired component or system is as good as new. denotes the constant repair rate of a failed unit in state i, where i = 1, 2, …, m + n − 1.
All failures including HR are statistically independent, and the switchover mechanism is perfect and instantaneous.
Based on the above assumptions and descriptions, the state transition diagram of the system
Figure 1 can be presented as below.
According to Yang and Dhillon [
4], the following partial differential integral equations describe the mathematical model of the HMS with HR and CCF.
where
and
. We use this denotation throughout the article.
In the following, we introduce some notations:
Take the following Banach space
as a state space:
Define the operators and their domain.
Then, (
7) and (
8) can be written as an abstract Cauchy problem:
X:
3. Well-Posedness of (9)
Theorem 1. generates a positive contraction - semigroup when .
The proof of Theorem 1 is omitted.
The dual space of
X is given by
Obviously,
is a Banach space. Let
Then,
, and Theorem 1 ensures that
. For
, take
; then,
, and
In (
10), we use
Equation (
10) implies that
is a conservative operator with respect to the set
Because of
and by using the Fattorini theorem [
18], we obtain the following result.
Theorem 2. is isometric for , i.e., We can obtain the system’s well-posedness from Theorems 1 and 2.
Theorem 3. If , then (9) has a unique positive TDS satisfying 4. Spectrum of the Operator
Lemma 1. If , then has at most finite eigenvalues in ; the geometric multiplicity of each eigenvalue is one, and 0 is a strictly dominant eigenvalue.
Proof. i.e.,
By substituting (
16)–(
18) into (
19), we deduce
From (
13) and (
14), we obtain
By Cramer’s rule, we have
Here,
U is the coefficient matrix of the above equations, and
is the matrix where the
i column of the coefficient matrix is replaced by the following vector:
By inserting (
20)–(
23) into (
12), we obtain
where
and
.
If
, then (
20)–(
23) imply
; that is to say,
. Thus,
is not an eigenvalue of
.
If
, then (
24) gives
By (
20)–(
23) and the condition of Lemma 1 we estimate
By (
26) and (
27), it is not difficult to know that all zeros of
in
are eigenvalues of
. Since
is analytic in
, it follows that
has, at most, countable isolated zero points in
from the zero-point theorem of the analytic function.
In the following, we verify the above results. If
has infinitely many zero points in
, and we assume that they are
, then we know that there is a convergent subsequence by the Bolzano–Weierstrass theorem. Without losing generality, assume
such that
,
By inserting
into (
25), we obtain
By
and the Riemann–Lebesgue theorem, we know that
From (
29)–(
31) and taking the limit
in (
29), we obtain that
. Obviously, this is a contradiction. Therefore,
has at most finite zero points in
; in other words,
has finite eigenvalues at most in
. Furthermore, according to (
20)–(
23), the eigenvectors corresponding to each
generate a linear space of one dimension. That is, the geometric multiplicity of each
is one. □
Remark 1. It is not difficult to prove that . Hence, 0 is the eigenvalue of with a geometric multiplicity of one. Because has finite eigenvalues and the real part of all non-zero eigenvalues is strictly less than 0 in Δ, 0 is a strictly dominant eigenvalue of .
Proof. denotes the matrix where the
i column of
is replaced by the vector
According to the properties of the determinant, we obtain
Lemma 2. is given bywhere Lemma 3. If , then has at most finite eigenvalues in , and if the geometric multiplicity of each eigenvalue is one, then 0 is a strictly dominant eigenvalue.
Proof. Consider
, i.e.,
By solving (
36), we deduce
Multiplying both sides of
by (
38), taking the limit
, and using (
37), we obtain
According to (
34) and (
35), we can obtain
The above equations are written in matrix form as
where
Thus, by Cramer’s rule, we have
Here,
D is the coefficient matrix of the above equations, and
is the matrix where the
i column of the coefficient matrix is replaced by the following vector:
We can substitute (
44) into (
33) to obtain
If
, (
38) and (
44) means
; that is,
, which indicates that
is not an eigenvalue of
.
Now, we prove that
. Because
and
we have
Therefore, (
46) is equivalent to
, which is to say,
According to (
38), we can estimate (assume
Combining (
48) with (
44), we obtain
(
47) and (
48) implies that all zeros of
in
are the eigenvalues of
. Because
is analytic in
, we know from the zero-point theorem for analytic functions that
has at most countable isolated zero points in
. Because this is the same as in Lemma 1, we can obtain that
has a finite number of zero points at most in
; in other words,
in
has at most finite eigenvalues. □
Remark 2. According to we can obtain that . Thus, 0 is the eigenvalue of with geometric multiplicity of one.
5. Asymptotic Behavior of the TDS of (9)
From
Section 3, we can obtain that the operator
also generates a positive contraction
semigroup,
. In this section, first of all, we prove that
is a quasi-compact operator. Since
and
are compact operators, it is obtained that the semigroup
generated by
is a quasi-compact
semigroup according to the perturbation of the quasi-compact operators. Next, we prove that
converges exponentially to a projection operator and provide a concrete expression for this convergence. Therefore, we obtain that the TDS of (
9) converges exponentially to its steady-state solution.
Proposition 1. If is a solution of (50), thenwhere is determined by (5)–(7). Theorem 4. If is Lipschitz-continuous and satisfies , then is a quasi-compact semigroup in X.
Proof. First of all, we define two operators for
From [
19] in Theorem 1.35 and the definition of
, we know that we only need to prove Condition 1 in Theorem 1.35 [
19]. For
, we set
; then
is a solution of (
50). Therefore, according to Proposition 1 we have, for
,
where
,
. Thus,
.
In the following, we estimate each term in (
53). According to the properties of the semigroup and the boundary conditions, we have
From (
54)–(
56), we can estimate the first, third, and fifth terms of (
53):
Using the boundary condition and Proposition 1, we obtain
Combining (
57)–(
62) with (
53), we obtain
If
then
and
. Thus,
For
, in the same way as (
63), the second, fourth, and sixth items in (
64) are obtained as follows:
In the following, we estimate the other three terms in (
64), and by using Propositions 1 and (
54)–(
56), we obtain
When
, we have
, (
65)–(
69)
. Therefore, (
69) and (
63) show that
is a compact operator.
Now, by the definition of
, we have, for
,
From these results, we obtain
Combining the definition of the quasi-compact operator (see Gupur [
19], Definition 1.85), we obtain that
is a quasi-compact
semigroup in
X. Obviously,
and
are compact operators on
X (see Gupur [
19], Definition 1.7). According to this result and Proposition 2.9 in Nagel [
20], we obtain the following results. □
Corollary 1. If the conditions are the same as in Theorem 4, then is a quasi-compact - semigroup in X.
By Lemmas 1 and 3, we know that the algebraic multiplicity of 0 is one.
Thus, according to Theorem 1, Lemma 1, Lemma 3, and Corollary 1 with Theorem 1.90 [
19], the following result is concluded.
Theorem 5. If is Lipschitz-continuous and satisfies , then there is a positive projection operator and appropriate constants such thatwhere and is a circle of sufficiently small radius with its center at 0. From Theorem 1, Lemma 1, and Corollary 1, we have
This means that only 0 is the spectral point of on the imaginary axis.
Next, we calculate the explicit expression of the project operator .
Theorem 6. If the conditions are the same as in Theorem 4, then the TDS of (9) converges exponentially to its steady-state solution, i.e., Proof. From this, together with Proposition 2.10 of Nagel and Engel [
16], we have
(i.e.,
, the essential growth bound of
(i.e.,
), satisfying
Since
and
are compact operators, by Proposition 2.12 in [
16], we have
Using this result and combining it with Theorem 4 and Corollary 2.11 of Engel and Nagel [
21], we obtain that 0 is a pole of
of order 1. Therefore, from Theorem 5 and the residue theorem, we have
To calculate this limit, we need to give the expression of .
For
, consider the equation
; that is,
Then, by substituting (
79) into (
70), we can derive
Furthermore, (
80), (
71), and (
72) can be written as
where
According to Cramer’s rule,
Here,
is the coefficient matrix of the above equations, and
is the matrix where the
column of the coefficient matrix is replaced by the following vector:
Simplifying
yields
where
Therefore,
where
.
Substituting (
83)–(
85) into (
74)–(
76), respectively, and using (
77), we derive
In particular, for
, we obtain
According to Theorem 3, (
89), and Theorem 6, we have
□
6. Asymptotic Expression of the TDS of (9)
Firstly, we can prove that the algebraic multiplicity of all eigenvalues of
in
is 1. In fact, if this state is wrong, then the algebraic multiplicities of all eigenvalues of
are greater than 1 [
22]. Without losing generality, suppose that their algebraic multiplicities are equal to 2; then,
has a solution in
, where
is an eigenvector in Lemma 1, namely
. In (
90), on either side of the role of
, where
is the eigenvector in Lemma 3, i.e.,
, we launch
which contradicts
. Therefore, the algebraic multiplicity of all eigenvalues of
is one in
.
Without losing generality, suppose that there are
real eigenvalues of
in
, and they are
Thus, combining Theorem 1.89 in [
19] (see also Nagel [
20]) with Theorem 1, we obtain
Here,
is a circle with a sufficiently small radius and a center
. Since the algebraic multiplicity of
is 1,
is a pole of
of order one. Thus, we know by the residual theorem that
The expression of
is given as follows:
where
is given by (
83)–(
88). By (
95) and (
96), we can determine all
.
where
and
satisfy
,
Finally, we deduce the following main results.
Theorem 7. If and are Lipschitz-continuous, then the TDS of (9) can be written as where are isolated eigenvalues of in and .
7. Numerical Results
In this section, we discuss some reliability indices of the system through specific examples, such as the system availability , reliability , and , and analyze the impact of changes in system parameters on system reliability indices. First of all, without loss of generality, let us consider the case of two active units and two standby units system, i.e., and , and assume that the repair time of the system is gamma-distributed and the repair rate is constant, i.e., . The influence of parameter changes on the instantaneous reliability index of the system is discussed below.
In
Figure 2, we describe the influence of different
changes with time
t on the instantaneous availability of the system (
is another parameter of the gamma distribution). It is easy to see from
Figure 2 that
decreases rapidly with increased time. After the system runs for a long time, it stabilizes and reaches a fixed value.
In the following, we assume (i.e., the repair time of the system is exponentially distributed) and continue to discuss the influence of different and on the instantaneous availability of the system.
Figure 3a,b show that
decreases with increases in
and
. In addition, as time goes to infinity, the instantaneous availability of the system converges to a certain value.
Figure 4 reveals the effect of different
on the instantaneous availability of the system. It is not difficult to find that
increases with increased
.
Figure 5 indicates the effect of
and
on the system reliability and mean time to failure (
). We note that
(
Figure 5a) and
(
Figure 5b) decrease as
increases. Obviously, reliability vanishes as time goes to infinity.
Obviously, a similar conclusion can be drawn for the system’s failure frequency,
, and the renewal frequency,
. Therefore, it can be seen from the above figure and discussion that when the time tends to infinity, the instantaneous reliability index of the system tends to a constant value, which verifies the main results obtained in
Section 5.
8. Conclusions
In this paper, we studied the dynamical solution problem of human–machine systems with human error and common-cause failure. We started from theory and used the theory of semigroups in functional analyses to model the system. The integral-differential equation was transformed into an abstract Cauchy problem in Banach space. Then, we proved the well-posedness of (
9), studied the asymptotic behavior of its time-dependent solution, and showed that the time-dependent solution converges exponentially to its steady-state solution, obtaining asymptotic expressions for the time-dependent solution. In addition, the influence of each parameter on the system reliability were analyzed through concrete numerical examples. Therefore, engineers can design a more reliable, safe, and cost-effective system by using the results obtained in this paper. To a certain extent, it provides a theoretical basis for system reliability management and optimal scheduling.
If we know the spectral distribution of in
and , we may directly estimate in Theorem 7, which is important for engineers. Based on our knowledge of this subject, we believe has a continuous spectrum in . However, it is necessary to verify this and investigate more results.