1. Introduction
The association of small chemicals or large biological molecules in a rapid, specific way is an essential step in various chemical or biological processes ranging from enzyme catalysis to regulation of immune responses [
1,
2]; extracting key thermodynamic information from those reactions will greatly benefit the understanding of those chemical or biological processes. Two of the most prominent thermodynamic properties that define a reaction are association rates and reaction intermediate state concentrations [
3].
The rate of association spans a range from
to
; it is limited either by diffusion or subsequent chemical processes such as conformational rearrangement [
1]. Association rates can be categorized as time-dependent and time-independent [
2,
4]. Time-dependent reaction rate constants are usually greatly influenced by diffusion, are more susceptible to disruptions in reaction environment, and can thus be described by non-equilibrium radial distribution functions [
5,
6]. Time-independent reactions rates look at molecule diffusion and interactions on a mass scale that exhibit overall reaction kinetics. Time-independent reaction rates can be modeled through approximation [
7].
Association reactions between chemical or biological molecules often involve an intermediate state (also known as transitional state) whose existence greatly expedite the efficiency of molecule association, thus increasing chemical or biological reaction rates [
1]. Intermediate states in the reaction pathway, also sometimes called transition states, refer to one or a set of configurations that possess rotational or translational correlations between the interacting species [
3,
8]. Intermediate states are the most unstable species in the reaction pathway, whose bonds are in the process of been constantly broken and remade [
8]. They are key to elucidating chemical or biological reaction pathways as their kinetic and structural characteristics shed light on the reaction process and energy landscape.
In order to reveal the structural or kinetic traits of intermediate states, kinetic experiments such as stop-flow were used in conjunction with fluorescence spectroscopy to capture this transient species [
9,
10,
11,
12]. Experimental detection of the transitional states can be done for some reactions, but for others, especially reactions involving large biological molecules, experimental detection of transient states might prove to be extremely difficult [
1,
8]. More generally, experimental data can only provide the kinetic and structural information of the more abundant states, which often exist as dominant fractions and are much easier to detect using regular experimental tools. Even though experimental data were unable to provide specific kinetic parameters associated with intermediate states, its existence can often be presumed based on a hyperbolic dependence of reaction time to reactant concentration [
8]. We speculate that if experimental evidence confirm the existence of the intermediate state but were unable extract the kinetic parameters involving the intermediate state turnover, it is crucial to reveal the information regarding the intermediate states via methods other than experimental. Therefore, we strive to build an efficient computational framework for extracting information encoded in experimental data regarding intermediate states [
13].
Here, we will present a novel method making use of artificial neural-networks to gather kinetic and concentration information of a possible transitional state in a biological or chemical reaction system [
14]. The objective of our method is to use computational methods to extract association/dissociation rate constants and concentration dynamics of the intermediate state with the aid of experimental datasets. To do so, firstly, we simulated the reaction systems using differential equations, secondly, we fit the modeled differential equations to experimental data, using neural networks as an approximator to differential equations [
15]. Finally, we utilized global search algorithm Particle Swarm Optimization (PSO) to solve the non-linear differential equations derived from reactions. We chose PSO because it can provides us a global best solution on the parameter space, other algorithms, such as gradient descent are prone to encounter local optima, are single based or have unknown search space issues; therefore, they are unable to produce an optimal solution [
16]. In sum, we have developed a simple, but effective model to simulate the reactions as differential equations using the adaptive strength of neural networks, and used a rapid global search method (PSO) [
17,
18] to minimize unsupervised errors combined from simulating reactions and from the fitting of experimental data to neural networks.
The main contributions of this paper are as follows:
Based on chemical reaction kinetics and experimental presumptions, a reaction model of differential equations involving intermediate states were constructed describing chemical or biological reactions.
Utilizing Artificial Neural Network (ANN) to model differential equations describing reactions. Adjusting ANN’s weights to best model the chemical reactions while minimizing the difference between the output of the neural network and experimentally measured data using swarm intelligence algorithms. Based on ANN and predefined reaction models, a novel model-based parameter extraction method called ANN-Dependent Kinetic Parameter Extraction method (AKPE) emerges that could obtain kinetic rate constants associated with lowly-populated species with high efficiency and low computational cost.
Extensive experiments employing the proposed method were carried out upon real-life examples. Kinetic parameters of biochemical reactions regarding the intermediate state were given. The derived parameters were further analyzed and subsequent in-depth kinetic and structural characterization of the intermediate state were accomplished.
Validation of our method were carried out by comparing kinetic rate constants produced by AKPE with corresponding experimental kinetic rates and stochastic simulation results, consistency is achieved. The agreement between parameter values produced by AKPE and other methods confirms the validity of AKPE.
The remainder of this paper is organized as follows.
Section 2 (Related Works) reviewed the area of chemical reaction simulation and introduced some basic ideas in model-based reaction parameter prediction.
Section 3 (Materials and Methods) proposed the general architecture and the computational details of our method. In
Section 4, (Results), an real-life case study using the C terminal domain of SARS-CoV virus’s main protease was analyzed using the proposed method, extensive biological characterization was achieved.
Section 5 (Discussion) validated the proposed method by comparing it with other algorithms or experimental data. Finally,
Section 6 (Conclusions) and
Section 7 (Future Perspective) concluded and provided future perspectives of this study.
2. Related Works
This section reviews some reaction simulation approaches and highlights the basic features of AKPE.
The two most widely used reaction simulation methods are deterministic and stochastic. Deterministic methods make use of reaction models, reaction models describe the rates of reactions and can demonstrate the concentration fluctuations of reactants. Large scale reactions that involve large quantities of reactants can be described by deterministic linear or non-linear ordinary differential equations in homogenous reaction space [
19,
20]. Small scale association reactions that can be described by a limited number of molecules, or when the key steps of the reactions depend on the behavior of a limited number of molecules are more accurately simulated by stochastic simulation algorithms (SSA), chemical reaction models can be incorporated into both of the aforementioned simulation categories [
21].
2.1. Deterministic Simulation Methods
For macroscopic deterministic reaction systems, differential equations can often describe reaction system and reactants by:
In the case of single reactant:
where
x is the reaction time and
y(
x) denotes reactant concentration as a function of time,
G is the differential function based on the chemical properties of the reaction system,
denotes the nth derivative of the term with respect to reaction time and is also the order of the reaction.
Sets of differential equations involving
k reacting species are presented below:
where
x is the reaction time,
to
are the concentrations of the
k reactants in the reaction system, the system of differential equations take into account the interaction between the
k species,
n is the reaction order.
2.2. Neural Networks as Universal Approximators
Neural networks can be universal approximators [
14]; it can be ideal to use neural networks to simulate reaction-derived differential equations. Methods utilizing feed forward neural networks was applied to solving differential equations describing gas-phase or liquid-phase reaction systems in steady states. The capability of neural networks as approximators of reaction-derived differential equations were exploited, postulating them as solutions for a given deferential equation with an unsupervised error. Neural networks can provide an expression for the analytical solution of differential equations over the entire definition domain. As it can be expressed as a set of compact mathematical expression, therefore, it can include any number of parameters and variables, making it adequate in solving optimization problem with an accuracy of around MSE/
to
. Although other analytical approaches at solving differential equations might arrive at identical solutions, the neural networks approach to solving differential equations is much more computationally efficient since it allows the evaluation of solutions at any desired number of input points with much less memory usage and computing time [
22,
23].
When describing gas-phase or liquid-phase reaction systems in non-equilibrium, the capability of neural networks to approximate stochastic diffusional systems can be revealed. A special class of neural networks, named Stochastic Neural Networks(SNN), are capable of approximating stochastic processes. SNN incorporated statistical randomness into its mathematical expression, thus are capable of generating random sample functions. SNN can be canonical solutions to random processes describing reaction systems just as ANN can represent solutions to ordinary differential equations. SNN is more memory efficient at representing solutions to random differential equations, since they are capable of generating a collection of sample functions while normal ANN only generate a single sample function. Some example studies have demonstrated SNN’s applicability at approximating solutions of non-linear random differential Equations [
24,
25].
2.3. Stochastic Simulation Methods
In microscopic systems where only a small number of molecules are present, stochastic diffusion takes a crucial role, therefore, differential equations describing the behavior of molecules in such a system will take diffusion and the location of each molecule into account. In such systems, stochasticity and variability is an perpetual problem to address. It has been shown experimentally that in some of these reaction systems, stochasticity of reaction resulted in large variability of reaction rates [
1].
Some theoretical work had been done to understand the consequences of stochastic fluctuations of molecule concentration on molecular interactions [
26]. Take two freely diffusing reacting particles with a center to center distance
r as example,
will be the probability density of two particles to remain unbound and separated with a distance
r at time
t and reaction dimensionality
[
5,
27]. Modeling the reaction of the two particles gives the reaction-diffusion master equation, which is given below:
The generalized form of the reaction-diffusion master equation can be derived when the space is divided into various compartments, the molecule
denote a molecule species I in subvolumn
. When describing reactions such as
, where
is the stoichiometry vector,
is the probability that the reaction will occur in compartment
. The generalized form of reaction-diffusion master equation can thus be expressed as [
28]:
where
t is the reaction time,
x is the coordinate position of the molecule, the factor
M governs reaction and
D diffusion. The form of
M and
D can be given below for reactant
X of species
i in compartment
j:
where
is the chemical reaction propensity function, which is also a measure of reaction rate.
Stochastic simulation algorithms (SSA) are widely used to implement the chemical master Equation [
29], it is a procedure for numerically simulating the time evolution of a small scale, well-stirred, diffusion influenced reacting system in which the reaction rate is subjected to stochastic variation [
30]. The most well-known and widely applied of which is developed by Gillespia and thus termed Gillespia’s Algorithm [
31], the general principle of which is described below:
For chemical reactions of the form:
or
, Gillespia’s Algorithm will first compute the propensity functions of each reaction with each reaction’s corresponding probability distribution, and then determine the time when the next chemical reaction takes place, finally, it provides the choice of which reaction occurs at which time step and the molecular number of each species in reaction will be updated accordingly, reporting the state of each reactant and the value of reaction kinetics as the reaction progress [
32].
However, despite recent major improvements in the efficiency of the SSA, its drawback remains the inaccuracy of simulation results, which is a common occurrence in experimental data independent computational simulations of reaction, this is a problem due to the unpredictability of chemical or biological systems. The great amount of computational time that were often required to simulate a desired system is also one of the disadvantage of SSA [
32]. People often resort to dynamic simulations that take into account the position and structural traits of every molecule in the reaction system, and it generate linear step response curves or reaction trajectories that are not necessary when the goal is only to obtain the thermodynamic parameters of the reaction system [
33,
34], but require much more computation power and need much more time [
33,
35].
Experimental results are valuable because they help to guide the process of reaction model building and simplify the complicated simulation of molecules in reaction. Establishing reaction models based on experimental data is beginning to gain more popularity in the field of chemical or biological computation, because it more accurately model the biological or chemical reaction process than hypothesized models alone.
Here we proposed an ANN-dependent Kinetic Parameter Extraction Method (AKPE) as an aid to experimental data to acquire kinetic rate constants from reactions. Features involved in AKPE include:
Presentation of reaction models describing chemical or biological reactions based on known experimental data.
Acquisition of reaction parameters based on reaction models will substantially simplify the simulation procedure and shorten the amount computational time.
The stochasticity of molecular reactions can be reflected on time-independent or time-dependent association rate constants, here, microscopic parameters such as center-to-center distance between two reacting molecules were also encoded in the kinetic rate constants, and AKPE will decode and extract these parameters from experimental data.
3. Materials and Methods
3.1. Introduction of the Computation Mechanism of ANN-Dependent Kinetic Parameter Extraction (AKPE)
In this section, a brief introduction of the method will be given. Firstly, a mathematical model of differential equations is presented describing chemical or biological reactions. The coefficients of the differential equations correspond to the kinetic parameters of the reactions. secondly, a neural network will be used to approximate the aforementioned differential equations according to experimental data, the coefficients in differential equations were incorporated into neural networks. Finally, errors from the differential equations and from the fitting of the experimental data to neural networks were combined to form a final error that will be evaluated by a global search algorithm that will yield optimum parameters for subsequent reaction analysis. The outline of the method is presented on a step-by-step basis below.
3.2. Mathematical Modeling of Reactions
In this section, mathematical models of reactions were presented as differential equations. For systems where the copy numbers of reactants are large enough that stochastic fluctuations in solvents have little effect on the macroscopic rate constants, Deterministic Reaction Rate Equations (RRE) are sufficient to describe the reaction system. Depending on whether the reactants are partitioned into different spatial compartments, we categorize RREs into ordinary differential equations and partial differential equations.
3.2.1. Differential Equation Describing Single Reactant Reaction
Ordinary deterministic chemical rate equations when describing reactions with a single reactant, such as material decay
A ->
A, can be expressed in the form:
where
x is the reaction time and
D designate a certain definition domain of
x,
A(
x) signifies the concentration of reactant
A,
denotes the nth order derivative of
A,
n is the reaction order. The boundary conditions were defined such that
is any predefined points in domain
D:
3.2.2. Differential Equations Involving Multiple Reactants
When the chemical reaction system involves multiple reacting species, such as in the nth order reaction A+B ->C, and the aim is to monitor the status of multiple reactants. We shall use a set of differential equations:
with initial conditions in the form of:
where
A,
B, and
C are reacting species,
n is the reaction order.
3.2.3. Partial Differential Equations Describing Compartmentalized Reactions
Partial differential equation was also employed when compartmentation or spatial inhomogeneity is present in the system with regard to a critical reactant [
36]. Reactions or interactions between species only occur when reacting species diffuse across reacting boundaries [
19]. The form of the partial differential equations that we used in AKPE for a single reactant
y are as follows [
22]:
where
and
are time measurements in two distinctive reaction environments;
y is the reactant concentration with regards to the sub-reactions.
Boundary conditions take the form of:
3.2.4. Time-Dependent Reaction Rate Constant
In classical chemical kinetics, it assumes a ample supply of reacting pairs in close proximity, kinetic parameters from reaction-derived differential equations often possess time-independent association rate constants. However, due to a small number of reacting molecules and low reaction rates, diffusion cannot effectively provide a continuous supply of closely reacting pairs [
27]. Anomaly in kinetic rate constants will shift the rate parameters away from a perfectly time-independent one. Therefore, a completely time-independent reaction rate constant would be misleading as microscopic fluctuations can propagate to cause a macroscopic effect on reaction rates. As a consequence, stochasticity turns out to be an important factor in reaction systems described above [
37]. To more accurately capture the variability in reactions, time-dependent rate coefficients were used, the complicated derivation of time-dependent association rate constants are not the main focus of our study here, for thorough description, please refer to
Supplementary Information (SI) Section S1 and references [
7,
38,
39].
3.3. Constructing Neural Networks Modeling Reaction Differential Equations for Experimental Data Approximation
The general form of the reaction-derived differential equations are given in the previous Section [
23]. Here, we utilize feedforward multi-layer neural networks for modeling any given set of experimental data. Any given function
y and its n-th order derivative
, can be mapped continuously by an neural network in the following way:
where
x is the independent variable,
,
,
are bounded real-valued adaptive parameters from neural networks,
k is the total number of hidden nodes and
is the activation function of neural networks, here it takes the form of a sigmoid function.
Any given experimental data can be formulated with a linear combination of neural nodes, with an arbitrary number of nodes. The coefficients of the reaction-modeled differential equations and inherent parameters of neural networks can be trained by a search algorithm both for the fitting of the experimental data and for solving differential equations. The general architecture of the neural network is illustrated in
Figure 1, the three layers of neural networks are demonstrated in the figure, The structure of the network can be represented as [
m,
,...
,
n] where
is the number of nodes in the kth hidden layer, we started with one hidden layer and 10 hidden nodes in the first setup, and then gradually increasing the number of hidden layers for better performance. The network consists of an input layer of m nodes representing m experimental data points. The second layer is the hidden layer, where input nodes are transformed by sigmoidal activation function
and adjustable parameters
and
for hidden layer
i, the first order and second order derivatives of each nodes were also computed in this layer, the specifics of transformations are slightly different depending on the requirements of each reaction systems; however, each reaction species in the reaction system can be represented by a certain transformation of the neural network nodes, which can subsequently be fitted to experimental data. The third layer is the outputs from the neural network, which are the coefficients of the reaction-derived differential equations, the parameters of neural networks and the search method, and errors from all the above-mentioned modules in AKPE.
3.4. Fitness Evaluation and Training in AKPE
The fitness function contains two errors arising from the fitting of experimental data and from solving reaction-derived differential equations. The linear combination of errors from the neural network’s modeling of reaction differential equation and from fitting of experimental data forms an unsupervised error function. The error to be minimized can be defined as an weighted error with minimization priority in either sub-errors according to specific conditions in application. The total error is given as:
where
is the weight in the range of 0–1. The weight factor
puts emphasis on one error over the other, and offer more freedom in user coordination, this leads to higher accuracy and better robustness.
From total error
e,
is the error that originate from differential equations and can be expressed as:
where function
G(.) takes the from of Equation (
1) for single reactant reaction or it can take the form of Equations (2) and (3) for multi-reactants reactions or compartmentalized reactions, respectively,
is the concentration of reactant,
denotes the nth order derivative of
,
m is the total number of points in the definition domain of independent variable
x.
is an error arising from neural network data fitting of experimentally measured values with an expression in the form of:
where
is single output neural network with parameter
p and
m input units fed with variable vector
x.
contains no adjustable parameters and is the experimental data vector in time point
i. The weight vector
p can be adjusted and trained according to the optimization algorithm used.
In the training process, the parameters to be adjusted in the minimization algorithms are the parameters from the differential equations, plus weights from the neural networks. The training data were normalized to be in the binary form to hasten the network training. About 70% of the data were used in the training and simulation steps and the rest of the data were used for validation of the network model. We initialize all the weights by assigning random values in the range , use the training data to learn the weights and record the value of the Root Mean Square Error(RMSE) in function of the times/cycles until the RMSE drops to a value below .
3.5. PSO as an Optimizer
We employed a global search method to find the set of optimum parameters that best fit the experimental data and solve the reaction-derived differential equations. The method we used was particle swarm optimization (PSO) [
17,
18]. PSO is a well-developed, computer optimization method aimed to find the global optimum by iterative refining a predefined measure of quality. It was introduced by Eberhart and Kennedy in 1995 as a stochastic global optimization technique inspired from the behavior of particles or individuals in a swarm. This algorithm integrated the information gathered by the parallel searching of individuals to enhance the total efficiency of the swarm. The main advantages of PSO algorithm are that it is simple in concepts, easy in implementation, stable in convergence and in global optimum finding, and efficient computationally compared to other heuristic optimization techniques [
40].
In PSO algorithm, each individual of the swarm represents a possible solution in the non-linear solution space. The problem of finding a global optimum is thus turned into finding the optimum individual. Each individual searches a problem space and its measure of quality is its position and velocity. PSO algorithm initializes with each particle placed randomly within an exploration space, and then each individual searches the space with its position and velocity updated iteratively according to previous local best
and global best
. The updating scheme of PSO is as follows:
where
i is the number of the particle,
j is the number of iteration,
is the inertia weight,
and
were defined by the authors as local and global acceleration constant,
and
are random vectors in the range of 0 and 1,
X and
V are positional and velocity vectors.
3.6. Pseudo Code and General Architecture of AKPE
The following is a sample pseudo-code of AKPE, annotations were provided in every line of the code after the percentage sign, function
is as defined in
SI S1,
is the reaction-derived differential equations,
has the expression
,
D,
MI and
denote reaction component dimer, intermediate state and starting monomer concentration, respectively. Executing AKPE module in Algorithm 1 and PSO module in Algorithm 2 will produce optimal values of all kinetic rate constants and intermediate state dynamics (concentration vs. reaction time curve) together with errors from every module, line 13, 17 and line 18, function D_to_MI and MI_to_D in in Algorithm 1 were inferred from reaction models, thus their expressions depend on the specific chemical or biological reaction under study.
Algorithm 1 AKPE |
Input: Experimental data , n is length of experimental data; Hypothesized reaction-derived differential equations, ; The range of kinetic rate constants, , ; Maximum number of function evaluations, ; Fitness weight, ; maximum neural network layers, k; Time points, t; Output: Kinetic rate constants, , y is the total number of kinetic rate constants to be evaluated, including , , , and reaction orders, , , ; Least square error from neural network, ; Intermediate state concentration at each time points, ; for each do initialize neural network nodes in each layer; initialize neural network weights , and offset ; initialize kinetic rate constants as X; end for functionError() for each do for each do ; %layers of neural network %value of D (Dimer) %dy/dx = y(1 − y) dDdt is the first order derivative of D ; %d2Ddt2 is the second order derivative of D ; % MI is the intermediate state for all do ; end for; ; %Back Calculation of D using MI according to specific reaction model for validation purpose ; %Error from Differential equation solution according to Equation ( 16) ; %Fitting module error, the general expression of D is ; %record D ; %record MI end for; end for; return error end function functionD_to_MI() %Expression of function D_to_MI in line 13 ; end function functionMI_to_D() %Expression of function MI_to_D in line 17 end function functiona(); %Expression of ; end function; functionsigmoid(x); ; %Expression of sigmoid function end function; function(x); end function;
|
Algorithm 2 PSO Module |
Input: Error function, ; Number of neural network nodes, ; Maximum number of iteration, ; Swarm size, ; Function relating swarm size to neural network nodes, ; Output: Optimal kinetic rate constants ; Optimal weights , , ; Errors; Initialize all PSO parameters %including N, w, , , etc; for each do ; ; end for whiledo PSO evaluation to minimize ; ; end while
|
The main steps of our application of neural networks and PSO in data fitting and parameter optimization can be summarized as follows and in
Figure 2.
- step 1:
Feedforward neural networks construction for experimental data fitting.
- step 2:
Constructing differential equations according to the reaction systems.
- step 3:
Fitness function construction by combining errors from data fitting and solving differential equations.
- step 4:
PSO optimization of kinetic parameters.
The novelty and contributions of AKPE method on biological or chemical reaction systems includes the following points:
Presentation of a method that integrate differential equation reaction modeling and neural networks for accurate reaction progress prediction.
Presentation of a computational method that predict the concentration dynamics and the reaction coefficients associated with the intermediate state which can not be measured experimentally due to limitations on experimental means.
Exploiting existing experimental data to aid de novo computational prediction of reaction coefficients and concentration dynamics of reactants and lowly-populated states (intermediate states).
Provided a simple and efficient system for fast reaction prediction, as opposed to having to describe every minute details of the reaction system in other simulation methods.
6. Conclusions
Simulation and prediction of chemical reactions can be an indispensable tool for understanding the mechanism of chemical or biological reactions [
37,
46,
47]. Many reactions call for the participation of an intermediate state which prompts the reactions to proceed in an energetically favorable way. Unfortunately, for some reactions, experimental measurement failed to capture the intermediate state dynamics because of its low abundance; however, computational simulations can show its excellency in aiding experimental data at measuring intermediate state dynamics.
In this paper, we have presented a new method AKPE for predicting kinetic rate constants and reacting species concentration. AKPE employs neural networks to approximate experimental data and solve reaction model-derived differential equations, it also implements swarm intelligence algorithm such as PSO to optimize parameters and coefficients. AKPE couples neural network data fitting and equation-based swarm intelligence algorithm error minimization for kinetic rates and lowly-populated state dynamics prediction. As an extension of experimental data fitting, it is aimed to extract more information from experimental data and defined chemical reaction models, It obtains more information than a simple two-state fitting, and outputs previously undetected kinetic rates and undiscovered transitional state concentrations.
The applicability of AKPE at predicting chemical reaction dynamics was illustrated using two case studies, which yielded corresponding kinetic rate constants and the concentration curve of the intermediate state as a function of time. The computational accuracy of AKPE was quantified in standard variation of kinetic parameters, and it is well within tolerable ranges. As one of the most illustrious points of our work, we have also demonstrated that AKPE properly describes intermediate state dynamics in reaction systems where experimental or other methods failed. AKPE produced the intermediate state concentration curve with respect to time, which describes in detail how the concentration of the intermediate changes as the reaction progress, providing us a peek into the mechanism of reactions.
Another point worth noticing is that even though we used a primarily deterministic model, we also incorporated stochasticity into our system since a particular reacting species(intermediate state) exist only in a very low concentration. The stochasticity was incorporated into a time dependent reaction rate constant . This stochasticity helped us in producing the correct intermediate state dynamics.
To evaluate the validity of AKPE, we used SSA simulations, which produced similar results as AKPE, with standard errors (MSE, RMSE) within acceptable ranges, demonstrating the validity of AKPE. Moreover, by mathematical modeling, we arrive at a equation linking experimentally measured kinetic rate constants and all the rate constants used in our model, thereby, making up another method for validating AKPE, the result showed that experimental association rate constant also confirmed the validity of the predictions produced by AKPE.
The method AKPE proposed in this article is a robust predictor of lowly-populated reacting species, it also greatly improves the accuracy of kinetic rate prediction. It helps to identify the intermediate state that was previously invisible to experimental detections.
The novelty of AKPE lies in the fact that it is a roust and generally accurate predictor of the reaction dynamics associated with scantly-populated reacting species, which were unable to detect by experimental methods alone. AKPE utilized existing experimental data on chemical or biological reaction systems, greatly improved the accuracy of kinetic rate prediction compared with de novo computational methods. In the method architecture, AKPE combines the flexibility of neural networks, the generalizability of differential equations in modeling reaction systems and the global optimizing properties of PSO to obtain an accurate simulation of reaction systems.