Next Article in Journal
An Innovative Method Based on Wavelet Analysis for Chipless RFID Tag Detection
Previous Article in Journal
Light Field Spatial Super-Resolution via View Interaction and the Fusing of Hierarchical Features
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Hybrid Preference Interaction Mechanism for Multi-Satellite Imaging Dynamic Mission Planning

by
Xueying Yang
,
Min Hu
*,
Gang Huang
and
Yijun Wang
Department of Aerospace Science and Technology, Space Engineering University, Beijing 101416, China
*
Author to whom correspondence should be addressed.
Electronics 2024, 13(12), 2374; https://doi.org/10.3390/electronics13122374
Submission received: 7 April 2024 / Revised: 12 June 2024 / Accepted: 16 June 2024 / Published: 17 June 2024
(This article belongs to the Section Microwave and Wireless Communications)

Abstract

:
The existing multi-satellite dynamic mission planning system hardly satisfies the requirements of fast response time and high mission benefit in highly dynamic situations. In the meantime, a reasonable decision-maker preference mechanism is an additional challenge for multi-satellite imaging dynamic mission planning based on user preferences (MSDMPUP). Therefore, this study proposes the hybrid preference interaction mechanism and knowledge transfer strategy for the multi-objective evolutionary algorithm (HPIM–KTSMOEA). Firstly, an MSDMPUP model based on a task rolling window is constructed to achieve timely updating of the target task importance degree through the simultaneous application of periodic triggering and event triggering methods. Secondly, the hybrid preference interaction mechanism is constructed to plan according to the satellite controller’s preference-based commands in different phases of the optimal search of the mission planning scheme to effectively respond to the dynamic changes in the environment. Finally, a knowledge transfer strategy for the multi-objective evolutionary algorithm is proposed to accelerate population convergence in new environments based on knowledge transfer according to environmental variability. Simulation experiments verify the effectiveness and stability of the method in processing MSDMPUP. This study found that the HPIM–KTSMOEA algorithm has high task benefit, short response time, and high task completion when processing MSDMPUP.

1. Introduction

Multi-satellite dynamic mission planning (MSDMP) is a large-scale combinatorial optimization problem that aims to quickly arrange multiple new tasks by adjusting the mission planning scheme in time through local/global replanning. The MSDMP system takes into account the uncertainties that may be encountered in the process of satellite observation of the Earth and demonstrates excellent performance in both routine and emergency safeguard mission scenarios [1,2]. However, with the continuous expansion of application scenarios, satellites are faced with a large number of dynamic factors interfering with the process of in-orbit operation, which puts forward higher requirements for the timely adjustment of planning schemes and the timeliness of planning. The existing MSDMP system is restricted by the limitations of constraints, and it is difficult to meet the requirements of intelligent applications under highly dynamic situations [3]; moreover, the satellite controllers are directly involved in the MSDMP system for preference decision-making for complex dynamic problems, which can make up for the system deficiencies, and make the system formulate a scheme that is more in line with the current environmental requirements [4]. Therefore, the investigation of multi-satellite dynamic mission planning based on user preference (MSDMPUP) can combine the user’s dynamically changing requirements with the mission planning system, making the system quickly respond to dynamic changes in the environment, which is the key to constructing a high-quality mission planning scheme; this approach represents the cutting-edge technology of current satellite mission planning research.
In recent years, researchers have conducted in-depth investigations on MSDMPUP problems and proposed many excellent models and algorithms. These approaches are classified into three main categories, namely, MSDMPUP planning models, planning strategies, and solution algorithms.
For MSDMPUP planning model research, common models include the integer planning model [5], the constraint satisfaction model [6], and the Markov decision model [7]. Qu et al. [8] have developed an intelligent scheduling framework that describes satellite mission planning as a mixed-integer planning model, solves it using an intelligent algorithm based on imitation learning, and proposes iterative views based on user preferences to improve the performance of the training strategy. Yu et al. [9] combine the constraints of visibility, optical axis (OA) adjustment capability, cooperation duration, and immunity to establish a user preference constraint satisfaction scheduling model based on user preference, which effectively improves the total observational benefit of the multi-satellite imaging dynamic mission planning system. Li et al. [10] introduce a Markov decision model based on user preferences in the task decision phase to effectively reduce the consumption time of the MSDMPUP system.
For the MSDMPUP planning strategy investigation, the main planning strategies of the MSDMPUP planning strategy investigation include proactive planning [11], reactive planning [12], and proactive reactive planning [13]. Gu et al. [14] propose an MSDMPUP model considering cloud prediction that relies on predictive recurrent neural networks and up-to-date satellite cloud images for continuous cloud forecasting and user preference selection through proactive planning. Wang et al. [15] proposed a reactive planning method considering user-preference- and time-driven strategies for the high dynamics and uncertainty of MSDMPUP, taking into account the two key objectives of observation benefit and robustness. Dai et al. [16] designed the MSDMPUP scheme through four phases of task preprocessing, task resource matching, task conflict assessment, and task resources, and proposed a user preference approach with task-slacking proactive–reactive planning to maximize the guaranteed task completion rate.
For the investigation of the MSDMPUP solution algorithm, the main solution algorithms include heuristic algorithms [17], deterministic algorithms [18], intelligent optimization algorithms [19], and reinforcement learning [20]. Wang et al. [21] provide an in-depth analysis of user preference selection based on environmental changes and propose a two-way rolling level heuristic to effectively improve the scheduling efficiency of MSDMPUP systems. Wu et al. [22] proposed a meta-heuristic approach for the large-scale MSDMPUP problem using a divide-and-conquer framework to improve the task execution benefit in different simulation environments with better allocation efficiency. Wang et al. [23] proposed a multi-objective differential evolutionary algorithm based on spatial partitioning–user preference policy to solve the MSDMPUP problem by simultaneously considering the three objectives of optimizing the overall task benefit, antenna load balancing, and task completion timelines. Lin et al. [24] propose a multi-satellite deep learning architecture for fusing user preferences, in which each satellite can train an efficient model based on deep learning algorithms, which significantly improves the timeliness of satellite mission decision-making. Song et al. [25] propose a deep reinforcement learning-based solution for the large-scale MSDMPUP problem, which utilizes deep neural networks for each action to estimate to prioritize the task scheduling.
Although researchers have achieved better planning results for the MSDMPUP problem through advanced planning models, planning strategies, and solution algorithms, there is still a difference between the output scheme of the mission planning system and the real user requirements due to a variety of dynamic factors, such as task cancellation, changes in task attributes, and changes in the allocated satellite resources in complex environments [26]. Presently, there are still many challenges in multi-satellite dynamic mission planning; the specific difficulties are described as follows:
  • Satellites performing on-orbit missions in space often face new emergencies, changes in user demands, and other dynamic events. Moreover, the existing target importance determination method has a limitation in that satellite resources cannot be deployed in time according to the dynamic changes in mission demand. Furthermore, it is difficult to complete the optimal matching of resource capacity and mission demand in the generated planning scheme. Therefore, the construction of a mission planning model that reasonably formulates the target importance is a major challenge for MSDMPUP.
  • In the existing research, users are mostly involved in the initial stage of the task and the final scheme selection stage, while dynamic tasks coexist with static tasks, and the task requirements present preference, complexity, and conflict, and there is the difficult problem of poor fitness of the output planning scheme. How to introduce the human capacity of planning and decision-making in the MSDMPUP optimized search process, design a reasonable and reliable user preference mechanism, and construct a high-quality task planning scheme is a new challenge for applying MSDMPUP in a complex environment.
  • The introduction of user preference requirements results in more complex real-world environmental constraints in MSDMP. Moreover, the problem space is more massive, resulting in difficulty in responding quickly to the environmental changes in the multi-satellite dynamic mission planning problem and low optimal efficiency. In addition, the design of a simple output planning scheme is a difficult challenge. Therefore, how to construct MSDMPUP solution algorithms adapted to a variety of changing modes, and effectively trade-off between the solution performance and the computational efficiency, is another challenge to applying MSDMPUP in complex environments.
To solve the above difficult problems, this work investigates MSDMPUP with imaging satellites. The main contributions of this work are as follows:
  • To address the real-time requirements of MSDMPUP due to dynamic changes in user requirements in complex environments, an MSDMPUP model based on a task rolling window is proposed. The model takes into account the three optimization objectives of task benefit, task completion, and task timeliness. Moreover, the model responds to environmental changes quickly by applying both periodic triggering and event triggering to achieve timely updating of the target task importance degree, which allows the task set to be scheduled in the optimal order based on the size of the target task importance degree, thus ensuring that each task can be completed within the optimal scheduling time.
  • To address the problem of poor adaptability of the MSDMPUP system output planning scheme, the hybrid preference interaction mechanism is proposed. In this mechanism, the pre-planning layer constructs a mapping strategy to map the MSDMPUP allocation sequence to the evaluation matrix of each sub-objective, the preference interaction layer generates a new generation of populations to participate in the algorithm optimization search according to the instructions sent by the satellite controller, and the interaction output layer constructs an inverse mapping strategy to reflect the evaluation value of the sub-objective to output the newest mission planning scheme selected by the satellite controller, which effectively deals with the dynamic changes in the environment.
  • To address the problem that the introduction of user preferences makes difficult for MSDMPUP, i.e., to cope with irregular environmental changes and poor solution efficiency, a knowledge transfer strategy for the multi-objective evolutionary algorithm is proposed. The knowledge transfer strategy constructs an MSDMPUP knowledge pool for knowledge storage, adaptively selects a knowledge transfer strategy according to the differences between the current environment and the historical environment under the new environment, retrieves similar knowledge in the historical environment to promote rapid convergence of the algorithm, adapts to the changes in mission requirements to obtain a high-quality multi-satellite imaging mission planning scheme, and enhances the universality of the solving algorithm.
To test the performance of the proposed method, we set up four sets of simulation experiments to verify the feasibility of the HPIM–KTSMOEA algorithm and compared the performance with other algorithms. The experimental results show that the method proposed in this work is superior and scalable, and provides a good reference for the multi-satellite dynamic mission planning problem in complex environments.
The rest of this work is organized as follows: Section 2 presents the basic MSDMP scenario and the motivation for investigating MSDMPUP. Section 3 introduces the MSDMPUP model based on a task rolling window and describes its objective function and constraints. Section 4 details the hybrid preference interaction mechanism of the HPIM–KTSMOEA algorithm and the specifics of the knowledge transfer strategy for the multi-objective evolutionary algorithm. Section 5 sets up simulation experiments to analyze the effectiveness and stability of the HPIM–KTSMOEA algorithm. Section 6 provides our conclusions and future perspectives.

2. Related Work

This section first analyses the fundamental MSDMP scenarios and then, describes the motivation for investigating MSDMPUP.

2.1. MSDMP Fundamental Scenario

MSDMP fundamental scenario is shown in Figure 1. Assuming that a set of satellites S a t = S a t 1 , , S a t 6 need to execute a set of target tasks T a s k = T 1 , , T 7 within a given task planning time S c h e d u l e T i m e = T t | t t b e g i n , t e n d , the visible time window of the target tasks is [ s t , e t ] , and the satellite executable orbit is O r b = O r b i t 1 , O r b i t 2 , the system generates an initial task planning scheme at the moment of T 1 as shown in Figure 1a, and the task planning scheme on the O r b i t 1 is ( S a t 1 , T 1 ) , ( S a t 2 , T 2 ) , ( S a t 3 , T 3 , T 4 ) , and that on the O r b i t 2 is ( S a t 4 , T 5 ) , ( S a t 5 , T 2 ) , ( S a t 6 , T 6 , T 7 ) .
From then, the MSDMP system is executed according to the initial mission planning scheme, which corresponds to the satellite allocation sequence { O r b i t 1 | o c c u p y k 1 j 1 , o 1 , o c c u p y k j 1 , o 1 , o c c u p y k + 1 j 1 , o 1 } , { O r b i t 2 | o c c u p y k 1 j 2 , o 2 , o c c u p y k j 2 , o 2 , o c c u p y k + 1 j 2 , o 2 } . The arrival of the new target task T 8 , T 9 , T 10 is monitored at time T 2 . The process of task assignment for the new target task using the local replanning approach is shown in Figure 1b. By comparing the start time of the visible time window of the new target task with the visible time window of each of the original target tasks one by one, the original search for the idle time window that exists in the satellite observation sequence is performed, and the new target task is inserted into the free time window, which is marked with a red square in Figure 1b.
This has completed the dynamic allocation of the multi-satellite imaging dynamic mission planning, and the final mission planning scheme obtained at T 3 time is shown in Figure 1c, with the mission planning scheme on Orbit 1 as ( S a t 1 , T 1 , T 9 ) , ( S a t 2 , T 2 ) , ( S a t 3 , T 3 , T 4 ) , and the mission planning scheme on Orbit 2 as ( S a t 4 , T 8 , T 5 ) , ( S a t 5 , T 2 ) , ( S a t 6 , T 6 , T 7 , T 10 ) [27,28].

2.2. Motivation

The existing MSDMPUP research user mostly adopts the proactive planning strategy for preference selection, which means that the user selects the target region or reference point based on the approximate range without preference P a r e t o , which helps the system search the high-performance region of the target space faster in order to retrieve a high-quality task planning solution [29,30]. Users are given preference information in the pre-task planning stage and do not make further changes; however, satellites in real applications are often subjected to unknown complex factors due to the complexity and variability in the environment (e.g., satellites are subjected to interferences caused by solar activity, electromagnetic storms, and on-orbit service failures; moreover, the number of on-orbit satellites in space has grown significantly, increasing the rate of satellite collisions). These factors result in dynamic changes in the user preference requirements during the entire process of search optimization for task planning, which is further complicated due to the output planning scheme not having high adaptability [31,32]. There is an urgent need to improve the participation method of user preference selection and to develop MSDMP schemes that are more relevant to real user requirements.
Therefore, this work proposes the hybrid preference interaction mechanism and knowledge transfer strategy for multi-objective evolutionary algorithms. In the pre-mission planning stage, the MSDMPUP model is constructed based on a task rolling window, and the user preference requirements are input through dynamic updating of the target task importance degree. The task planning optimal search period constructs the hybrid preference interaction mechanism. Then, the user preference demand feature is used via the preference command to generate a new set of population data in order to optimize the search results of the algorithm. In addition, to address the problem of the user preference requirements dynamically changing, which makes it difficult for the multi-satellite dynamic mission planning system to respond quickly to environmental changes, a knowledge transfer strategy for the multi-objective evolutionary algorithm is constructed to effectively trade-off the solution performance and computational efficiency.

3. Problem Description

This section first describes the MSDMPUP model based on a task rolling window and then further describes the objective function and constraints of the model.

3.1. MSDMPUP Model Based on a Task Rolling Window

In this section, the MSDMPUP model based on a task rolling window (MSDMPUP–TRW) is constructed. The model is oriented toward the synthesis of satellite imaging requirements for complex environments in terms of six indicators, namely, safeguard type S u p , application area a p p , task requirements r e q , resource characteristics r e s , urgency degree u r g , and environmental perturbations d i s , to generate the target task importance degree required to produce the initial task set for determining the observation order of the target task.
The target task importance degree is dynamically updated by using a task rolling window to cope with changes in the complex environment and a knowledge transfer strategy for the multi-objective evolutionary algorithm is used to generate the initial task planning scheme. The specific method of the MSDMPUP–TRW model is as follows.
Firstly, a unified mathematical description of the indicators affecting the target task importance degree is provided as follows:
w = f ( g = 1 6 I n d e x g ) = f ( S u r , a p p , r e q , r e s , u r g , d i s ) ,
where I n d e x g denotes the gth target task indicator, and the indicator evaluation matrix for any target task is as follows:
X = ( x i j ) n × 6 = x 11 x 13 x n 1 x n 6 , x i j ( 1 i n , 1 j 6 ) ,
where x i j ( 1 i n , 1 j 6 ) is the quantitative value of the gth indicator for the ith target observation task, and a vector of weights w is obtained for the quantitative values of the targets x i j as follows:
w i = ( j = 1 n x i j / 10 ) 1 N i = 1 N ( j = 1 n x i j / 10 ) 1 N , W = w 1 , w 2 , w 3 , , w n T ,
According to the weight vector w , the initial target task importance degree is further calculated as follows:
S i g = g = 1 6 W I n d e x g ,
Secondly, the task rolling window method is used to dynamically update the task importance degree during the task planning cycle, while using periodic triggers and emergency event triggers. The task rolling window method is used to dynamically update the task importance degree during the task planning cycle, and both the periodic trigger method and the event trigger method are used. Among them, the periodic trigger is a trigger mechanism based on time intervals, which recalculates the target task importance degree according to a preset period without being affected by external tasks. Event triggering is a triggering mechanism based on the occurrence of a specific event. When the system monitors a specific event (such as a change in satellite resources, a change in the demand for the target task, and other emergency events), the importance of each target task is recalculated, and to avoid repeated triggering and interference with the stability of the system, an event triggering threshold is set to achieve a dynamic balance between the number of triggering times and the task importance degree.
The initial moment of the task is executed using the periodic triggering method, which allows the dynamic change in the importance of each target task to be independent of the external task and to be rolled over only once every fixed period of time. During cyclic triggering execution, if an emergency event occurs when the value of the event triggering function is greater than the threshold value set for event triggering, it triggers the execution of the emergency event triggering method and the cycle is completed until the end of the task planning. The event trigger threshold is set to T h r e s h o l d E t r i g g e r , and the event trigger function is as follows:
E t r i g g e r = i = 1 N T a s k ( s i g i + t e n d t b e g i n / t i d e a l i t c u r r e n t i ) .
where N T a s k is the number of target tasks, s i g i is the target task importance of the ith task, t b e g i n , t e n d denotes the start and end time of the task planning cycle, and t i d e a l i and t c u r r e n t i denote the ideal planning time and the current moment of the first task, respectively. t e n d t b e g i n / t i d e a l i t c u r r e n t i denotes the timeliness parameter of the i th target task, the smaller the difference between the time of the moment t c u r r e n t i and the ideal planning time t i d e a l i , the larger the timeliness parameter, and the larger the value of the corresponding event trigger function. Target task importance degree s i g i combined with the timeliness parameter quantifies the urgency and importance of target task scheduling.
Figure 2 introduces the execution process of the task scrolling window. During the task planning time t b e g i n , t e n d , a group of target tasks T 1 , T 2 , , T n is dynamically updated with the target task importance, where the event trigger threshold is set to T h r e s h o l d E t r i g g e r = 2 ; the blue and yellow squares represent the regular and emergency tasks, respectively, and the dotted lines of different color are used to split the different task scrolling windows, which overlap. The initial moment t b e g i n is followed by the preparation time Δ T , which is triggered by the periodicity to initiate the mth task roll; the m + 1 th task roll is triggered by the contingency tasks T 3 , T 4 at time t 1 , the m + 2 th task roll is triggered by the contingency tasks T 5 , T 7 at time t 2 ; and the m + 3 th task roll is initiated by a periodic trigger at the moment t 3 ,which loops until the end of the task planning time.

3.2. Establish Model Objective Functions and Constraints

3.2.1. Model Objective Functions and Decision Variable

The objective function construction in this study mainly considers the task benefit objective function, the task timeliness objective function, and the task completion objective function.
  • Sub-objective function 1—maximize task benefit objective function:
max f 1 ( x i , j , o ) = i = 1 N T a s k j = 1 N S a t o = 1 N O r b x i , j , o   i = 1 N T a s k w i , i T a s k , j S a t , o O r b , w i W ,
where Equation (6) represents the maximum value of the task benefit of the satellite performing the target task in all orbits, and the satellite should fulfill as many tasks of high target importance as possible. T a s k denotes the set of target tasks, S a t denotes the set of satellites, O r b denotes the set of satellite execution orbits, W denotes the target importance degree matrix of target tasks, x i , j , o denotes the decision factor, w i denotes the target task importance degree of target tasks i , N T a s k denotes the number of target tasks, N S a t denotes the number of satellites, and N O r b denotes the number of executable orbital circles. Among them, the mathematical expression of the decision factor x i , j , o is as follows:
x i , j , o = 0 , T i   i s   n o t   e x e c u t e d   b y   t h e   S a t j   i n   o r b i t o 1 , T i   i s   e x e c u t e d   b y   t h e   S a t j   i n   o r b i t o ,
where x i , j , o 0 , 1 , x i , j , o = 0 indicates that the i -th target mission is not carried out by the j -th satellite in o -th orbit, indicating that the i -th target mission is carried out by the j -th satellite in the o -th orbit.
2.
Sub-objective function 2—minimize task timeliness objective function:
min f 2 ( x i , j , o ) = i = 1 N T a s k j = 1 N S a t o = 1 N O r b x i , j , o   i = 1 N T a s k W i r e s p o n s e , i T a s k , j S a t , o O r b ,
where Equation (8) indicates the minimum response time of the satellites in all orbits to perform the target tasks, and the satellites should complete as many fast response tasks as possible. W i r e s p o n s e indicates the observed response time of the target task.
3.
Sub-objective function 3—maximize task completion objective function:
max f 3 ( x i , j , o ) = i = 1 N T a s k j = 1 N S a t o = 1 N O r b x i , j , o   N T a s k 100 % , i T a s k , j S a t , o O r b .
where Equation (9) represents the number of all target tasks performed for this mission planning as a percentage of the total number of tasks, and satellites should be assigned as many target tasks as possible to perform.

3.2.2. Model Constraints

This subsection describes the MSDMPUP constraints, specifically the target task allocation constraints, the target visibility constraints, the satellite performance constraints, and the target mission assignment constraints.
  • The target task allocation constraints are as follows: each satellite can observe multiple target tasks in sequence and each target task can be observed by multiple satellites:
S A j = 1 i = 1 n [ 1 S A i j x i j ] , i = 1 N T a s k j = 1 N S a t x i j 1 ,
where S A i j denotes the probability of successful observation of the j th satellite performing the i th target task. S A i j denotes the probability that the j th satellite successfully observes at least one target mission, and x i j denotes the decision variable for the j th satellite to perform the i th target mission;
2.
The target visibility constraint is as follows: the satellite is visible to the target mission within the elevated node pitch angle interval:
ϑ v e r t i c a l arccos ( R R + h ) θ v i s i b l e ϑ v e r t i c a l + arccos ( R R + h ) ,
where ϑ v e r t i c a l represents the elevation angle of the intersection point when the imaging satellite is perpendicular to the target task, R represents the linear distance between the target task and the center of the earth, h represents the linear distance between the imaging satellite and the ground, and θ v i s i b l e represents the visible angle between the imaging satellite and the target task.
3.
The satellite performance constraints are as follows: these constraints are limited by the satellite’s own resources, side-swing energy, and execution time:
( i = 1 N T a s k j = 1 N S a t o = 1 N O r b x i , j , o O C o n o 1 ( W i , j , o e t W i , j , o s t ) + i = 1 N T a s k j = 1 N S a t o = 1 N O r b   x i , j , o S C o n i , j , o 1 ) E n e r g y s a t .
where O C o n o 1 denotes the resource consumption per unit time in orbit O , S C o n i , j , o 1 denotes the side-swing energy consumption of the satellite j performing the target task i in orbit O , W i , j , o e t , W i , j , o s t denotes the observation end and start time of the satellite j performing the target i mission in orbit O , and E n e r g y s a t denotes the satellite’s own energy.
4.
The target mission assignment constraints are as follows: these constraints mainly consist of the side-swing angle constraints, the resolution constraints, and the field-of-view angle constraints:
( θ a t T A j ) & ( r j S a t r min ) & θ a t T θ a t T 1 θ j .
where θ a t T denotes the mission swing angle of the observation target, A j denotes the maximum lateral swing angle of the imaging satellite, r j S a t indicates the remote sensor resolution of the imaging satellite j , r min denotes the minimum resolution required by the target task, and θ j denotes the individual field of view of the satellite remote sensor.

4. Design of HPIM–KTSMOEA Algorithm

In this section, an MSDMPUP model based on the hybrid preference interaction mechanism and knowledge transfer strategy for the multi-objective evolutionary algorithm (HPIM–KTSMOEA) is proposed. The framework of the HPIM–KTSMOEA algorithm is shown in Algorithm 1.
Algorithm 1. The framework of HPIM–KTSMOEA
Input :   o b j   ( Parameters   of   the   satellites   S n ,   Parameters   of   the   target   task   T a s k ,   Population   size   P n );
Output: MSDMPUP optimal scheme;
1   MSDMPUP   model   based   on   task   rolling   window   is   used   to   generate   the   initial   set   of   tasks   T a s k = T 1 , , T N t a s k ;
2 The MSDMPUP fitness function is constructed according to Equations (6)–(8) and the MSDMPUP constraints are constructed according to Equations (9)–(12).
3       max f ( x i , j , o ) = f 1 ( x i , j , o ) , f 2 ( x i , j , o ) , f 3 ( x i , j , o )
4   / *   Call   the   pre - planning   layer   to   obtain   the   initial   allocation   reference   knowledge   pool   P o o l a l l o c a t i o n   and   the   target   task   evaluation   knowledge   pool   E v a l u a t e G a i n ,   E v a l u a t e T i m e ,   E v a l u a t e C o m p l e t i o n */
P o o l a l l o c a t i o n ( S a t , T a s k , G a i n , C o m p l e t i o n , T i m e ) = N S G A II ( o b j ) .
   for gen = 1:n
P k = D e s c e n d S o r t ( f k ( x i , j , o ) ) ./* Generating subpopulation */
f k φ : P k ( S a t j , T a s k i ) E v a l u a t e i n d e x ( i , j ) ./*Mapping strategy */
   end
5   / *   Call   the   preference   interaction   layer   to   generate   the   user   preference   planning   schemes   S o l u p r e f e r e n c e */
    P n e w = u s e r P , P 1 , P 2 , P 3 /* Generate a new generation of population*/
S o l u p r e f e r e n c e = H P I M K T S M O E A _ o p e r a t e ( P n e w ) /* HPIM-KTSMOEA algorithm */
6   / * Call   the   Interactive   output   layer   to   obtain   the   MSDMPUP   scheme S o l u n e w for satellite management user */
f k φ 1 : E v a l u a t e i n d e x ( i , j ) P k ( S a t j , T a s k i ) ./*Inverse mapping strategy */
S o l u n e w = k = 1 3 P k ( S a t j , T a s k i ) ./* MSDMPUP scheme */
7 End

4.1. Hybrid Preference Interaction Mechanism

This study proposes the hybrid preference interaction mechanism (HPIM) based on the MSDMPUP model to address the mismatch of planning scenarios applied in real environments caused by the limitations of user preference demand input in existing mission planning methods. In various stages of the mission planning process of the solution optimization search, the planning method is based on the preference instructions of satellite controllers so that it can be applied to different mission scenarios and effectively respond to the dynamic changes in the environment. The hybrid preference interaction mechanism consists of three layers: the pre-planning layer, the preference interaction layer, and the interaction output layer.
In the pre-planning layer, the satellite controller updates the external comprehensive index that evaluates the importance index of the target task according to the actual needs, to intervene in the observation sequence of the target task. In the preference interaction layer, the satellite controller sends demand instructions to the system according to the actual application environment, and the user needs, to realize the intervention of the optimization search process on the user preference task planning scheme. The interaction output layer outputs multiple imaging task planning schemes generated by the system for satellite management and control personnel to select, and realize the intervention of the final scheme selection.
The introduction of the hybrid preference mechanism allows satellite controllers to specify the allocation of resources based on preference-based instructions, ensuring that the mission planning scheme matches the actual needs under different mission scenarios and improving the efficiency and success rate of mission execution. In addition, the hybrid preference interaction mechanism emphasizes the interaction and collaboration between satellite controllers and the mission planning system, making the mission planning scheme flexible and scalable.
In addition, the existing investigation methods usually use the integer discrete set of satellite and target mission assignment sequences as the candidate solution of the MSDMPUP problem. Moreover, invalid and duplicate values of the candidate solutions are obtained during the optimization search process, which causes the failure of the MSDMPUP scheme. In this section, the physical significance of the satellite and the target mission is considered, and the individual sub-objective fitness values of the satellite and the target task are used as a unified medium. Furthermore, the mapping strategy and inverse mapping strategy are set up in the hybrid preference interaction mechanism to convert the MSDMPUP problem from a discrete space to a continuous space, which increases the reasonableness of the target task allocation and the efficiency of the solution.

4.1.1. Pre-Planning Layer

The pre-planning layer aims to obtain an initial allocation reference knowledge pool based on the initial set of tasks, and to construct a mapping strategy to map the multi-satellite imaging task allocation sequence into an evaluation matrix for each sub-objective, generating a knowledge pool of target task evaluations to inform the MSDMPUP planning allocation. Figure 3 shows the schematic diagram of the pre-planning layer, where the gray circles indicate the initial population individuals, the purple circles indicate the task benefit sub-population P 1 individuals, the green circles indicate the task timeliness sub-population P 2 individuals, and the blue circles indicate the task completion sub-population P 3 individuals. It is assumed that six satellites S a t = j = 1 , ...6 | S a t j carry out observations of six target tasks T a s k = i = 1 , , 6 | , T a s k i distributed in different regions.
Firstly, according to the initial task set T a s k = T 1 , , T N t a s k generated by using the MSDMPUP model based on a task rolling window, multiple sets of multi-satellite imaging task assignment sequences are randomly generated as the initial parent population P , in which each set of multi-satellite imaging task assignment sequences is an individual. The NSGA-II algorithm is then applied to the MSDMPUP model to quickly obtain the initial Pareto frontier surface of the initial parent population using a non-dominated sorting method.
Secondly, the frontier surface is optimized using a reference point decomposition method to generate the optimal Pareto frontier surface for the initial allocation reference knowledge pool. The unit of knowledge of the initial allocation reference knowledge pool consists of a matrix of size 5 * N t a s k . It specifically contains five parts, namely, the executing satellite number sequence, target task sequence, task benefit, task response time and task completion, with the following expressions:
P o o l a l l o c a t i o n = S a t 1 S a t 2 S a t N S a t T a s k 1 T a s k 2 T a s k N t a s k G a i n 1 G a i n 2 G a i n N t a s k C o m p l e t i o n 1 C o m p l e t i o n 1 C o m p l e t i o n N t a s k T i m e 1 T i m e 2 T i m e N t a s k ,
Thirdly, the fitness values of the task benefit objective function, the task timeliness objective function, and the task completion objective function are used as the evaluation criteria, and the individuals on the optimal Pareto frontier are sorted in descending order of the fitness values of the three sub-objective functions to generate the task benefit sub-population P 1 , the task timeliness sub-population P 2 , and the task completion sub-population P 3 .
Finally, a mapping strategy is constructed to map the sequence of execution satellites and target tasks in the discrete evaluation space to the continuous scheme evaluation space through the corresponding function values of each sub-objective, which is specified as follows:
f n φ : P n ( S a t j , T a s k i ) E g ( i , j ) C s , n 1 , 2 , 3 , g E G a i n , E T i m e , E C o m p ,     j S a t = S a t 1 , , S a t N S a t , i T a s k = t 1 , , t N t a s k .
where φ uses the objective function evaluation matrix E g ( i , j ) as the replacement matrix for the spatial mapping, and the individuals in the task benefit sub-population P 1 , the task timeliness sub-population P 2 , and the task completion sub-population P 3 are mapped to the task benefit evaluation matrix E G a i n , the task timeliness evaluation matrix E T i m e , and the task completion evaluation matrix E C o m p using the mapping strategy. The evaluation matrices for each objective function form a continuous vector space C s and are stored in the target task evaluation knowledge pool P o o l E v a l u a t e = E G a i n , E T i m e , E C o m p .

4.1.2. Preference Interaction Layer

The preference interaction layer aims to generate a new population generation based on the commands sent by satellite controllers and to generate user preference planning schemes by optimizing the search using a knowledge transfer strategy for the multi-objective evolutionary algorithm; these data generation steps are carried out to realize the dynamic adjustment of the user preference task planning schemes using the optimization search process. Satellite controllers can send demand instructions to the system to provide empirical knowledge according to the actual application requirements. Figure 4 shows the schematic diagram of the preference interaction layer.
Firstly, the satellite controllers perform single command, double command, and multiple command sending according to the actual requirements. According to the difference in the sent command compared with the new generation population based on individual preference selection, the new generation population construction method is as follows:
s i n g l e   i n s t r u c t i o n , P n e w = 50 % P + 50 % P i , i = 1 , 2 , 3 D o u b l e   i n s t r u c t i o n , P n e w = 40 % P + 30 % P i + 30 % P j , i j , i , j = 1 , 2 , 3 M u l t i p l e   i n s t r u c t i o n , P n e w = 40 % P + 20 % P 1 + 20 % P 2 + 20 % P 3 .
where P n e w denotes the new generation population, P denotes the initial parent population, and P i denotes the sub-populations, specifically the task benefit sub-population P 1 , the task timeliness sub-population P 2 , and the task completion sub-population P 3 .
Secondly, the knowledge transfer strategy for the multi-objective evolutionary algorithm (HPIM–KTSMOEA) is used to perform evolutionary operations on the new generation population by calculating the crowding distance between the knowledge as well as the otherness of the knowledge. According to the otherness of the knowledge as a medium of knowledge transfer between the historical environment and the current environment, the algorithm adaptively selects the knowledge transfer strategy for solving and outputting the MSDMPUP user preference planning scheme when reaching the termination condition.

4.1.3. Interaction Output Layer

The interaction output layer is aimed at reflecting the sub-target evaluation values according to the constructed inverse mapping strategy to obtain a sequence of multiple satellite and target mission assignments. The system outputs the latest mission planning scheme for the satellite controllers in order to achieve manual intervention in the final scenario selection. The inverse mapping strategy matches the current evaluation matrix obtained from the evolution operation with the historical sub-objective evaluation values of the corresponding evaluation matrix in the target task evaluation knowledge pool one by one. Next, the strategy matches the historical sub-objective evaluation value with the smallest difference with the current sub-objective evaluation value in the evaluation matrix and then maps it to the corresponding sequence relationship. The inverse mapping strategy is specified as follows:
  • The satellite controller sends a single command:
f n φ 1 : { min | E g C u r r e n t ( i , j ) E g H i s t o r y ( i , j ) | } P n ( S a t j , T a s k i ) , g E G a i n , E T i m e , E C o m p , j S a t = S a t 1 , , S a t N S a t , i T a s k = t 1 , , t N t a s k ,
where E G a i n C u r r e n t ( i , j ) , E T i m e C u r r e n t ( i , j ) , E C o m p l e t i o n C u r r e n t ( i , j ) represent the task benefit value, task timeliness value, and task completion degree of the current satellite performing the target task, respectively; φ 1 denotes the inverse mapping function; and E G a i n H i s t o r y ( i , j ) , E T i m e H i s t o r y ( i , j ) , E C o m p l e t i o n H i s t o r y ( i , j ) represent the corresponding historical task benefit value, historical task timeliness value, and historical task completion degree of the satellite performing the target task in the target task evaluation knowledge pool, respectively.
2.
Satellite controllers send double instructions:
( f n & f m ) φ 1 : { min | E G a i n C u r r e n t ( i , j ) + E T i m e C u r r e n t ( i , j ) 2 E G a i n H i s t o r y ( i , j ) + E T i m e H i s t o r y ( i , j ) 2 | } P ( S a t j , T a s k i ) , n m E G a i n , E T i m e , E C o m p ,   j S a t = S a t 1 , , S a t N S a t , i T a s k = t 1 , , t N t a s k ,
3.
Satellite controllers send multiple commands:
( f 1 , f 2 , f 3 ) φ 1 : min | E G a i n C u r r e n t ( i , j ) + E T i m e C u r r e n t ( i , j ) + E C o m p l e t i o n C u r r e n t ( i , j ) 2 E G a i n H i s t o r y ( i , j ) + E T i m e H i s t o r y ( i , j ) + E C o m p l e t i o n H i s t o r y ( i , j ) 2 | P ( S a t j , T a s k i ) .
Taking the reflective output of the task benefit priority command sent to the satellite controller as an example, the principle of the interactive output layer is introduced, as shown in Figure 5. The target assessment matrix of task benefits obtained by the evolutionary operation of the new generation population is E G a i n C u r r e n t = [ 191 , 405 , 156 , 323 , 334 , 417 ] , and the historical task benefit assessment matrix E G a i n H i s t o r y is used as a reflective reference. The current sub-objective assessment value of 191 is matched one by one with the assessment value in the first column of the historical task benefit assessment matrix E G a i n H i s t o r y , and the historical sub-objective assessment value of 186, which has the smallest difference with the current sub-objective assessment value is matched, and its reflective projections are retrieved to obtain the satellite and target task sequence relationship as [ S a t 1 , T a s k 4 ] . The rows in which the historical sub-objective assessment value of 186 is located in are invalid values; these values are removed from the assessment matrix and matched against the next sub-objective assessment value of 405. The loop traversal method is used to output the final allocation sequence: [ S a t 1 , T a s k 4 ] , [ S a t 2 , T a s k 2 ] , [ S a t 3 , T a s k 5 ] , [ S a t 4 , T a s k 3 ] , [   S a t 5 , T a s k 6 ] , [   S a t 6 , T a s k 4 ] .

4.2. Knowledge Transfer Strategy for Multi-Objective Evolutionary Algorithm

The multi-objective evolutionary algorithm (MOEA) is an intelligent search algorithm that finds the optimal solution to a problem with multiple optimization objectives through continuous iterative evolution [33,34]. The MSDMPUP problem requires the simultaneous consideration of the three optimization objectives of task benefit, task completion, and task timeliness, which is a typical multi-objective optimization problem. In addition, MSDMPUP is affected by the dynamically changing requirements of users in complex environments; the problem space is more massive, facing much greater problem uncertainty. Moreover, how to ensure that the algorithm responds efficiently to changes in the environment while reducing computational consumption is a key issue to be solved.
Therefore, this section proposes a knowledge transfer strategy for the multi-objective evolutionary algorithm (KTSMOEA), which is capable of considering both regular and stochastic changes, efficiently tracking environmental changes, and accelerating the convergence of populations in new environments. The schematic diagram of the algorithm is shown in Figure 6.
Firstly, the MSDMPUP knowledge pool is constructed for storing the current environmental knowledge and historical knowledge as a pool of solutions for solving the MSDMPUP problem as follows:
R e p o s i t o r y k n o w l e d g e = v = 1 n 1 t = 1 m 1 H i s _ k v , t + k v , t K = Q , C , Q P O S ( t ) , C = μ 1 , , μ n ,
where denotes the t -th knowledge of generation v-th in the MSDMPUP knowledge pool. Knowledge is the optimal solution to solve the MSDMPUP problem, represented in the optimization search process, as the P a r e t o optimal subset, as well as the implicit features of the P a r e t o optimal subset, may provide valuable information for exploring future environments in each historical environment. H i s _ k v , t denotes the historical knowledge. Q P O S ( t ) denotes the P a r e t o optimal subset in the t -th generation of the historical environment, C = μ 1 , , μ n denotes the set of implicit features, and μ j is the average of all solutions in the optimal subset Q of the historical environment P a r e t o along each dimension of the decision space, which is mathematically described as follows:
μ j = 1 | Q | i = 1 | Q | x i j , x i Q ,
where x i j denotes the j -th dimension of the i -th element in the optimal subset Q of the historical environment P a r e t o .
Secondly, the maximum storage capacity of the MSDMPUP knowledge pool is C capacity , and in order to store more diverse historical knowledge in a limited space, historical knowledge updates are performed for cases beyond C capacity to assist new environments in quickly exploring P a r e t o -optimal mission planning schemes as follows:
D i s _ k = min E u c ( C i , C j ) max E u c ( C i , C j ) , i , j [ 1 , k ] R e p o s i t o r y k = R e p o s i t o r y k 10 % k ( min ( D i s _ k ) ) ,
where E u c ( M i , M j ) denotes the Euclidean distance between knowledge and D i s _ k denotes the crowding distance between knowledge. Smaller D i s _ k indicates greater similarity between the knowledge and provide redundant information for future generations to optimize the search. Therefore, the last 10 % of knowledge from D i s _ k is removed from the MSDMPUP knowledge pool to ensure that diverse historical knowledge is always stored within the knowledge pool.
Determining the otherness between the historical and current environments and identifying the most appropriate knowledge from the knowledge pool are key issues for efficient knowledge transfer. The multi-objective optimization problem is similar in the corresponding P a r e t o -optimal subsets and P a r e t o -fronts in two environments with small otherness, and the algorithm search process can be accelerated by selecting the historical environment with small differences from the current environment for history transfer, as follows:
o t h e r n e s s t = i = 1 n + 1 d i s ( F ( x i , t ) , F ¯ ( x i , t ) ) n + 1 ,
where o t h e r n e s s t denotes the environmental otherness of the t -th knowledge, and the F ( x i , t ) , F ¯ ( x i , t ) ,respectively, denote the global optimization objective function values corresponding to the i -th position of the t -th knowledge in the historical and current environments.
Finally, knowledge transfer strategies are adaptively selected based on knowledge otherness to transfer knowledge from the historical environment to the current environment, including the following:
C k ( k ) = k r a n d H i s _ k v , t , o t h e r n e s s t e × 10 4 H i s _ k v , t S D M S , o t h e r n e s s t e × 10 4 ,
If o t h e r n e s s t e × 10 4 indicates that the current environment is similar to the historical environment, the knowledge corresponding to the historical environment will be directly transferred to the current environment, which means that the knowledge in the historical environment will be combined with the random individuals in the current environment to participate in the optimization search as the next-generation population.
If o t h e r n e s s t e × 10 4 indicates that the current environment is different from the historical environment, and the knowledge in the historical environment is closer to the P a r e t o of the real environment than the knowledge in the current environment, then the subspace distribution alignment method S D M S is used to mine the association rules between the historical environment and the current environment, which then transfers the knowledge from the historical environment to the current environment. The subspace distribution alignment method S D M S is as follows:
S D M S = H s A 1   C T A 2 λ 1 λ 2 λ 2 λ 3 cov ( H i s _ k v , t , C k ) H s A 1   C T A 2 T .
where H s is the eigenvector of H i s _ k , C T is the eigenvector of C k . O S is the orthogonal complement of H s , and A 1 , A 2 are orthogonal matrices of H s T C T and O s T C T , respectively. λ 1 , λ 2 , λ 3 are the diagonal matrices of H s and C T . The subspace distribution alignment method S D M S enables the transfer of knowledge from the historical environment to the current environment in order to quickly adapt to the current environment. The specific description is shown in Algorithm 2.
Algorithm 2. The framework of the knowledge transfer strategy for multi-objective evolutionary algorithm
Input: Initial population P, Task benefit subpopulation P1, Task timeliness subpopulation P2, Task completion subpopulation P3
Output: A series of MSDMPUP scheme
1 while termination criterion not met
2 /* Construct the MSDMPUP knowledge pool and knowledge dynamic updating*/
Set   R e p o s i t o r y k =
if R e p o s i t o r y k C capacity
  R e p o s i t o r y k = v = 1 n 1 t = 1 m 1 H i s _ k v , t + k v , t ./* MSDMPUP knowledge pool */
  else
D i s _ k = C d i s t a n c e ( k ) /* Determine the crowding distance between knowledge*/
R e p o s i t o r y k = R e p o s i t o r y k 10 % k ( min ( D i s _ k ) ) /* knowledge update*/
  end
3 /* knowledge transfer strategies according to knowledge otherness */
o t h e r n e s s t = i = 1 n + 1 d i s ( F ( x i , t ) , F ¯ ( x i , t ) ) / n + 1 ./* Determine otherness between environments */
if   o t h e r n e s s t e × 10 4
C k ( k ) = k r a n d H i s _ k v , t /*Direct transfer*/
  else
C k ( k ) = H i s _ k v , t S D M S ./*Subspace distribution alignment method*/
  end
4 Perform basic multi-objective evolutionary algorithm
5 End

4.3. Computational Complexity of HPIM–KTSMOEA

In this section, the computational complexity of the HPIM–KTSMOEA algorithm is described in terms of two dimensions, time complexity and space complexity. Firstly, time complexity is affected by four components: task set and population initialization, whose time complexity is usually proportional to the number of tasks and population size, denoted as O ( N T a s k × n ) . N T a s k denotes the target number of tasks and n denotes the population size; the time complexity of calculating the fitness value of the objective function and processing the constraints depends on the objective function and the objective constraint complexity, denoted as O ( f ( n ) ) . f ( n ) is the objective function for a population size of n . Preference interaction mechanism, time complexity depends on the number of user preferences, is denoted as O ( n × P ) . The time complexity of the multi-objective evolutionary algorithm based on knowledge transfer strategy depends on the construction and updating of MSDMPUP knowledge base, denoted as O ( K ) , and K is the size of knowledge pool. Therefore, the HPIM–KTSMOEA time complexity is O ( T × n ) + N _ g e n e r a t i o n × ( O ( f ( n ) ) + O ( n × P ) + O ( K ) ) , where N _ g e n e r a t i o n represents the number of population iterations.
Secondly, the space complexity is related to the amount of data that the algorithm needs to store, which in HPIM–KTSMOEA is determined by three components: input parameters and task set, population and individual information, and MSDMPUP knowledge pool. Input parameters and task sets, whose spatial complexity depends on the number of tasks, denoted as O ( N T a s k ) ; population and individual information, whose spatial complexity depends on the size of the population, denoted as O ( n ) ; and MSDMPUP knowledge base, whose spatial complexity depends on the size of the knowledge base, denoted as O ( K ) . Therefore, the HPIM–KTSMOEA space complexity is O ( N T a s k + n + K ) .

5. Experimental Result and Analysis

To verify the effectiveness and stability of the HPIM–KTSMOEA algorithm for solving the MSDMPUP problem, the following four sets of simulation experiments were carried out in this work using MatlabR2020b software: Experiment 1. Effectiveness of the MSDMPUP model based on a task rolling window; Experiment 2. Effectiveness and stability of HPIM–KTSMOEA under small-scale MSDMPUP; Experiment 3. Effectiveness and stability of HPIM–KTSMOEA under large-scale MSDMPUP; and Experiment 4. Performance analysis of the HPIM–KTSMOEA algorithm.

5.1. Experiment Setting

According to the MSDMPUP mission requirements, this experiment generated 10 imaging satellites from the simulation scenario set of the Satellite Simulation Kit (STK). Each satellite carried only one sensor and the maximum satellite yaw angle did not exceed 40°. Satellite orbit parameters and payload parameters are shown in Table 1. The target tasks are generated randomly, and the target task set with target task assignment constraints, target visibility, satellite performance, and target task assignment are generated by using the MSDMPUP model based on a task rolling window. The experimental simulation time was set as 00:00:00~24:00:00, 9 February 2024 (UTCG). In addition, the HPIM–KTSMOEA algorithm parameters were as follows: the individual population size was set to 100 and the maximum number of iterations was set to 1000 generations.

5.2. Experiment 1: Effectiveness of the MSDMPUP Model Based on a Task Rolling Window

To verify the effectiveness of the MSDMPUP model based on task rolling window (MSDMPUP–TRW), the MSDMPUP–TRW model in this study was compared with the multi-satellite imaging mission planning model based on event triggering (MSIMP–ET) [35], and the multi-satellite imaging mission planning model based on periodic-triggered rolling, MSIMP–PTR [36]. Task sets carrying the target task importance degree were generated using three types of methods.
The MSIMP–ET model addresses the problem of high dimensionality and time-varying feature information. The model formulates the multi-satellite scheduling problem as a dynamic combinatorial optimization problem, which reduces losses due to irrational allocation of satellite resources by triggering events, which means that the rolling is triggered when a new demand arrives. The MSIMP–PTR model improves the quality of the multi-satellite imaging mission planning scheme by introducing a periodic trigger rolling window into the multi-satellite scheduling problem in response to the common problem of uncertainty in the arrival time of dynamic missions and updating the target mission importance according to a fixed period.
In this experiment, 45 target tasks are randomly generated, and the initial moment is used to generate the task set T a s k = T 1 , T 2 , , T 45 of the target task importance degree using the three types of models, MSDMPUP–TRW, MSIMP–ET, and MSIMP–PTR. where the MSDMPUP–TRW model sets the trigger threshold to 2, the MSIMP–PTR model is given a trigger period of Δ T i m e , and the target task importance degree is specified as shown in Table 2. “Target” indicates the target mission number, “ P M S D M P H M A T R W ”, “ P M S I M P E T ”, “ P M S I M P P T R ” denote the target task importance degree generated by using the MSDMPUP–TRW, MSIMP–ET, and MSIMP–PTR models, respectively.
During the planning period of the experiment, the satellite control center monitored task changes at different moments. The target task importance degree of the target tasks at different moments is shown in Figure 7, where “*” indicates the target task importance degree corresponding to each target task and connects the target tasks “*” with straight lines of different colors.
Figure 7a shows the target task importance degree of the 45 target tasks generated at the initial moment in the models MSDMPUP–TRW, MSIMP–ET, and MSIMP–PTR, and it can be seen from the figure the same target task is generated under the three types of models with a slightly different degree of importance, but the gap in the degree of importance of the target tasks is small. Figure 7b–d show the changes in the importance of the target tasks for the three types of models, MSDMPUP–TRW, MSIMP–ET, and MSIMP–PTR, during the planning time, and the red boxes indicate the importance of the target tasks that changed more significantly after the dynamic update during the planning cycle.
As can be seen from Figure 7, all three types of models are capable of generating dynamically updated target task importance according to the user’s task requirements. The MSIMP–ET model is more sensitive to the external environment, and the rolling completion of the update of the target task importance for each triggered task is capable of handling emergency tasks in a timely manner, but it will lead to frequent scheduling when facing more triggering events. The MSIMP–PTR model performs periodic scrolling according to the set time interval to realize dynamic updating of the importance of the target task, but the planning cycle is long and cannot sufficiently satisfy the high timeliness required of the emergency task. The MSDMPUP–TRW model combines the updating methods of periodic triggering and event triggering and sets the trigger threshold for the adaptive processing of multiple emergency tasks, which avoids frequent triggering and at the same time possesses the capability of timely processing of the emergency task.
The simulation results of Experiment 1 show that the MSDMPUP–TRW model proposed in this work can respond quickly to environmental changes and avoid frequent calculation and updating of the planning scheme to achieve timely updating of the importance of the target mission. This, in turn, affects the observation order of the satellite performing the target mission, enabling the satellite resources to be deployed in a timely manner according to the mission requirements and improving the planning efficiency.

5.3. Experiment 2: Effectiveness and Stability of HPIM–KTSMOEA under Small-Scale MSDMPUP

To verify the effectiveness of the HPIM–KTSMOEA algorithm in solving small-scale MSDMPUP, 45 target tasks were randomly generated in the local range of longitude [3, 54] and latitude [73, 135]. It is assumed that at time T, the target tasks located in the sea area (T1, T3, T4, T9, T17, T20, T24, T27, T29, T31, T35, T37, T40, and T42) dynamically update the importance of the target tasks using the MSDMPUP–TRW model. We generated a task set with the target task importance. The geographical locations of 45 target tasks and their corresponding target task importance values are shown in Figure 8. The geographic distribution of target tasks is shown in Figure 8a, where red circles represent target tasks. The target task importance degree at the initial time and time T is shown in Figure 8b. The red box indicates the importance of target tasks that changed significantly after the dynamic update process.
The experiment was performed as follows: Set the case of sending preference instructions without satellite controllers as Instance1. The HPIM–KTSMOEA algorithm solves instance1 to obtain the initial optimal assignment task planning scheme of MSDMPUP, as shown in Table 3, where, “T” indicates the target task number, “S” indicates the satellite number, “St” and “Et” indicate the start time (hour: minute: second) and end time (hour: minute: second) of satellite observation, respectively, and “--” indicates that the task assignment time window does not meet the target task assignment constraints, and, thus, no imaging task is arranged. As can be seen from Table 3, all 45 target tasks are effectively assigned. Therefore, the HPIM–KTSMOEA algorithm achieves better task planning results.
Thereafter, to verify the effectiveness of the hybrid preference interaction mechanism on the basis of the same target task set generated in the same scene, the HPIM–KTSMOEA algorithm was used to solve eight groups of instances according to different user preference requirements. The instruction input for each instance and the optimal simulation results of the HPIM–KTSMOEA algorithm solving different instances are shown in Table 4.
The following evaluation indexes were used to analyze the results of this experiment:
  • Ideal task benefit: I b e n e f i t = i = 1 N T a s k w i . This represents the sum of the target task importance degree in the ideal case where all target tasks are observed.
  • Actual task benefit: R b e n e f i t = i = 1 N T a s k j = 1 N S a t o = 1 N O r b x i , j , o   i = 1 N T a s k w i . This represents the sum of the target task importance degree of the actual observed target.
  • Average task response time: T r e s p o n s e = i = 1 N T a s k j = 1 N S a t o = 1 N O r b x i , j , o   W i r e s p o n s e / N T a s k . This represents the average task response time W i r e s p o n s e of the actual observation target.
  • Task completion rate: R c o m p l e t i o n = i = 1 N T a s k j = 1 N S a t o = 1 N O r b x i , j , o   / N T a s k . This represents the proportion of the number of actual observed target tasks to the total number of tasks.
The performance analysis of the HPIM–KTSMOEA algorithm for solving different instances is shown in Figure 9. Figure 9 analyses the performance of different instances using target task ideal benefit, actual task benefit, task response time, and task completion rate, which is significant in revealing the impact of different user preference instructions on task planning performance. In particular, Instance1 represents the no user preference input instance in the ideal case, Instance2, Instance3, and Instance4 represent the task benefit priority instruction, task timeliness priority instruction, and task completion priority instruction input instances in the case of a single preference instruction, respectively. Instance5, Instance6, and Instance7 denote the double preference instruction case with task benefit priority instruction and task timeliness priority instruction inputs, task benefit priority instruction and task completion priority instruction inputs, task timeliness priority instruction and task completion priority instruction inputs, respectively. Instance8 denotes the multiple instruction inputs.
The performance analysis of the HPIM–KTSMOEA algorithm for solving different instances is shown in Figure 9. Figure 9a–c represent the convergence curves of the HPIM–KTSMOEA algorithm under the three dimensions of the task benefit objective function, task timeliness objective function, and task completion objective function, respectively, and Figure 9d represents the average values of the HPIM–KTSMOEA algorithm under the different evaluation indexes.
In Figure 9a, the HPIM–KTSMOEA algorithm for each instance under the target task benefit evaluation index is as follows: Instance2 > Instance6 > Instance5 > Instance8 > Instance4 > Instance3 > Instance7 > Instance1. This index can reflect the performance of the HPIM–KTSMOEA algorithm in terms of benefits, and by using the target task benefits for different instances can assess the impact of the user input task benefit priority instruction on the system benefits.
In Figure 9b, the response time of each instance under the task timeliness evaluation index is as follows: Instance3 < Instance7 < Instance5 < Instance8 < Instance2 < Instance4 < Instance6 < Instance1. This metric reflects the performance of the HPIM–KTSMOEA algorithm in terms of timeliness, by comparing the average response time under different instances, and by evaluating how different settings of the user’s timeliness requirements affect the efficiency of task planning. Shorter response time means more efficient task planning.
From Figure 9c, it can be seen that the order of each instance under the task completion evaluation index is as follows: Instance4 > Instance6 > Instance7 > Instance8 > In-stance2 > Instance5 > Instance3 > Instance1. By comparing task completion under different instances, we can understand how different user requirements for task completion affect the execution of task planning. Higher task completion means that task planning is better able to satisfy users’ needs.
Combined with the experimental results in Table 4 and Figure 9, Experiment 2 shows that the HPIM–KTSMOEA algorithm proposed in this work is able to solve the MSDMPUP model effectively in small-scale target task scenarios, and the hybrid preference interaction mechanism allows satellite controllers to adjust the preferences of the mission planning scheme according to the actual needs to improve the adaptability of the multi-satellite dynamic mission planning scheme.

5.4. Experiment 3: Effectiveness and Stability of HPIM–KTSMOEA under Large-Scale MSDMPUP

To verify the effectiveness of the HPIM–KTSMOEA algorithm for solving MSDMPUP on a large scale, Experiment 3 randomly generates 100 target tasks for simulating user imaging task requirements on a global scale in longitude [−90, 90] and latitude [−180, 180], and the geographical distribution of the 100 target tasks on a global scale is shown in Figure 10.
It is assumed that the observed target tasks for demand changes globally include (T10, T23, T31, T39, T49, T57, T68, T79, and T94) at the moment of T. The initial moment and the importance of the target tasks at the moment of T generated by using the MSDMPUP–TRW model are shown in Figure 11a. To more directly demonstrate the allocation relationship between the satellite and the target mission in the case of the no-satellite controller preference command sending, the HPIM–KTSMOEA algorithm is used to solve the visualization of the initial multi-satellite dynamic mission planning scheme in this scenario. As shown in Figure 11b, the x-axis denotes the target mission number, the y-axis denotes the target mission corresponding to the target mission importance, and the z-axis denotes the executing satellite number. To show more clearly the target missions performed by different satellites, different colored squares are used to represent the target missions performed by different satellites.
The performance analysis of the HPIM–KTSMOEA algorithm for solving different instances of MSDMPUP at a large scale is shown in Table 5 and Figure 11. Table 5 evaluation metrics included ideal target task benefits I b e n e f i t , actual task benefits R b e n e f i t , average task response time T r e s p o n s e , and task completion rate R c o m p l e t i o n .
As can be seen from the data in Table 5, the actual task benefits in Instance2, Instance5, Instance6, and Instance8 are more similar to the ideal benefits of the target task. Moreover, the average task response time in Instance3, Instance5, Instance7, and Instance8 is shorter than that in other instances, and the task completion rate is greater than 90% in Instance4, Instance6, Instance7, and Instance8.
From the multiple convergence curves in Figure 12, it can be seen that the HPIM–KTSMOEA algorithm for solving the MSDMPUP model finds as many allocation schemes as possible at the beginning of the iteration, and as the number of iterations increases, it quickly converges to the optimal scheme among different instances. This result proves the effectiveness of the knowledge transfer strategy of the HPIM–KTSMOEA algorithm, which is able to invoke similar knowledge in the historical environment to facilitate the fast convergence of the algorithm.
To further validate the scalability of the MSDMPUP validation of the HPIM–KTSMOEA algorithm at a large scale, we randomly generate 150, 200, 250, and 300 target tasks to simulate the user imaging requirements at the same latitude and longitude range with a step size of 50.
The performance analysis of the HPIM–KTSMOEA algorithm for solving different instances of MSDMPUP at a large scale is shown in Table 6 and Figure 13. The solving instances include no preference instruction input, single instruction input, dual instruction input, and multiple instruction input, with a total of four types and eight sets of instances, as shown in the column of instruction input in Table 6. Meanwhile, the evaluation indexes in Table 6 include target task ideal benefit I b e n e f i t , actual task benefit R b e n e f i t , task response time T r e s p o n s e , and task completion rate R c o m p l e t i o n .
The performance analysis of the HPIM–KTSMOEA algorithm for solving MSDMPUP at four scales of 150, 200, 250, and 300 target tasks is shown in Figure 13. Figure 13a demonstrates the target task benefits that the HPIM–KTSMOEA algorithm can obtain under different target task sizes, reflecting that the algorithm has a good ability to optimize resource allocation and task scheduling. Figure 13b demonstrates the time response of the HPIM–KTSMOEA algorithm under different target task sizes, and the algorithm is always able to complete the target task allocation with a low response time. Figure 13c demonstrates that the instances of the HPIM–KTSMOEA algorithm under different target task sizes all have high task completion.
Experiment 3 shows that the HPIM–KTSMOEA algorithm proposed in this work demonstrates significant advantages when dealing with large-scale target task scenarios, and it cannot only effectively solve the MSDMPUP model, but also shows excellent effectiveness, stability, and scalability for target tasks at all scales. At the same time, based on the method proposed in this work, satellite controllers can send preference commands to the system, which can optimize the search process to adjust the preferences of the MSDMPUP scheme, and the HPIM–KTSMOEA algorithm is able to solve the optimal scheme within a limited number of iterations and with a high degree of task completion in all instances. In addition, single instruction delivery significantly outperforms dual and multiple instructions for task planning adjustment for individual preference requirements.

5.5. Experiment 4: Performance Analysis of the HPIM–KTSMOEA Algorithm

This section compares the proposed HPIM–KTSMOEA algorithm with the knowledge transfer for multi-objective evolutionary algorithm, (KT–DMOEA [37]), bidirectional knowledge transfer for MOEA algorithm (BKT–MOEA [38]), hybrid knowledge transfer for MOEA algorithm (HKT–MOEA [39]), bi-objective knowledge transfer for MOEA algorithm (BOKT–MOEA [40]), and the helper objective-based multifactorial evolutionary algorithm (HO–MEA [41]); in total, five algorithms are compared for performance.
The KT–MOEA algorithm uses a weighted approach to guide the environment to generate a new generation of populations by combining knowledge of inflection points during the dynamic evolution process, which can lead to effective tracking of environmental changes in the multi-satellite imaging mission planning system. The BKT–MOEA algorithm designs a bi-directional knowledge migration strategy that is able to identify preferred target tasks in the process of searching for knowledge-optimized searches, making the task planning scheme closer to the actual requirements. The HKT–MOEA algorithm evaluates task relevance based on the distributional characteristics of the evolving population, adaptively performs knowledge transfer, and is able to retrieve more globally optimal task planning solutions. The BKT–MOEA algorithm uses two different measures to accurately measure the similarity between different types of tasks to transfer knowledge and improve the solution efficiency of the task planning system. The HO–MEA algorithm uses a multi-mission optimization framework to transfer knowledge between the auxiliary algorithms and the original problem, improving the accuracy of the dynamic mission planning scheme for multi-satellite imaging.
Eight sets of instance simulation experiments were carried out under 100 target task scenarios, in which the instance interactive operation is consistent with the descriptions of Experiments 2 and 3, and the instances all use the same number of iterations and population size. The experimental results of the performance comparison between the HPIM–KTSMOEA algorithm and the other algorithms are shown in Table 7 and Figure 14.
To comprehensively evaluate the effectiveness of the HPIM–KTSMOEA algorithm on the MSDMPUP problem, this section uses the actual task benefit R b e n e f i t , average task response time T r e s p o n s e , and task completion rate R c o m p l e t e as the evaluation indexes. Table 7 and Figure 14 shows that the HPIM–KTSMOEA algorithm has better stability, and has obvious advantages over other algorithms in terms of the actual task benefit R b e n e f i t , average task response time T r e s p o n s e , and task completion rate R c o m p l e t e in solving the eight sets of instances under MSDMPUP. This demonstrates that the HPIM–KTSMOEA algorithm is able to effectively deal with the dynamic mission planning problem of multi-satellite imaging under different user requirement inputs.
We further analyzed the HPIM–KTSMOEA algorithm, KT–MOEA algorithm, BKT–MOEA algorithm, HKT–MOEA algorithm, BOKT–MOEA algorithm, and HO–MEA algorithm data using box plots in Instance8. Figure 15a–c represent the box plots of different algorithms under the three dimensions of the task benefit objective function, the task timeliness objective function, and the task completion objective function, respectively. The top and bottom of the box plots represent the upper quartile and lower quartile, respectively, the top and bottom edge lines represent the maximum and minimum values, the middle line segment represents the average, and the black plus sign represents the outliers.
Figure 15 shows that the HPIM–KTSMOEA algorithm has advantages over other algorithms in the three dimensions of the task benefit objective function, the task timeliness objective function, and the task completion objective function, with a larger range of converged datasets and fewer outliers, indicating that the introduction of the knowledge transfer strategy is able to quickly find more high-quality task planning solutions in the search space. The KT–MOEA algorithm and BOKT–MOEA algorithm show smaller and faster convergence among the convergence data set range, which indicates that the algorithm may fall into a local optimum. The BKT–MOEA algorithm and HO–MEA algorithm median task benefits and response times are off-center and show a tendency to be biased, indicating that the algorithm solution data are not evenly distributed in the optimal search space. The HKT–MOEA algorithm has small task benefit values, long response times, and poor task completion, which indicates that it cannot sufficiently adapt to dynamically changing task requirements.
The simulation results of Experiment 4 show that the HMIM–KTSMOEA algorithm proposed in this study introduces the knowledge transfer strategy, which can effectively cope with the stochastic changes in the task planning environment in the case of multiple-user demand, effectively making a trade-off between the solution performance and the computational efficiency, and solving the optimal solution in a shorter time. Moreover, the HMIM–KTSMOEA algorithm outperforms the other five algorithms in terms of optimization objective fitness value and optimization efficiency. Therefore, the HMIM–KTSMOEA algorithm is the superior choice to solve the MSDMPUP model.

6. Conclusions

In this work, we proposed the hybrid preference interaction mechanism and knowledge transfer strategy for the multi-objective evolutionary algorithm (HPIM–KTSMOEA). Firstly, an MSDMPUP model based on a task rolling window was proposed to achieve timely updating of target task importance by responding quickly to environmental changes through the simultaneous application of periodic triggering and event triggering methods. Secondly, the hybrid preference interaction mechanism was proposed to construct a mapping strategy to map the MSDMPUP allocation sequence to a sub-objective evaluation matrix, which directs the command inputs from the satellite controllers to guide the new population generation of the new generation and achieves the preference-based adjustment of the mission planning scheme. Finally, a knowledge transfer strategy for the multi-objective evolutionary algorithm was proposed to transfer knowledge according to the difference between the current and historical environments and invoke similar knowledge in the historical environment to promote the fast convergence of the algorithm. The simulation results show that this method can quickly solve the MSDMPUP problem under different task requirements, with outstanding performance such as high task benefit, short response time, and high task completion rate.
The HPIM–KTSMOEA method proposed in this paper works through a hybrid preference mechanism to satisfy different user preference requirements, however, the hybrid preference interaction mechanism may face the problem of insufficient dynamic adaptability to cope with rapidly changing dynamic environments in real application scenarios. To overcome this limitation, the next study will combine advanced prediction techniques and reinforcement learning techniques to develop an intelligent interactive collaborative-based multi-satellite imaging dynamic mission planning system, which utilizes deep learning prediction models to predict future environmental changes and enables the algorithms to quickly adjust the preference mechanism through intelligent interactive collaboration to achieve more efficient and intelligent multi-satellite imaging dynamic mission planning. Therefore, it is significant to investigate the problem of an intelligent interactive collaborative-based multi-satellite imaging dynamic mission planning system.

Author Contributions

Methodology, X.Y. and M.H.; supervision, M.H.; validation, G.H. and Y.W.; writing—original draft, X.Y. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by National Natural Science Foundation of China, grant number 61403416.

Institutional Review Board Statement

Not applicable.

Data Availability Statement

Data are contained within the article.

Acknowledgments

Thanks to Min Hu for his important technical help and for providing experimental equipment and related materials.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Lee, K.; Kim, D.; Chung, D.; Lee, S. Application of Optimal Scheduling for Synthetic Aperture Radar Satellite Constellation: Multi-Imaging Mission in High-Density Regional Area. Aerospace 2024, 11, 280. [Google Scholar] [CrossRef]
  2. Song, Y.; Ou, J.; Wu, J.; Wu, Y.; Xing, L.; Chen, Y. A Cluster-Based Genetic Optimization Method for Satellite Range Scheduling System. Swarm Evol. Comput. 2023, 79, 101316. [Google Scholar] [CrossRef]
  3. Chen, L.; Zheng, Y. Satellite Communication System Resource Scheduling Algorithm Based on Artificial Intelligence. Procedia Comput. Sci. 2023, 228, 551–558. [Google Scholar] [CrossRef]
  4. Wang, D.; Guan, L.; Liu, J.; Ding, C.; Qiao, F. Human–Machine Interactive Learning Method Based on Active Learning for Smart Workshop Dynamic Scheduling. IEEE Trans. Hum.-Mach. Syst. 2023, 53, 1038–1047. [Google Scholar] [CrossRef]
  5. Zhibo, E.; Shi, R.; Gan, L.; Baoyin, H.; Li, J. Multi-Satellites Imaging Scheduling Using Individual Reconfiguration Based Integer Coding Genetic Algorithm. Acta Astronaut. 2021, 178, 645–657. [Google Scholar] [CrossRef]
  6. Rigo, C.A.; Seman, L.O.; Camponogara, E.; Morsch Filho, E.; Bezerra, E.A. A Nanosatellite Task Scheduling Framework to Improve Mission Value Using Fuzzy Constraints. Expert Syst. Appl. 2021, 175, 114784. [Google Scholar] [CrossRef]
  7. Ren, L.; Ning, X.; Wang, Z. A Competitive Markov Decision Process Model and a Recursive Reinforcement-Learning Algorithm for Fairness Scheduling of Agile Satellites. Comput. Ind. Eng. 2022, 169, 108242. [Google Scholar] [CrossRef]
  8. Qu, Q.; Liu, K.; Li, X.; Zhou, Y.; Lu, J. Satellite Observation and Data-Transmission Scheduling Using Imitation Learning Based on Mixed Integer Linear Programming. IEEE Trans. Aerosp. Electron. Syst. 2022, 169, 1989–2001. [Google Scholar] [CrossRef]
  9. Yu, Y.; Hou, Q.; Zhang, J.; Zhang, W. Mission Scheduling Optimization of Multi-Optical Satellites for Multi-Aerial Targets Staring Surveillance. J. Frankl. Inst. 2020, 357, 8657–8677. [Google Scholar] [CrossRef]
  10. Li, G.; Li, X.; Li, J.; Chen, J.; Shen, X. PTMB: An Online Satellite Task Scheduling Framework Based on Pre-Trained Markov Decision Process for Multi-Task Scenario. Knowl.-Based Syst. 2024, 284, 111339. [Google Scholar] [CrossRef]
  11. Valicka, C.G.; Garcia, D.; Staid, A.; Watson, J.-P.; Hackebeil, G.; Rathinam, S.; Ntaimo, L. Mixed-Integer Programming Models for Optimal Constellation Scheduling given Cloud Cover Uncertainty. Eur. J. Oper. Res. 2019, 275, 431–445. [Google Scholar] [CrossRef]
  12. Li, Z.; Li, X. A Multi-Objective Binary-Encoding Differential Evolution Algorithm for Proactive Scheduling of Agile Earth Observation Satellites. Adv. Sp. Res. 2019, 63, 3258–3269. [Google Scholar] [CrossRef]
  13. Liang, J.; Zhu, Y.; Luo, Y.; Zhang, J.; Zhu, H. A Precedence-Rule-Based Heuristic for Satellite Onboard Activity Planning. Acta Astronaut. 2021, 178, 757–772. [Google Scholar] [CrossRef]
  14. Gu, Y.; Han, C.; Chen, Y.; Xing, W.W. Mission Replanning for Multiple Agile Earth Observation Satellites Based on Cloud Coverage Forecasting. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2022, 15, 594–608. [Google Scholar] [CrossRef]
  15. Jianjiang, W.; Xuejun, H.; Chuan, H. Reactive Scheduling of Multiple EOSs under Cloud Uncertainties: Model and Algorithms. J. Syst. Eng. Electron. 2021, 32, 163–177. [Google Scholar] [CrossRef]
  16. Dai, C.-Q.; Zhang, Y.; Yu, F.R.; Chen, Q. Dynamic Scheduling Scheme With Task Laxity for Data Relay Satellite Networks. IEEE Trans. Veh. Technol. 2024, 73, 2605–2620. [Google Scholar] [CrossRef]
  17. Li, F.; Wan, Q.; Wen, F.; Zou, Y.; He, Q.; Li, D.; Zhong, X. Multi-Satellite Imaging Task Planning for Large Regional Coverage: A Heuristic Algorithm Based on Triple Grids Method. Remote Sens. 2024, 16, 194. [Google Scholar] [CrossRef]
  18. Liu, J.; Zhang, G.; Xing, L.; Qi, W.; Chen, Y. An Exact Algorithm for Multi-Task Large-Scale Inter-Satellite Routing Problem with Time Windows and Capacity Constraints. Mathematics 2022, 10, 3969. [Google Scholar] [CrossRef]
  19. Luo, Q.; Peng, W.; Wu, G.; Xiao, Y. Orbital Maneuver Optimization of Earth Observation Satellites Using an Adaptive Differential Evolution Algorithm. Remote Sens. 2022, 14, 1966. [Google Scholar] [CrossRef]
  20. Ou, J.; Xing, L.; Yao, F.; Li, M.; Lv, J.; He, Y.; Song, Y.; Wu, J.; Zhang, G. Deep Reinforcement Learning Method for Satellite Range Scheduling Problem. Swarm Evol. Comput. 2023, 77, 101233. [Google Scholar] [CrossRef]
  21. Wang, J.; Song, G.; Liang, Z.; Demeulemeester, E.; Hu, X.; Liu, J. Unrelated Parallel Machine Scheduling with Multiple Time Windows: An Application to Earth Observation Satellite Scheduling. Comput. Oper. Res. 2023, 149, 106010. [Google Scholar] [CrossRef]
  22. Wu, G.; Luo, Q.; Du, X.; Chen, Y.; Suganthan, P.N.; Wang, X. Ensemble of Metaheuristic and Exact Algorithm Based on the Divide-and-Conquer Framework for Multisatellite Observation Scheduling. IEEE Trans. Aerosp. Electron. Syst. 2022, 58, 4396–4408. [Google Scholar] [CrossRef]
  23. Wang, T.; Luo, Q.; Zhou, L.; Wu, G. Space Division and Adaptive Selection Strategy Based Differential Evolution Algorithm for Multi-Objective Satellite Range Scheduling Problem. Swarm Evol. Comput. 2023, 83, 101396. [Google Scholar] [CrossRef]
  24. Lin, Z.; Ni, Z.; Kuang, L.; Jiang, C.; Huang, Z. Satellite-Terrestrial Coordinated Multi-Satellite Beam Hopping Scheduling Based on Multi-Agent Deep Reinforcement Learning. IEEE Trans. Wirel. Commun. 2024, 27, 1. [Google Scholar] [CrossRef]
  25. Song, Y.; Ou, J.; Pedrycz, W.; Suganthan, P.N.; Wang, X.; Xing, L.; Zhang, Y. Generalized Model and Deep Reinforcement Learning-Based Evolutionary Method for Multitype Satellite Observation Scheduling. IEEE Trans. Syst. Man Cybern. Syst. 2024, 54, 2576–2589. [Google Scholar] [CrossRef]
  26. Chatterjee, A.; Tharmarasa, R. Multi-Stage Optimization Framework of Satellite Scheduling for Large Areas of Interest. Adv. Sp. Res. 2024, 73, 2024–2039. [Google Scholar] [CrossRef]
  27. Chen, M.; Wen, J.; Song, Y.-J.; Xing, L.; Chen, Y. A Population Perturbation and Elimination Strategy Based Genetic Algorithm for Multi-Satellite TT&C Scheduling Problem. Swarm Evol. Comput. 2021, 65, 100912. [Google Scholar] [CrossRef]
  28. Han, X.; Yang, M.; Wang, S.; Chao, T. Continuous Monitoring Scheduling for Moving Targets by Earth Observation Satellites. Aerosp. Sci. Technol. 2023, 140, 108422. [Google Scholar] [CrossRef]
  29. Şahin Zorluoğlu, Ö.; Kabak, Ö. An Interactive Multi-Objective Programming Approach for Project Portfolio Selection and Scheduling. Comput. Ind. Eng. 2022, 169, 108191. [Google Scholar] [CrossRef]
  30. Wang, D.; Qiao, F.; Guan, L.; Liu, J.; Ding, C. Human–Machine Collaborative Decision-Making Method Based on Confidence for Smart Workshop Dynamic Scheduling. IEEE Robot. Autom. Lett. 2022, 7, 7850–7857. [Google Scholar] [CrossRef]
  31. Yoo, S.; Jo, Y.; Bahn, H. Integrated Scheduling of Real-Time and Interactive Tasks for Configurable Industrial Systems. IEEE Trans. Ind. Inform. 2022, 18, 631–641. [Google Scholar] [CrossRef]
  32. Zhang, L.; Deng, Q.; Wen, X.; Zhao, Y.; Gong, G. Optimal Production Scheduling with Multi-Round Information Interaction for Demander-Dominated Decentralized Scheduling Problem. Eng. Appl. Artif. Intell. 2023, 123, 106228. [Google Scholar] [CrossRef]
  33. Tian, Y.; Cheng, R.; Zhang, X.; Cheng, F.; Jin, Y. An Indicator-Based Multiobjective Evolutionary Algorithm With Reference Point Adaptation for Better Versatility. IEEE Trans. Evol. Comput. 2018, 22, 609–622. [Google Scholar] [CrossRef]
  34. Lin, W.; Lin, Q.; Feng, L.; Tan, K.C. Ensemble of Domain Adaptation-Based Knowledge Transfer for Evolutionary Multitasking. IEEE Trans. Evol. Comput. 2024, 28, 388–402. [Google Scholar] [CrossRef]
  35. Cui, K.; Song, J.; Zhang, L.; Tao, Y.; Liu, W.; Shi, D. Event-Triggered Deep Reinforcement Learning for Dynamic Task Scheduling in Multisatellite Resource Allocation. IEEE Trans. Aerosp. Electron. Syst. 2023, 59, 3766–3777. [Google Scholar] [CrossRef]
  36. Li, H.; Li, Y.; Meng, Q.Q.; Li, X.; Shao, L.; Zhao, S. An Onboard Periodic Rescheduling Algorithm for Satellite Observation Scheduling Problem with Common Dynamic Tasks. Adv. Space Res. 2024, 73, 5242–5253. [Google Scholar] [CrossRef]
  37. Wu, L.; Wu, D.; Zhao, T.; Cai, X.; Xie, L. Dynamic Multi-Objective Evolutionary Algorithm Based on Knowledge Transfer. Inf. Sci. 2023, 636, 118886. [Google Scholar] [CrossRef]
  38. Wu, Y.; Ding, H.; Gong, M.; Qin, A.K.; Ma, W.; Miao, Q.; Tan, K.C. Evolutionary Multiform Optimization With Two-Stage Bidirectional Knowledge Transfer Strategy for Point Cloud Registration. IEEE Trans. Evol. Comput. 2024, 28, 62–76. [Google Scholar] [CrossRef]
  39. Cai, Y.; Peng, D.; Liu, P.; Guo, J.-M. Evolutionary Multi-Task Optimization with Hybrid Knowledge Transfer Strategy. Inf. Sci. 2021, 580, 874–896. [Google Scholar] [CrossRef]
  40. Jiang, Y.; Zhan, Z.-H.; Tan, K.C.; Zhang, J. A Bi-Objective Knowledge Transfer Framework for Evolutionary Many-Task Optimization. IEEE Trans. Evol. Comput. 2023, 27, 1514–1528. [Google Scholar] [CrossRef]
  41. Yang, X.; Li, W.; Wang, R.; Yang, K. Helper Objective-Based Multifactorial Evolutionary Algorithm for Continuous Optimization. Swarm Evol. Comput. 2023, 78, 101279. [Google Scholar] [CrossRef]
Figure 1. MSDMP fundamental scenario. (a) Description of initial allocation scheme for mission planning; (b) description of local replanning method; (c) description of replanning scheme for mission planning.
Figure 1. MSDMP fundamental scenario. (a) Description of initial allocation scheme for mission planning; (b) description of local replanning method; (c) description of replanning scheme for mission planning.
Electronics 13 02374 g001
Figure 2. Schematic of the target task rolling window.
Figure 2. Schematic of the target task rolling window.
Electronics 13 02374 g002
Figure 3. Schematic diagram of the pre-planning layer.
Figure 3. Schematic diagram of the pre-planning layer.
Electronics 13 02374 g003
Figure 4. Schematic diagram of the preference interaction layer.
Figure 4. Schematic diagram of the preference interaction layer.
Electronics 13 02374 g004
Figure 5. Schematic diagram of the interaction output layer.
Figure 5. Schematic diagram of the interaction output layer.
Electronics 13 02374 g005
Figure 6. Schematic diagram of KTSMOEA algorithm.
Figure 6. Schematic diagram of KTSMOEA algorithm.
Electronics 13 02374 g006
Figure 7. Target task importance degree at different moments (a) task importance degree in the initial moment; (b) target task importance degree changes for MSDMPUP–TRW; (c) target task importance degree changes for MSIMP–ET; (d) target task importance degree changes MSIMP–PTR.
Figure 7. Target task importance degree at different moments (a) task importance degree in the initial moment; (b) target task importance degree changes for MSDMPUP–TRW; (c) target task importance degree changes for MSIMP–ET; (d) target task importance degree changes MSIMP–PTR.
Electronics 13 02374 g007
Figure 8. Experiment results under 45 target task scenarios (a) the geographical locations of 45 target tasks; (b) target task importance degrees at different moments.
Figure 8. Experiment results under 45 target task scenarios (a) the geographical locations of 45 target tasks; (b) target task importance degrees at different moments.
Electronics 13 02374 g008
Figure 9. HPIM–KTSMOEA simulation results for 45 target task scenarios (a) task benefit value of various instances; (b) task timeliness value of various instances; (c) task completion value of various instances; (d) objective function value of different instances.
Figure 9. HPIM–KTSMOEA simulation results for 45 target task scenarios (a) task benefit value of various instances; (b) task timeliness value of various instances; (c) task completion value of various instances; (d) objective function value of different instances.
Electronics 13 02374 g009aElectronics 13 02374 g009b
Figure 10. Global location distribution of 100 target tasks.
Figure 10. Global location distribution of 100 target tasks.
Electronics 13 02374 g010
Figure 11. The results of mission planning under no preference instructions by satellite controller (a) target task importance degree in different moment; (b) multi-satellite mission assignment.
Figure 11. The results of mission planning under no preference instructions by satellite controller (a) target task importance degree in different moment; (b) multi-satellite mission assignment.
Electronics 13 02374 g011
Figure 12. HPIM–KTSMOEA simulation results under 100 target task scenarios (a) task benefit value of various instances; (b) task timeliness value of various instances; (c) task completion value of various instances; (d) objective function value of different instances.
Figure 12. HPIM–KTSMOEA simulation results under 100 target task scenarios (a) task benefit value of various instances; (b) task timeliness value of various instances; (c) task completion value of various instances; (d) objective function value of different instances.
Electronics 13 02374 g012aElectronics 13 02374 g012b
Figure 13. HPIM–KTSMOEA simulation results in different population sizes (a) task benefit value of various instances; (b) task timeliness value of various instances; (c) task completion value of various instances.
Figure 13. HPIM–KTSMOEA simulation results in different population sizes (a) task benefit value of various instances; (b) task timeliness value of various instances; (c) task completion value of various instances.
Electronics 13 02374 g013
Figure 14. Performance analysis of the HPIM–KTSMOEA algorithm with different algorithms (a) task benefit value of different instances; (b) task timeliness value of different instances; (c) task completion value of different instances.
Figure 14. Performance analysis of the HPIM–KTSMOEA algorithm with different algorithms (a) task benefit value of different instances; (b) task timeliness value of different instances; (c) task completion value of different instances.
Electronics 13 02374 g014
Figure 15. Performance analysis of different algorithms in box plots (a) task benefit value of instance8; (b) task timeliness value of instance8; (c) task completion value of instance8.
Figure 15. Performance analysis of different algorithms in box plots (a) task benefit value of instance8; (b) task timeliness value of instance8; (c) task completion value of instance8.
Electronics 13 02374 g015
Table 1. Satellite orbital parameters and satellite payload parameters.
Table 1. Satellite orbital parameters and satellite payload parameters.
No. Sat a
/(km)
e
/(°)
i
/(°)
R A A N ω
/(°)
φ
/(°)
θ j
/(°)
t max S j
/(s)
A j
(°)
w j
/(°/s)
r j s
/(m)
Sat176200.01495.009159.84994.764109.1535400300.22
Sat274780.08296.582155.90897.034168.1525400300.22
Sat375190.10393.669117.29190.972162.3175400300.22
Sat478730.09395.910112.55892.612119.5626450300.32
Sat577970.06792.125195.81892.860157.3736450350.32
Sat678960.07592.365166.19392.172188.2546450350.31.5
Sat775470.03193.201153.47597.629172.7687500350.31.5
Sat876570.02797.767161.21093.795103.5537500350.41.5
Sat978010.11691.377159.75190.657103.3258500400.41.5
Sat1076220.01496.312107.56394.130122.0728500400.41.5
Table 2. MSDMPUP–TRW, MSIMP–ET, MSIMP–PTR model generated target task importance degree at the initial moment.
Table 2. MSDMPUP–TRW, MSIMP–ET, MSIMP–PTR model generated target task importance degree at the initial moment.
TargetT1T2T3T4T5T6T7T8T9T10T11T12T13T14T15
P M S D M P H M A T R W 9.517.774.967.746.479.805.818.237.255.228.114.455.294.469.28
P M S I M P E T 9.557.584.827.666.279.855.657.997.165.377.904.295.064.288.98
P M S I M P P T R 9.407.644.867.546.539.735.768.147.155.338.364.475.164.379.13
TargetT16T17T18T19T20T21T22T23T24T25T26T27T28T29T30
P M S D M P H M A T R W 9.034.836.544.737.429.695.569.886.145.338.069.787.614.596.40
P M S I M P E T 8.774.656.464.777.359.355.459.746.065.277.959.657.314.386.14
P M S I M P P T R 8.934.766.764.567.459.315.579.656.125.267.949.537.764.616.39
TargetT31T32T33T34T35T36T37T38T39T40T41T42T43T44T45
P M S D M P H M A T R W 5.145.464.499.308.746.368.948.567.716.446.264.626.127.336.85
P M S I M P E T 4.975.164.189.258.486.378.678.257.506.236.164.485.937.116.62
P M S I M P P T R 5.115.524.509.518.906.649.108.687.876.326.444.706.247.476.95
Table 3. The optimal allocation scheme of instance 1.
Table 3. The optimal allocation scheme of instance 1.
TSStEtTSStEtTSStEt
T1S501:45:4001:46:01T16S801:47:3001:47:44T31S900:07:1600:07:29
T2S600:32:6300:40:18T17S305:14:4305:14:54T32S501:44:5201:45:12
T3S800:05:2200:05:36T18S1003:25:4003:25:54T33S901:39:2001:39:34
T4S600:30:0600:30:13T19S404:28:0804:28:15T34S401:01:3701:01:44
T5S306:51:5406:52:04T20S1000:10:4300:10:57T35S600:34:1400:34:23
T6S107:46:1407:46:35T21S1001:42:4601:42:59T36S803:24:2303:24:37
T7S609:39:3909:39:46T22S105:56:1705:56:37T37S401:03:5601:04:03
T8S306:48:1506:48:24T23S501:42:0601:42:27T38------
T9S1000:03:4300:03:57T24S900:10:2900:10:43T39S400:54:4800:55:15
T10S300:17:5700:18:06T25S204:52:5804:53:05T40S701:13:4501:13:58
T11S903:14:5403:15:07T26S609:37:2109:37:28T41S523:53:0923:53:22
T12S501:38:3901:38:59T27S501:48:3901:49:04T42S501:11:0301:11:26
T13S107:54:1307:54:31T28S523:59:3023:59:50T43S107:49:1507:49:35
T14S206:23:5306:24:06T29S101:06:2801:0648T44S207:56:5507:57:03
T15S710:56:0510:56:18T30S306:47:5006:06:49T45S206:17:0806:17:16
Table 4. The experiment results of the HPIM–KTSMOEA algorithm solving different instances under Small-Scale MSDMPUP.
Table 4. The experiment results of the HPIM–KTSMOEA algorithm solving different instances under Small-Scale MSDMPUP.
InstanceInstruction TypeInstruction Input Condition I b e n e f i t R b e n e f i t T r e s p o n s e R c o m p l e t e
Instance1No instructionNo priority38027227179%
Instance2Single instructionTask benefit priority38035623386%
Instance3Single instructionTask timeliness priority38029716381%
Instance4Single instructionTask completion priority38030624095%
Instance5Double instructionTask benefit,
timeliness priority
38033119884%
Instance6Double instructionTask benefit,
completion priority
38033625194%
Instance7Double instructionTask timeliness,
completion priority
38028218992%
Instance8Multiple instructionTask benefit, timeliness,
completion priority
38032021389%
Table 5. The experiment results of the HPIM–KTSMOEA algorithm solving different instances under Large-Scale MSDMPUP.
Table 5. The experiment results of the HPIM–KTSMOEA algorithm solving different instances under Large-Scale MSDMPUP.
InstanceInstruction TypeInstruction Input Condition I b e n e f i t R b e n e f i t T r e s p o n s e R c o m p l e t e
Instance1No instructionNo priority75062833781%
Instance2Single instructionTask benefit priority75074229489%
Instance3Single instructionTask timeliness priority75067221784%
Instance4Single instructionTask completion priority75068230397%
Instance5Double instructionTask benefit,
timeliness priority
75071225787%
Instance6Double instructionTask benefit,
completion priority
75072031695%
Instance7Double instructionTask timeliness,
completion priority
75065624693%
Instance8Multiple instructionTask benefit, timeliness,
completion priority
75070227391%
Table 6. The simulated experiment results of HPIM–KTSMOEA algorithm in different task sizes.
Table 6. The simulated experiment results of HPIM–KTSMOEA algorithm in different task sizes.
N T a s k Instance1Instance2
I b e n e f i t R b e n e f i t T r e s p o n s e R c o m p l e t e I b e n e f i t R b e n e f i t T r e s p o n s e R c o m p l e t e
150105086243280%1050102837888%
2001440116753479%1440138246687%
2501780142561677%1780169153785%
3002190174167876%2190204759584%
N T a s k Instance3Instance4
I b e n e f i t R b e n e f i t T r e s p o n s e R c o m p l e t e I b e n e f i t R b e n e f i t T r e s p o n s e R c o m p l e t e
150105093027283%105094541194%
2001440123833882%1440128150893%
2501780151339080%1780156658591%
3002190182850379%2190189463590%
N T a s k Instance5Instance6
I b e n e f i t R b e n e f i t T r e s p o n s e R c o m p l e t e I b e n e f i t R b e n e f i t T r e s p o n s e R c o m p l e t e
150105098133487%105099741695%
2001440132441286%1440135351394%
2501780161947184%1780165559692%
3002190196053883%2190200465791%
N T a s k Instance7Instance8
I b e n e f i t R b e n e f i t T r e s p o n s e R c o m p l e t e I b e n e f i t R b e n e f i t T r e s p o n s e R c o m p l e t e
150105090232693%105096735490%
2001440122440592%1440131044689%
2501780149546490%1780160250788%
3002190180652189%2190193856987%
Table 7. MSDMPUP simulation results of various instances using different algorithms.
Table 7. MSDMPUP simulation results of various instances using different algorithms.
InstancesHPIM–KTSMOEAKT–MOEABKT–MOEA
R b e n e f i t T r e s p o n s e R c o m p l e t e R b e n e f i t T r e s p o n s e R c o m p l e t e R b e n e f i t T r e s p o n s e R c o m p l e t e
Instance162233782%58337075%56836273%
Instance273229488%68832482%67231680%
Instance366521784%61324177%59923575%
Instance468030397%61833488%60432686%
Instance571125787%66028480%64427778%
Instance671731695%66634887%65033986%
Instance765024693%60827286%59226584%
Instance869827391%64230183%62929482%
Avg68428090%63430982%62030181%
InstancesHKT–MOEABOKT–MOEAHO–MEA
R b e n e f i t T r e s p o n s e R c o m p l e t e R b e n e f i t T r e s p o n s e R c o m p l e t e R b e n e f i t T r e s p o n s e R c o m p l e t e
Instance154235470%59637278%55836072%
Instance263231075%70032884%66231879%
Instance357323072%62724679%60523876%
Instance457731982%63233891%60832786%
Instance561227174%67128882%64327978%
Instance661833281%68335390%65734285%
Instance756726080%61427887%59026984%
Instance859828878%65430685%63029682%
Avg58929577%64731385%61830380%
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Yang, X.; Hu, M.; Huang, G.; Wang, Y. A Hybrid Preference Interaction Mechanism for Multi-Satellite Imaging Dynamic Mission Planning. Electronics 2024, 13, 2374. https://doi.org/10.3390/electronics13122374

AMA Style

Yang X, Hu M, Huang G, Wang Y. A Hybrid Preference Interaction Mechanism for Multi-Satellite Imaging Dynamic Mission Planning. Electronics. 2024; 13(12):2374. https://doi.org/10.3390/electronics13122374

Chicago/Turabian Style

Yang, Xueying, Min Hu, Gang Huang, and Yijun Wang. 2024. "A Hybrid Preference Interaction Mechanism for Multi-Satellite Imaging Dynamic Mission Planning" Electronics 13, no. 12: 2374. https://doi.org/10.3390/electronics13122374

APA Style

Yang, X., Hu, M., Huang, G., & Wang, Y. (2024). A Hybrid Preference Interaction Mechanism for Multi-Satellite Imaging Dynamic Mission Planning. Electronics, 13(12), 2374. https://doi.org/10.3390/electronics13122374

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop