In recent years, researchers have dedicated considerable attention towards operations in close-proximity to uncooperative orbiting artificial objects. Within this context, the onboard reconstruction of the chaser-target relative state vector is a crucial capability for incoming mission scenarios such as formation flying missions (FF), on-orbit servicing demonstrators (OOS), and active debris removal, as well as small bodies exploration [
1]. While these missions are currently in the spotlight of discussions, attaining feasibility still hinges on substantial technological advancements. The necessity for close-proximity manoeuvring introduces a requirement for a guidance, navigation, and control chain to be autonomously solved onboard to ensure timeliness, reactivity, effectiveness, and robustness in both nominal and off-nominal operations. The initial component of this chain is the relative state reconstruction and navigation. This is especially challenging when dealing with artificial uncooperative targets, which calls for a robust solution relying solely on the chaser’s capabilities, as highlighted in [
2,
3]. In this operational context, imaging with passive sensors emerges as the best sensor architecture. A comprehensive review of initial pose determination techniques based on a VIS monocular camera is provided in [
4]. Further, VIS-based optical navigation has been successfully applied within both cooperative and uncooperative rendezvous, as pointed out in [
5,
6]. Nonetheless, the effectiveness of visible imaging is heavily reliant on illumination conditions. Consequently, OOS missions encounter substantial constraints when illumination requirements for VIS image acquisition are integrated into the design and definition of the close proximity operations. Elements such as the target orbit beta angle, attitude history, solar aspect angle induced by the chaser’s fly-around, and the camera axis could significantly jeopardize the ability to detect and track the target appropriately. This limitation may lead to an unacceptable increase in either mission duration or risk. Illumination bottlenecks become particularly pronounced for targets in Low Earth Orbit (LEO) experiencing prolonged eclipses, as highlighted in [
7]. In recent years, the Hayabusa 2 mission successfully exploited its thermal-infrared (TIR) imager for vision-based navigation purposes [
8]. Such an outcome has highlighted the possibility of combining sensors operating in different spectral bands to enhance the navigation solution accuracy. This work aims at exploiting a TIR imager leveraging its insensitivity to illumination conditions, to overcome the limitation imposed by imaging sensors operating in the visible spectrum. Some preliminary work on the topic can be found in [
9], in which the author performs an assessment of the best feature detector and descriptors for thermal infrared images. Nevertheless, TIR sensors are usually characterized by a smaller array size compared to visible ones, thus they have a lower resolution and poorer contrast with respect to VIS sensors, which negatively affects image processing algorithms, as highlighted in [
10]. To overcome these limitation, sensor fusion strategies can play a major role, as pointed out in [
11]. Multispectral data fusion is a dominant technique within the field of Unmanned Aerial Vehicle (UAV) navigation [
12], yet its role remains marginal in the domain of spacecraft relative navigation. The different multispectral data fusion schemes can be divided into two main approaches: image fusion and high-level data fusion, as portrayed in
Figure 1.
Multispectral image fusion aims at creating a new and more informative image type by combining the complementary strengths of the two distinguished spectral bands. The newly obtained image type can then be fed to the subsequent navigation chain to enhanche its robustness to illumination conditions. Image fusion has been successfully applied within the context of remote imaging [
13], while its application in spacecraft navigation scenarios remain marginal. The work presented in [
14] is concerned with the evaluation of the different pixel-level image fusion techniques within the context of spacecraft relative navigation. The validity of image fusion techniques for navigation purposes is assessed in [
15,
16]; where the authors test the effectiveness of pose initialization algorithms on VIS-TIR fused images. A further step is then presented in [
17], in which a Convolutional Neural Network (CNN) based pose estimation algorithm is successfully tested on this new image type. Despite the effectiveness of pixel-level image fusion, this work focuses on high level data fusion. Decision-level multispectral data fusion allows for more flexibility and robustness of the whole navigation chain. In fact, the two source images are processed separately and they are treated as two independent sensors. In this way, the information channels can be treated as redundant, and thus it is easier to isolate faulty measurements or to exclude one sensor when it is no longer providing meaningful information. The work presented in [
18] presents a navigation architecture in which feature tracking is performed simultaneously on both VIS and TIR images, and this information is then fused within a Kalman Filter to achieve robustness in person motion tracking. Subsequently, the idea of performing high-level multispectral data fusion was adopted for relative navigation and mapping of asteroids and unknown spacecraft by [
19,
20], respectively. Further research on asteroid relative navigation is carried out in [
21], in which the authors perform a CNN-based feature map fusion to enhance the accuracy of centroid detection during proximity operations of the HERA mission. In the presented research, the author build on the approach presented in [
22], to propose a flexible navigation strategy to fuse multispectral information. Specifically, the aforementioned work introduces the concept of sensor handover from onecamera to the other to support night-time navigation for a mobile robot. The results presented in the paper suggest that the navigation accuracy and robustness can clearly benefit from the introduction of a thermal infrared imager. However, the handover from the VIS to the TIR camera is controlled by an external user; whereas we implement a fully automatic switch between the two different sensing modalities. Feature detection and feature tracking are performed separately for VIS and TIR images, and the feature points position is fed as observable to an Extended Kalman Filter (EKF). Model-to-image matching is then employed to establish 3D-2D correspondences between the known model and the extracted feature points. Such tightly coupled architecture can be affected in terms of robustness whenever the target’s shape is particularly complex, since it may be more challenging to accurately detect and track a high number of features. However, given the relatively simple shape of the Tango spacecraft and the utilization of two cameras simultaneously, these robustness issues are easily mitigated. Further, we introduce autonomous sensor handover from one sensing modality to the other, in such a way to retain only the most meaningful measurements source. This approach may be extremely useful during eclipses, where the VIS camera is automatically excluded from the navigation chain, since it cannot contribute to the pose estimation task. The presented navigation scheme is applicable to known uncooperative targets, for which at least a simplified geometrical model is available. It is important to remark that the target geometry of an uncooperative space object may not be always known. The most common way of dealing with unknown uncooperative objects is to rely on Simultaneous Localization and Mapping (SLAM) techniques, which enable the chaser to both reconstruct the target’s shape and to perform relative navigation. However, such techniques tend to be numerically heavy, due to the high number of map points which shall be stored in the memory. As proposed in [
23], it can be useful to split the mission into two operative phases: the first one relies on SLAM to gather information about the target’s shape; while in the second phase of the mission the known geometry of the target object is exploited to perform model-based relative navigation. Building on this concept, we can assume to be beyond the preliminary mapping phase, and that at least partial information regarding the target’s shape can be exploited. The major contributions of the paper can be then summarized as follows:
The paper is organized as follows:
Section 2 presents a schematic outline of the developed navigation architecture, highlighting the functionalities of each building block. The Image Processing (IP) functional block is thoroughly described in
Section 3; while
Section 4 details the relative navigation filter. The simulation environment is introduced in
Section 5, while the results are presented and discussed in
Section 6. Conclusion summarizing the main outcomes of our study are reported in
Section 7.