Next Article in Journal
A Real-Time Distance Measurement System for a Digital Twin Using Mixed Reality Goggles
Previous Article in Journal
Detecting Emotions through Electrodermal Activity in Learning Contexts: A Systematic Review
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Simulation of RSO Images for Space Situation Awareness (SSA) Using Parallel Processing

Department of Earth and Space Science, York University, 4700 Keele Street, Toronto, ON M3J 1P3, Canada
*
Author to whom correspondence should be addressed.
Sensors 2021, 21(23), 7868; https://doi.org/10.3390/s21237868
Submission received: 28 September 2021 / Revised: 1 November 2021 / Accepted: 17 November 2021 / Published: 26 November 2021
(This article belongs to the Section Optical Sensors)

Abstract

:
With the rapid increase in resident space objects (RSO), there is a growing demand for their identification and characterization to advance space simulation awareness (SSA) programs. Various AI-based technologies are proposed and demonstrated around the world to effectively and efficiently identify RSOs from ground and space-based observations; however, there remains a challenge in AI training due to the lack of labeled datasets for accurate RSO detection. In this paper, we present an overview of the starfield simulator to generate a realistic representation of images from space-borne imagers. In particular, we focus on low-resolution images such as those taken with a commercial-grade star tracker that contains various RSO in starfield images. The accuracy and computational efficiency of the simulator are compared to the commercial simulator, namely STK-EOIR to demonstrate the performance of the simulator. In comparing over 1000 images from the Fast Auroral Imager (FAI) onboard CASSIOPE satellite, the current simulator generates both stars and RSOs with approximately the same accuracy (compared to the real images) as STK-EOIR and, an order of magnitude faster in computational speed by leveraging parallel processing methodologies.

1. Introduction

Space Situational Awareness (SSA) is becoming an increasingly important issue around the world as the number of resident space objects (RSOs) is continually increasing. SSA requires innovative, robust, and reliable solutions to identify, track, and characterize RSOs. Previously, we demonstrated the feasibility of using low-resolution on-orbit images (such as star tracker images) for RSO detection. In [1], we examined the number and frequency of RSOs that are detectable (given the physical and optical parameters of the objects and imager such as distance, brightness, motion) when a commercial-grade star tracker is used instead of dedicated high-resolution imagers. Using the simulation study [1], we demonstrated that hundreds of objects in low earth orbit can be observed using a star tracker in a day. The starfield simulator—named space-based optical image simulator (SBOIS)—developed for this feasibility study was modeled after FAI imager onboard CASSIOPE spacecraft. Similar simulators; such as the ones outlined in [2,3] use a propagation and ray-tracing software where SBOIS performs all calculations in MATLAB with no additional software required. The simulation represents low-resolution star-tracker-like images invisible to IR range. More information on the preliminary starfield image simulator is also provided in [4]. SBOIS also serves as the training tool to enable artificial intelligence (AI) algorithm design to identify and characterize the RSOs from low-resolution images. AI algorithms have been shown in recent years to perform accurately and efficiently after training, such as [5] for RSO detection and [6] for RSO characterization. Research such as [7,8] highlights the need for high quality and quantity of training data about starfield images and corresponding labels. A labeled starfield database, as produced by SBOIS, enables novel model development for resolving imagery for classification purposes. As detailed in [9] machine learning algorithms adapted to classification and characterization can leverage large dataset variations which are only achievable via tools like SBOIS. Therefore from the perspective of image processing algorithms, the key-value metrics of an image simulator like SBOIS are labeled image generation accuracy, variability in simulation, and image generation time.
While SBOIS provides powerful means to generate and simulate a large number of images, the simulator’s performance was limited in its computation efficiency, as well as, the versatility in providing images from multiple sources. In this paper, we present the next-generation starfield simulator design that features object centroid estimation; implementation of parallel processing for optimal computation; and robustness of input simulation parameters. To validate the accuracy of the simulator, we compared the resulting simulation of the starfield images to the actual images obtained from FAI as well as the simulated images from the Systems Tool Kit (formerly Satellite Tool Kit, STK) a commercially available application.
The accuracy and efficiency of the simulation in starfield images is a critical step in developing SSA algorithms. In [10], the importance of ‘high-fidelity and innovative simulation architecture’ is described to engage SSA mission design. PROXORTM/RT- PROXORTM by Bell Aerospace is one of the few commercially available starfield simulators using multi-thread architecture. STK offers an alternative tool to visualize and analyze stars and RSOs in orbit. For the purpose of this study, we compared the proposed simulator design to STK with electro-optical and infrared sensors (EOIR) toolkit. Originally developed by the Space Dynamics Laboratory for missile defense, STK-EOIR supports radiometric sensor modeling of optical sensors. Combined with STK’s various mission design capabilities, STK-EOIR is a unique platform where optical RSO tracking missions can be planned, simulated, and studied with a graphical user interface for convenience. More details of STK-EOIR features are described in [11]. Ref. [12] presents a study on space object identification using deep neural networks using STK-EOIR where the photometric observations are generated from STK-EOIR with the GEODSS sensor. While STK-EOIR proves to be a versatile platform to simulate starfield images for mission planning and proof-of-concept demonstration, it still lacks the flexibility and efficiency we seek for the current studies applications. We recognize that a commercial product like STK offers a unique capability to perform complex analyses as well as training and research opportunities. AGI reports that STK has more than 50,000 installations at more than 800 global organizations. While limited STK functionalities are available for free, advanced functions still require expensive licensing. Furthermore, for the purpose of the current study, flexibility in low-level implementation is required to accommodate the large set of data and multiple sensor platforms. As such, a custom starfield simulator specifically designed for the RSO identification study was designed and based on the simulator we had developed for the feasibility study. In developing SBOIS, three key parameters were considered to enable RSO tracking algorithms: (1) RSO centroid estimation; (2) implementation of parallel processing; and (3) versatility in simulation parameters.

2. Simulator Architecture

SBOIS was originally designed to examine the feasibility of RSO detection in space. To understand the characteristics of on-orbit images with RSOs, we simulated various scenarios based on commercial grade star trackers currently available in the market. We considered several star trackers, including AD-1 Star Tracker by Mars Bureau [13], and the BOKZ-MF Star tracker by IKI RAN [14]. All models are great candidates for star imaging on a small satellite platform for attitude determination. However, for the purpose of RSO detection, we focused on Fast Auroral Imager (FAI) parameters as the baseline imager, due to a large number of images being publicly available with well-known host satellite positions. The FAI sensor is onboard the CASSIOPE satellite as part of the ePOP (Enhanced Polar Outflow Probe) science mission. The purpose of FAI is to measure large-scale auroral emissions. It measures auroral emissions with near-infrared wavelengths (650–1100 nm) and a monochromatic wavelength of 630 nm. The near-infrared range was chosen instead of the chromatic wavelength as it is a more realistic representation of low-resolution images like what would be found from star tracker images. The parameters for FAI are shown in Table 1. Images from FAI are publicly available on the ePOP website [15]. FAI sensor has an effective f-number of f/0.8, an effective focal length of 13.8 mm. SBOIS simulates the optical environment from various sources of direct and reflected flux from sources that include; the starfield, Earth, Earth Limb, Moon, Sun, and Zodiac. To combine these different flux sources, as well as, the RSO/s into one image the architecture in Figure 1 is used. SBOIS uses the common assumption that the RSOs are not active sources of flux but rather, reflectors of solar radiation, which is common in recently published RSO models [16,17,18]. To account for the active sources of flux, they are considered either as point sources or area sources with different calculations performed for each. Point sources include the Sun, and stars that are calculated from their known magnitude. The equations to calculate the given magnitude to the Digital Number (DN) values are shown in Equations (1) and (2) [18].
EObj = Ezero × 10−0.04mobj
DNx,y = Eobj × eDN × QE × τ × Aapp × tint × PSF(x, y, xobj, yobj)
In Equations (1) and (2): EObj represents the photon irradiance from the objects in photons per second per meter squared ph/s/m2. Ezero represents the zero-point photo irradiance found from integrating the solar spectrum over the effective band, in ph/s/m2. mobj represents the absolute magnitude of the object. DNx,y represents the Direct Number (DN) value at pixel x and y on the sensor. eDN represents the ratio of photons received to the electrons given by the sensor properties, the units are electrons per DN, e/DN. QE represents the quantum efficiency of the sensor system in the band used. τ represents the optical feed loss of the sensor. Aapp represents the area of the aperture given in m2. tint represents the integration time of the sensor in seconds. PSF(x, y, xobj, yobj) represents the point spread function on pixel x, y for a source that is located at xobj, yobj. For more information on point spread functions and the function implemented please see [1]. Note that the Sun is never included in the image due to the Sun exclusion angle.
For area sources such as Earth, Earth’s limb, the Moon, Moonglow, and zodiac glow are calculated from their surface brightness. The equations to transform the given magnitude to the DN values are shown in Equations (3)–(5).
L o b j = E z e r o × 10 0.04 m a s c ,   o b j ( 180 π ) 2 3600 2
E o b j = L o b j × I F O V 2
D N p p = E o b j × e D N × Q E × τ × A a p p × t i n t
In Equations (3)–(5): Lobj represents the photon radiance given in units of photons per second per meter squared per steradian ph/s/m2/sr. Ezero represents the zero-point photo irradiance found from integrating the solar spectrum over the effective band, in units of ph/s/m2. masc, obj represents the magnitude per square arc-second of the surface object given in magnitude per arc-second squared. EObj represents the photon irradiance from the objects in photons per second per meter squared, ph/s/m2. IFOV represents the instantaneous field of view which is found by dividing the pixel size by the focal length. DNpp represents the DN values per pixel across the sensor. eDN represents photoelectron the DN ratio given in electrons per DN, e/DN. QE represents the quantum efficiency of the sensor system in the band being used. τ represents the optical feed loss of the sensor. Aapp represents the area of the aperture given in units meters squared m2. tint represents the integration time of the sensor in seconds. Note that Equation (3) represents the conversion from arcseconds squared to degrees squared, with ( 180 π ) 2 representing the conversion from degrees squared to steradians.
To calculate the brightness of the RSO’s in the image, a Bidirectional Reflectance Distribution Function (BRDF) is used with the incoming solar flux. To perform the BRDF modeling of RSO’s, we assume that satellites consist of several well-defined surfaces called facets. These facets are surfaces with well-known BRDF’s. Complex 3-dimensional shapes consist of these facets transforming the complex and unknown BRDF function of RSO’s into a sum of its known facets. In the current study, all RSO’s are assumed to be perfectly spherical objects with a 0.2 reflectance value, as this is the simplest case to determine detections. Equations (6)–(11) show the calculation of DN values from the spherical facet model of RSO’s. More details on the BRDF and facet modeling of SBOIS are described in [19].
ρ A F = ρ A ( B d i f f F d i f f + B s p e c F s p e c )
F d i f f , S p h e r i c a l = 2 3 π 2 [ ( π Φ ) cos ( Φ ) + sin ( Φ ) ]
F s p e c , S p h e r i c a l = 1 4 π 2
m o b j = m s u n 2.5 log 10 ( ρ A F a 2   ) + 5 log 10 ( d r s o , s u n × d r s o , s e n s o r a 2   )
E O b j = E z e r o × 10 .04 m o b j
D n x , y = E o b j × e D N × Q E × τ × A a p p × t i n t × P S F ( x , y , x o b j , y o b j )
In Equations (6)–(11), ρAF represents the reflective area phase function which is an intermittent step on calculating the magnitude of the RSO. ρ represents the reflectivity of the object. A represents the effective area of the RSO in meters squared. Bdi f f and Bspec represent the diffuse and specular Hejduk mixing coefficients, respectively [17]. Fdi f f and Fspec represent the diffuse and specular phase function, respectively. Φ represents the solar phase angle of the Sun, RSO, sensor system. mobj represents the magnitude of the RSO. msun represents the magnitude of the sun, which is constant at −26.73. a represents one astronomical unit, the distance from the Sun to Earth. dx,y represents the distance from object x to object y. The rest of the variables are the same as in Equations (1) and (2) which are described above.
After all the sources of flux are included in the image various noise are considered and added including, shot noise, dark current, read noise, multi-pixel radiation events, and the point spread function. For more information on the noise modeling of SBOIS164 see [4].
The final stage of the simulator combines all the flux sources and noise to create an image with the labeled information outputted as auxiliary data with the image. One powerful feature of SBOIS is the ability to simulate different types of RSO’s, such as active and inactive, that have different types of auxiliary data available. This allows for robustness in the inputs allowing accurate ephemeris data to be entered for satellites with well-known positions, or TLE propagation for RSO’s with less accurately known orbits. Robustness in algorithms generally leads to a decrease in either computation speed or accuracy. By leveraging parallel processing capabilities in the input, RSO propagation, and image generation, the overall computation speed can be reduced while keeping the robustness of the inputs. One of the challenges in developing a robust RSO image simulator is the different types of active and inactive RSO’s, which contain fidelity of auxiliary data. SBOIS is designed to generate a robust image simulator, accepting multiple types of data inputs, as well as, being more efficient than currently available image simulators by leveraging parallel processing capabilities.
The robustness of the inputs was designed with the assumption that different types of observations and RSO’s have different levels of information available. Active satellites would have the most information available, normally having ephemeris data, attitude, and general shape information available. The ephemeris data allows for better positioning of the target RSO compared with propagation methods. The attitude and general shape allow for a more accurate estimate of the facet model of the satellite allowing the light reflected off the RSO to be simulated with more accuracy. This information provides the most accurate simulation with SBOIS but is not always available for all RSO’s. For deactivated satellites and rocket bodies, ephemeris data and attitude are not always available. When the ephemeris data is not available a TLE and PSGP4, mentioned in more detail below, are used to determine the position of the RSO. With no attitude information, and in the case of debris where the shape is sometimes not known, the RSO’s are simulated as a perfectly reflective sphere. This is commonly done in different RSO simulators to estimate the effective light coming off the object which allows for the training and evaluation of detection algorithms [5], as well as detection prediction [1]. The robustness of the outputs was also designed to provide any information that the user requires as auxiliary data to allow for easy training of machine learning algorithms. These outputs include; coordinate rotation matrices, object brightness, labeled star location, and most importantly labeled RSO information. The labeled RSO information gives the RSO’s that are visible in the image with their position, relative motion, relative magnitude, and signal to noise ratio. This allows for the SBOIS simulated images to be used in the training of the detection algorithms without the need for any external labeling code/software.

3. Parallel Processing and PSGP4 Architecture

Image simulation is a computationally intensive task, this is only increased with the simulation of thousands of moving RSO’s possibly in the image, background zodiac light, moon glow, etc. The impact of this is requiring seconds to minutes to generate one simulated image using standard sequential computation methods. To allow for large data sets (thousands to millions of images) to be generated in a reasonable time, the computation time per image needs to be reduced when looking at commercially available simulators such as STK’s EOIR. To accomplish this parallel processing in the form of Single Instruction Multiple Data (SIMD) parallel processing is introduced. SIMD leverage the improvements in computer hardware allowing a single instruction to be performed on multiple sets of data at once, reducing the computation time required to perform the same calculation on multiple data points. For image simulation, this is commonly used with ray-tracing algorithms performing the same calculation on thousands to millions of rays. The SBOIS uses a forward ray tracking method with the different sources of light mentioned above. The implementation of SIMD allows for the light received from different sources, as well as, all the light reflected off of RSO’s to be performed at once greatly reducing the computation time required. SIMD cannot be implemented on every part of SBOIS, though it was able to be implemented on the most computationally intensive steps which include; coordinate transformations, received flux, propagation, and assigning values to pixels.
The simulation of RSO images requires the known position of the RSO’s at the given time an image is taken. As for most RSO’s ephemeris data is not available TLE’s are used to propagate an RSO to approximately its position in the image. This has many uses such as estimating the distance to the RSO, seeing if the RSO is in the Field of View (FOV), and determining the pixel location of the RSO if it is in the FOV. Propagation is one of the more computationally intensive portions of simulating RSO images, this scales with the amount of RSO’s propagated. For an image where the RSO is not known propagation of an entire catalog is required to determine what RSO is in the image. Currently, there are approximately 20,000 tracked objects in the NORAD Satellite Catalog, with that number expected to jump an order of magnitude over the next decade [19]. Due to the proliferation of tracked RSO’s more efficient methods for satellite propagation are required to generate large datasets efficiently. This need has led to the creation of the Parallel Specialized Perturbation Method 4 (PSGP4) Propagator.
PSGP4 is a version of the SGP4 propagator that is set up to propagate RSO’s in parallel. Due to the propagation of RSO’s following almost the same architecture and having no correlation between different RSO’s, this allows for the implementation of SIMD parallel processing. This is shown from the algorithm block diagram, Figure 2. While there are some sections where SIMD is not possible, such as the solving of Kepler’s equation which is performed iteratively, it is implemented in a significant portion to see improvements in computation time without a reduction in accuracy. This can be seen from the implementation of PSGP4 for maneuver detection in Parallel Propagation for Orbital Maneuver Detection [20]. Since the paper was published in 2020 the PSGP4 algorithm has been updated to include the Specialized Deep Space Perturbation Method 4 (SDP4). The difference between SGP4 and SDP4 is that SDP4 is used for objects with an orbital period above 225 min with SGP4 being for objects with periods below 225 min. The SDP4 algorithm follows the same architecture as the SGP4 algorithm with some added functions and different coefficients, this allows SDP4 to be integrated with a set of sub-functions which are represented on the left side of Figure 2.

4. Simulator Performance Metrics

4.1. RSO Pixel Position Comparison

To validate the accuracy of the SBOIS to a readily available commercial product such as STK-EOIR, a comparison study using the identical orbital parameters in both simulators was conducted. Images taken by FAI on 15 June 2019, 23:35:50 to 23:36:15 were simulated. This time period was chosen to capture the first sighting of the RADARSAT Constellation Mission (RCM) spacecraft after they were launched on 12 June 2019. Their distinct formation of three spacecraft, combined with other reference stars within FAI’s field of view help not only in determining the pixel locations but also in identifying and correcting for potential lens distortions since it forms a geometrical shape rather than a single data point. Three stars from the Bright Star V5 catalog—#6111, #6158, and #6195—were used as references to prevent errors caused by discrepancies in the simulation of the attitude of the observer spacecraft CASSIOPE. They were chosen for their unique in-line formation that is almost parallel to the RCM formation during the observation period used in this study.
Figure 2 illustrates a set of images from FAI (left), SBOIS (center), and STK-EOIR (right). Highlighted in the red ellipses are the RCM constellation imaged on 15 June. Yellow ellipses depict the locations of three reference stars (Bright Star V5 #6111, #6158, #6195). The RCM, a three-satellite constellation providing daily revisits of Canadian land and ocean, was launched three days prior to the study period; thus, not fully deployed and still retains its distinctive formation. The formation remains visible within FAI’s field of view for 32 s, with a total of 31 images taken.
Table 2 illustrates sample data with three satellite locations with corresponding errors in pixels defined as the difference between FAI images and simulated images from SBOIS and STK-EOIR, respectively. More examples of the pixel locations and errors are listed in Appendix A.
The comparison results are very promising. Pixel accuracy for RCM satellites are mostly within 5 pixels for SBOIS. This error can be attributed to slight position and attitude discrepancies between the real host satellite and the host within the simulator. Observing satellite (CASSIOPE satellite that hosts FAI) ephemeris, for example, plays a critical role in presenting RSO orbit in simulated images. The error of RSO position, as seen from a CCD, is minimal compared to the effect of host satellite position. While the shape and size of the objects affect the brightness of the object, the position of the RSO in the starfield is mostly defined by the accuracy of the RSO position, observer satellite position, and attitude of the observer satellite. RSO attitude and facets (surfaces or shape of the RSO) contribute to the light curve in the form of brightness variation; however, they do not affect the position of the RSO on the simulated images. For the purpose of this study, it was assumed that all space objects are 10-m diameter with a perfect sphere for visibility on the simulated image. The spheres are estimated to have 0.2-reflectance across FAI’s effective spectral band. Statistics, such as light reflectance of materials used, the physical size of the satellite, solar panel placement, and orientation could also affect the centroid of the reflected light perceived by another sensor; but only on a subpixel level. The comparison study between SBOIS and STK-EOIR did not consider the above-listed physical parameters. Instead, all objects were assumed to be of the same size for comparison and to eliminate potential errors which may occur from the continuous attitude changes. STK’s SGP4 propagator was used for the RSO positioning to keep the propagation method the same between SBOIS and STK.
The pixel locations of RCM satellites and reference stars, as shown in Figure 3, were compared by extracting their pixel location from both simulators; then, by calculating differences in the X- and Y-axes. The data from SBOIS are extracted directly from the simulator, as there is a built-in function to the output pixel location. STK, on the other hand, has no such functions to export a file with pixel locations of objects within the field of view of a sensor. It does have a details window which can be viewed via EOIR synthetic scene window, listing the information within each pixel as the user selects it. The information includes X and Y coordinates, objects within the pixel, and object distance from the sensor. The lack of batch-export ability in STK significantly hinders any future analysis if the user does not know the general location of the target at the time period, and likely have to cross-reference with the 3D scenario window in STK to determine the pixels the objects are in before analyzing them.
The reference stars are compared in a similar manner. Since both simulators use the same star catalog (Bright Star V5), significant differences between the two simulators’ outputs were not expected nor observed. It is worth noting that on rare occasions—at 23:36:15, for example—the scale of the star formation seems to be altered slightly (pixel differences between first and last star is different on both simulators), with STK showing a formation that is slightly larger than that of SBOIS. Also, noteworthy that in the last 2 images, as the satellites move towards the edge of the image, the lens distortion effect of the FAI sensor caused the image to warp, decreasing the accuracy of both simulators compared to the real image, as seen in the comparison shown in Appendix A.
As stated in the RCM analysis, SBOIS images depict stars with 6.6 and RSO’s with 5.1-pixel accuracy, comparable to the images generated from STK (mean difference of 6.5 compared to original FAI image. Note that we reported accuracy of 9.4 pixels using the original simulator without the sub-pixel feature [21]. As the simulator was used only to examine the feasibility of RSO identification using low-resolution, the reported pixel accuracy was considered acceptable. With the improvement made since then, we now represent on-orbit images with better than 5.1-pixel accuracy, making the simulator pertinent for various applications. In the current study, we focus on the validity of the simulator in creating a large dataset for AI algorithm training where accuracy and variance of the simulated data is key characteristic. Regardless of the approach, we take in developing AI for this purpose, simulated data needs to represent the real dataset with as much accuracy and flexibility as we can afford.
It is also noted that in comparison with SBOIS, STK-EOIR does not have a built-in sub-pixel output option; if one needs to identify the exact location of an object, it can only be done through a third-party algorithm that calculates the centroid of a multi-pixel shape. This may lead to some albeit small, but frequent inaccuracies in future studies.

4.2. Formatting of Mathematical Components

Currently, there are over 20,000 in-orbit objects such as satellites, debris, rockets, revolving around Earth in various orbits, and this number is expected to rise by about an order in magnitude with the newly constructed US space fence coming operational [22]. To account for the large scale of propagation required a Parallel SGP4 method (PSGP4) as outlined in [20] was implemented. In addition, the Bright Star catalog adds around 10,000 objects in the form of stars into the scenario. Combining all the objects in any simulation is a challenging task, as well as the synthetic scene generation required to simulate the field of view and images of a sensor installed on a spacecraft. To successfully complete given tasks, a simulator needs to be capable of simulating the propagation of a great amount of RSOs, while taking the Bright Star Catalog, or any chosen catalog, into account when generating a simulated scene for a sensor; this should be done in a way that does not require significant computation time. This requires the simulator to be optimized, robust, flexible, and customizable. We compare SBOIS and STK-EOIR in the above-listed aspect, with emphasis on computation speed, since SBOIS utilizes such methods in its computation. It is worth noting that STK has the advantage of having a graphical user interface (GUI) and a visual 3D scenario, but these can take valuable computational resources away from the actual simulation; this can be a burden if the user does not require the visualization nor the GUI.
The comparison study was designed with both simulators generating 200 and 1000 images, then comparing their processing time. These results can be seen in Table 3. The results from SBOIS were collected via an output function built into the simulator akin to other auxiliary data, which gives them time to generate the specified time period. The initialization time for SBOIS is between 0.13 to 0.15 s—insignificant for this comparison. Results from STK were taken by subtracting the image creation time of the 1st image from the last image generated by checking the option "auto-generate scene bitmap", which gives the in-between time of the two images. It is worth noting that the size of the EOIR window affects STK calculation speed, but closing it will also stop the generation of images. Therefore, for the following comparisons, all windows were adjusted to the smallest dimension possible to achieve the best results.
The data above compares the computational speed of STK12.0.1 and its compatible EOIR to SBOIS in terms of total time to generate a specified number of images for each comparison category. From the comparison, it is clear that SBOIS is superior to STK in the processing time required under the circumstances that are tested. Apart from the total time needed, the average time taken for each image was also compared. The time needed for SBOIS increases alongside the number of images within a set. the average time per image for a 200-image sequence was 0.34 s and decreased to 0.28 s when the number of images was increased to 1000. STK takes significantly longer than SBOIS to produce the images in both cases with the images produced not being automatically labeled. This implies that SBOIS is more computationally efficient for generating large labeled datasets.
As mentioned earlier, the 3D visualization capability and the GUI of STK help the user greatly in many scenarios but can slow down the program significantly if a large number of objects are active at the same time. To test the limitations of STK, the entire GPS constellation and some independent satellites were imported into a scenario. STK crashed several times during the import process after there are over 60 objects in the scenario. Moreover, the import process can be time-consuming for the user if TLE or other satellite files are not readily available. On average, it takes 1.5 to 2 s to import an object from the AGI database via GUI. Importing all objects within SBOIS into an STK scenario can be extremely time-consuming. Although object import time can be reduced significantly when considering the best-case scenario when the user has satellite orbit files ready, the possibility of STK crashing during the import is still present.

5. RSO Detection Algorithms Using Simulated Images

Detection algorithms are out of the scope of this paper, with the goal just to provide simulated data for the training and verification of these algorithms. Here some of the different types of RSO detection algorithms are mentioned and how simulated images are used with them. Firstly, RSO detection algorithms are used to determine if an RSO is in the image, as well as, distinguish between stars and RSO’s. The main way of determining the difference between stars and RSO’s is from the motion that they produce in the image, with stars streaking due to the motion of the camera and host platform. The streaking of the RSO’s is due to the motion of the camera, host platform, and RSO’s motion. When accounting for the sensor and platform motion the only objects that should be moving are the RSO, allowing for determination between the stars and RSO’s. For more information on different detection and centroiding methods, such as machine learning and analytical algorithms see [21,23,24]. For the training of machine learning detection algorithms, hundreds to thousands of images and image sequences are required, which are not always available for space-based platforms. To make up for the lack of data simulated images can be used to train these algorithms. One example of detection algorithms being trained of SBOIS simulated images is presented in Dave’s AMOS technical presentation [5].

6. Conclusions

In this paper, we described the architecture of the custom design starfield simulator, SBOIS, and compared its performance with the real images as well as the simulated images generated using a commercial product, STK-EOIR. In comparing FAI images taken on 15 June 2019, 23:35:50 to 23:36:15 where a distinct RADARSAT Constellation Mission (RCM) formation was observed, an average of approximately 5-pixel accuracy was observed in both simulators, demonstrating very similar performance in generating realistic starfield images. SBIOS, however, features an annotation function that each image generated has labeled data for the simulated objects (both RSOs and stars) with position, velocity, pixel centroid, and SNR. In comparing the processing time, SBOIS outperforms for 200 and 1000-image cases with less than 1/3 of the processing time to generate the images.
As part of future algorithm development, we are extracting light curve information of RSOs to estimate the target range and improve the accuracy of the temporal classifier with further characterization parameters. This will be similar to the light curve inversion methods outlined in [6,25,26,27] to estimate an unresolved RSO’s shape, attitude, and optical properties. An additional area of research is the point spread function modeling of sensors using convolution filters from the proposed classification algorithm. The algorithm proposed for PSF modeling would improve the accuracy of the simulator in recreating images and reduce centroiding and classification for future sensor simulations. Performing a similar analysis on a varying range of image datasets from ground and space, from varying sensor types and resolutions could also widen the application of the simulator.

Author Contributions

A brief sentence giving each authors contribution follows. S.D. contributed by providing centroiding algorithm, writing the machine learning section, and helped identify the real images to use. Y.F. contributed by generating the STK EOIR simulated images, performing timing and position comparison, and writing and editing the paper. R.L. contributed by supervising, writing and editing the paper, and securing funding for the research. R.C. contributed SBOIS simulator and simulated images, data analysis and results finding, as well as, the writing and editing of the paper. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by Natural Sciences and Engineering Research Council of Canada (NSERC) grant number RGPIN-2019-06322; and Canadian Space Agency (CSA) grant number 19FAYORA12. And The APC was funded by Canadian Space Agency (CSA) grant number 19FAYORA12.

Institutional Review Board Statement

Not Applicable.

Informed Consent Statement

Not Applicable.

Data Availability Statement

The data presented in this study are available on request from the corresponding author. The data is not publicly available due to continuing research and containing sensitive data.

Conflicts of Interest

The authors declare no conflict of interest.

Appendix A

Table A1. Accuracy Analysis of Simulated Images Using SBOIS and STK-EOIR.
Table A1. Accuracy Analysis of Simulated Images Using SBOIS and STK-EOIR.
ObjectsSBOIS STK-EOIR SBOIS ErrorSTK Error
Row(X)Col(Y)Row(X)Col(Y)
15 June 2019 23:35:50
Ref 6111143.0151.2143.6151.0(1.1, −8.6)(1.7, −8.8)
Ref 6158121.6161.3120.6160.7(−4.4, −5)(−5.4, −5.6)
Ref 6195101.1170.2100170.0(−1.5, −4.8)(−2.6, −5)
RCM-122.6155.013.7152.6(3.8, 5.1)(−5.1, 2.7)
RCM-235.5147.830.0144.6(1.3, 4.8)(−4.2, 1.6)
RCM-349.9139.645.4136.6(−0.3, 4)(−4.8, 1)
15 June 2019 23:35:55
Ref 6111143.8152.3144.5150.8(1.3, −6.7)(2, −8.2)
Ref 6158122.5162.5122.6161.6(−4.2, −3.5)(−4.1, −4.4)
Ref 6195102.0171.5101.7171.7(−0.6, −3.5)(−0.9, −3.3)
RCM-155.5139.550.0140.6(1.1, 4.5)(3.9, 5.6)
RCM-2--65.6131.6(2.9, 4.2)
RCM-382.9123.979.6123.0(−1.6, 3.7)(4, 2.8)
15 June 2019 23:36:00
Ref 6111144.7153.5144.7153.7(2.5, −5.6)(2.5, −5.4)
Ref 6158123.3163.7122.0163.4(−3, −2.4)(−4.3, −2.7)
Ref 6195102.9171.8100.6172.1(0.2, −3.4)(−2.1, −3.1)
RCM-188.7123.785.0121.7(1.6, 3.9)(2.1, 1.9)
RCM-2101.5116.4100.6113.6(−1, 3.7)(8.1, 1)
RCM-3115.7108.6114.6106.0(−1.5, 3.4)(−2.6, 0.8)
15 June 2019 23:36:05
Ref 6111144.2151.8144.0151.6(1.7, −6.2)(1.5, −6.4)
Ref 6158122.9162.2122.0162.6(−4.3, −3.1)(−5.2, −2.7)
Ref 6195102.6171.4101.8171.7(−0.7, −3.8)(−1.5, −3.5)
RCM-1120.3105.1116.0104.8(−0.4, 1.7)(−4.7, 1.3)
RCM-2132.797.8130.696.0(−2.5, 1.4)(−4.6, 0.4)
RCM-3146.589.6144.087.7(−3.6, 1.9)(−6.1, 0)
15 June 2019 23:36:10
Ref 6111143.1148.6143.6148.4(0.2, −9.2)(0.7, −9.4)
Ref 6158121.9159.2121.6159.6(−4.1, −6.6)(−4.4, −6.2)
Ref 6195101.7168.6100.8169.6(−1.2, −6.4)(−2.1, −5.4)
RCM-1150.384.8147.783.6(−2.8, −2.4)(−5.4, −3.6)
RCM-2161.277.6161.674.7(−6.6, −2.9)(−6.2, −5.9)
RCM-3175.369.5174.665.4(−6.5, −3.3)(−7.2, −7.4)
15 June 2019 23:36:15
Ref 6111142.0145.3142.0145.4(−0.4, −11.8)(−0.4, −11.7)
Ref 6158120.9156.2119.7157.3(−6.6, −10.2)(−7.8, −9.1)
Ref 6195100.8165.898.6167.3(−2.1, −9.4)(−4.3, −7.9)
RCM-1179.064.5178.662.4(−7.1, −7.1)(−7.5, −9.2)
RCM-2190.457.5192.353.6(−7.1, −6.1)(−5.2, −10)
RCM-3202.649.6205.045.5(−11.6, −7.3)(−9.2, −11.5)

References

  1. Clemens, S.; Lee, R.; Harrison, P.; Soh, W. Feasibility of Using Commercial Star Trackers for On-Orbit Resident Space Object Detection. In Proceedings of the Advanced Maui Optical and Space Surveillance Technologies Conference, Maui, HI, USA, 11–14 September 2018. [Google Scholar]
  2. Zhang, W.; Jiang, Z.; Zhang, H.; Luo, J. Optical Image Simulation System for Space Surveillance. In Proceedings of the Seventh International Conference on Image and Graphics, Qingdao, China, 26–28 July 2013; pp. 721–726. [Google Scholar]
  3. Haherty, S.P.; Ellis, H.B., Jr. A High Fidelity Approach to Data Simulation for Space Situational Awareness Missions. In Proceedings of the Advanced Maui Optical and Space Surveillance Technologies Conference, Maui, HI, USA, 20–23 September 2016; p. 57. [Google Scholar]
  4. Clemens, S. On-Orbit Resident Space Object (RSO) Detection Using Commercial. Master’s Thesis, York University, Toronto, ON, Canada, 2019. [Google Scholar]
  5. Dave, S.; Clark, R.; Chianelli, G.; Lee, R. Machine Learning Implementation for In-Orbit RSO Orbit Estimation Using Star Tracker Cameras. In Proceedings of the Advanced Maui Optical and Space Surveillance Technologies Conference, Maui, HI, USA, 15–18 September 2020. [Google Scholar]
  6. Linares, R.; Crassidis, J.L.; Jah, M.K. Particle Filtering Light Curve Based Attitude Estimation for Non-Resolved Space Objects. In Proceedings of the 24th AAS/AIAA Space Flight Mechanics Meeting, Santa Fe, NM, USA, 26–30 January 2014; p. 210. [Google Scholar]
  7. Varela, L.; Boucheron, L.; Malone, N.; Spurlock, N. Streak detection in wide field of view images using Convolutional Neural Networks (CNNs). In Proceedings of the Advanced Maui Optical and Space Surveillance Technologies Conference, Maui, HI, USA, 17–20 September 2019; p. 89. [Google Scholar]
  8. McQuaid, I. Autonomous Association of Geo RSO Observations Using Deep Neural Networks; Air Force Institude of Technology: Dayton, OH, USA, 2018. [Google Scholar]
  9. Liu, B.; Yao, L.; Han, D. Harnessing ontology and machine learning for RSO classification. SpringerPlus 2016, 5, 1655. [Google Scholar] [CrossRef] [Green Version]
  10. Hagerty, S.; Ansari, K. An Innovative Near-Real-Time Approach to Space SItualtional Awareness (SSA) Data Simulation. In Proceedings of the 35th Space Symposium, Technical Track, Colorado Springs, CO, USA, 8–11 April 2019. [Google Scholar]
  11. McQuaid, I.; Merkle, L.; Borghetti, B.; Cobb, R.; Fletcher, J. Space object identification using deep neural networks. In Proceedings of the Advanced Maui Optical and Space Surveilance Technologies Conference, Maui, HI, USA, 11–14 September 2018. [Google Scholar]
  12. Zakharov, A.I.; Prokhorov, M.E.; Tuchin, M.S.; Zhukov, A.O. Minimum Star Tracker Specification Required to Achieve a Given Attitude Accuracy. Astrophys. Bull. 2013, 68, 481–493. [Google Scholar] [CrossRef]
  13. Lobanov, V.S.; Tarasenko, N.V.; Zboroshenko, V.N. Orientation and Stabilization Systems of Space Vehicles for Different Purposes: Lones of Development. Gyroscopy Navig. 2016, 7, 50–57. [Google Scholar] [CrossRef]
  14. Cogger, L.; Howarth, A.; Yau, A.; White, A. Fast Auroral Imager (FAI) for the e-POP Mission. Space Sci. Rev. 2014, 189, 15–25. [Google Scholar] [CrossRef]
  15. Curti, F.; Spiller, D.; Schiattarella, V.; Orsi, R. Recognition of Orbiting-Objects Through Optical Measurements of Light-Reflecting-Targets by Using Star Sensors. In Proceedings of the IAA Conference on Space Situational Awareness, Orlando, FL, USA, 13–15 November 2017. [Google Scholar]
  16. Hejduk, M.D. Specular and Diffuse Components in Spherical Satellite Photometric Modeling. In Proceedings of the Advanced Maui Optical and Space Surveillance Technologies (AMOS) Conference, Maui, HI, USA, 13–16 September 2011. [Google Scholar]
  17. Fulcoly, D.O.; Kalamaroff, K.I.; Chun, F. Determining Basic Satellite Shape from Photometric Light Curves. J. Spacecr. Rocket. 2012, 49, 76–82. [Google Scholar] [CrossRef]
  18. Shell, J.R. Optimizing Orbital Debris Monitoring with Optical Telescope. In Proceedings of the Advanced Maui Optical and Space Surveillance Technologies (AMOS) Conference, Maui, HI, USA, 14–17 September 2010. [Google Scholar]
  19. Pechkis, D.L.; Pacheco, N.S.; Botting, T.W. Statistical Approach to the Operational Testing of Space Fence; IEEE Aerospace and Electronic Systems Magazine: Piscataway, NJ, USA, 2016. [Google Scholar]
  20. Clark, R.; Lee, R. Parallel Processing for Orbital Maneuver Detection. Adv. Space Res. 2020, 66, 444–449. [Google Scholar] [CrossRef]
  21. Jia, B.; Pham, K.D.; Blasch, E.; Wang, Z.; Shen, D.; Chen, G. Space Object Classification using Deep Neural Networks. In Proceedings of the IEEE Aerospace Conference, Big Sky, MT, USA, 3–10 March 2018. [Google Scholar]
  22. Haimerl, J.A.; Fonder, G.P. Space Fence System Overview. In Proceedings of the AMOS, Maui, HI, USA, 15–18 September 2015. [Google Scholar]
  23. Samed, A.L.; Karagoz, I.; Dogan, A. An Improved Star Detection Algorithm Using a Combination of Statistical and Morphological Image Processing Techniques. In Proceedings of the 2018 IEEE Applied Imagery Pattern Recognition Workshop, Washington, DC, USA, 9–11 October 2018. [Google Scholar]
  24. Fitzmaurice, J.; Bedard, D.; Lee, C.H.; Seitzer, P. Detection and correlation of geosynchronous objects in NASA’s Wide-field Infrared Survey Explorer images. Acta Astronaut. 2021, 183, 176–198. [Google Scholar] [CrossRef]
  25. Bradley, B.K.; Axelrad, P. Lightcurve Inversion for Shape Estimation of GEO Objects from Space-Based Sensors; University of Colorado, International Space Symposium for Flight Dynamicsm: Boulder, CO, USA, 2015. [Google Scholar]
  26. Gasdia, F. Optical Tracking and Spectral Characterization of Cubesats for Operational Missions; Embry-Riddle Aeronautical University: Daytona Beach, FL, USA, 2016. [Google Scholar]
  27. Clark, R.; Dave, S.; Wawrow, J.; Lee, R. Performance of Parameterization Algorithms for Resident Space Object (RSO) Attitude Estimates. In Proceedings of the Advanced Maui Optical and Space Surveillance Technologies, Maui, HI, USA, 15–18 September 2020. [Google Scholar]
Figure 1. Space-based optical image simulator (SBIOS) system architecture is in green and purple. The image generation sequence for steps 2 through 6 are shown in blue. noisesources include: read-out noise, shot noise, and the point spread function; pseudoimage refers to the simulated image that contains the single light source, namely Earth glow, moon, zodiac glow, and RSO.
Figure 1. Space-based optical image simulator (SBIOS) system architecture is in green and purple. The image generation sequence for steps 2 through 6 are shown in blue. noisesources include: read-out noise, shot noise, and the point spread function; pseudoimage refers to the simulated image that contains the single light source, namely Earth glow, moon, zodiac glow, and RSO.
Sensors 21 07868 g001
Figure 2. PSGP4 Algorithm Block Diagram, where the color represents the different implementations of SIMD processing. On the left of the main flow, the additional PSDP4 functions are shown with the location of the implementation.
Figure 2. PSGP4 Algorithm Block Diagram, where the color represents the different implementations of SIMD processing. On the left of the main flow, the additional PSDP4 functions are shown with the location of the implementation.
Sensors 21 07868 g002
Figure 3. Sample Images from: (a) FAI on-orbit observation [15]; (b) Simulated using SBOIS (c) Simulated using STK-EOIR.
Figure 3. Sample Images from: (a) FAI on-orbit observation [15]; (b) Simulated using SBOIS (c) Simulated using STK-EOIR.
Sensors 21 07868 g003
Table 1. Fast Aurora Imager Parameters [11].
Table 1. Fast Aurora Imager Parameters [11].
ParameterValue
Pixels256 × 256
Quantum Efficiency, QE0.66
Optical Transmittance Loss, τ0.9
Aperture Diameter1.7 cm *
Pixel Size26 µm
Effective Focal Length1.38 cm *
Integration Time0.1 s
* The focal length of the FAI sensor is 68.9 mm and it has an f-number of f/4. This gives it an aperture of 17 mm.
Table 2. Summary of Simulator Accuracy.
Table 2. Summary of Simulator Accuracy.
ObjectsSBOISSTK-EOIRSBOISESTK
Row(X)Col(Y)Row(X)Col(Y)ErrorError
15 June 2019 23:35:50
Ref 6111143.0151.2143.6151.0(1.1, −8.6)(1.7, −8.8)
Ref 6158121.6161.3120.6160.7(−4.4, −5)(−5.4, −5.6)
Ref 6195101.1170.2100170.0(−1.5, −4.8)(−2.6, −5)
RCM-122.6155.013.7152.6(3.8, 5.1)(−5.1, 2.7)
RCM-235.5147.830.0144.6(1.3, 4.8)(−4.2, 1.6)
RCM-349.9139.645.4136.6(−0.3, 4)(−4.8, 1)
Table 3. Computation Time Comparison.
Table 3. Computation Time Comparison.
Processing Time with 200 Image SequenceAverage Time per ImageProcessing Time with 1000-Image SequenceAverage Time per Image
SBOIS67 s0.34 s280 s0.28 s
STK681 s3.4 s3322 s3.3 s
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Clark, R.; Fu, Y.; Dave, S.; Lee, R. Simulation of RSO Images for Space Situation Awareness (SSA) Using Parallel Processing. Sensors 2021, 21, 7868. https://doi.org/10.3390/s21237868

AMA Style

Clark R, Fu Y, Dave S, Lee R. Simulation of RSO Images for Space Situation Awareness (SSA) Using Parallel Processing. Sensors. 2021; 21(23):7868. https://doi.org/10.3390/s21237868

Chicago/Turabian Style

Clark, Ryan, Yanchun Fu, Siddharth Dave, and Regina Lee. 2021. "Simulation of RSO Images for Space Situation Awareness (SSA) Using Parallel Processing" Sensors 21, no. 23: 7868. https://doi.org/10.3390/s21237868

APA Style

Clark, R., Fu, Y., Dave, S., & Lee, R. (2021). Simulation of RSO Images for Space Situation Awareness (SSA) Using Parallel Processing. Sensors, 21(23), 7868. https://doi.org/10.3390/s21237868

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop