1. Introduction
Light in the visible region has the ability to substantially penetrate into water bodies. A common wavelength of laser sources used for bathymetry is 532 nm which is visible as green light and is close to the optimum. Light detection and ranging (LiDAR) bathymetry measures the time a pulse of such green light needs to travel from the laser source to an underwater target and back to the instrument’s receiver. Furthermore, the power density is reduced through beam geometry and scattering. As the pulse propagates through air and water, it looses energy by interaction with the medium. As a result, the pulse amplitude at the receiver can become very weak, i.e., the signal-to-noise ratio (SNR) becomes very small, which can make reliable detection of the ground difficult.
One conceivable remedy to make the pulse stronger at the receiver would be to increase transmission power. Alternatively another method is to try to reduce the noise. On the one hand, it is conceivable to further improve the optical and electronic components, such as narrower optical filters and amplifiers that contribute very little noise. On the other hand, storing the received signal samples, a.k.a., Full-Waveform, enables SNR enhancement during post-processing.
Gutierrez et al. [
1] mention the possibility of sensitivity enhancement by “averaging multiple waveforms over homogeneous surfaces” and call this stacking. Stacking generally refers to methods of averaging a number of measurements in order to enhance the coherent parts and attenuate the noise, thereby increasing the SNR. For a review of the history of stacking as used in seismic exploration, see Rashed [
2]. Stacking methods already have been successfully used in LiDAR by Stilla et al. [
3] who describe a method of how to automatically align neighboring waveforms in order to maximize coherency and detect weak pulses. They assume a plane geometric structure in the local neighborhood and sweep over a range of angles until a maximum is found. Mader et al. [
4] use waveform stacking in LiDAR bathymetry. For this purpose, the origin and direction of the laser beam axes are first registered from the local sensor coordinate to the project coordinate system using the position information of the measuring platform. This makes it possible to use the water surface as a prominent geometric feature to align the single waveforms by shifting along their axes before averaging. For mainly horizontal ground geometries, they find that the recognizability of ground points is actually improved. It appears noteworthy that they do not directly use the stacked waveforms for estimation of the resulting target distances, but use the distance found from the stacked waveform to define a search range within which the final range is estimated from the single waveform.
Averaging, which is at the core of waveform stacking, has a long history in remote sensing. Motivated by a work of Lee et al. [
5], George and Zamanakos [
6] investigated so-called Comb Filters for enhancement of SNR in radar. Comb filters result from coherently adding up a periodic signal and can be implemented as a tapped delay line. This type of signal processing was called pre-detection averaging because the summation of the repeated pulses, taking into account the phase of the echo signal, takes place before signal detection, see, e.g., [
7].
In LiDAR with the advent of digital signal processing, it became feasible to record the output signal of the optical detector. The coherent addition of several such signals before processing, i.e., detecting targets, is another closely related form of pre-detection averaging and has been used by Riegl [
8] and Pfennigbauer et al. [
9] for some time.
In this technical note, we will describe an efficient variant of pre-detection averaging that can be applied to full-wave laser data and discuss it specifically for use in bathymetry. In
Section 2, we describe the datasets used for the assessment and explain the algorithms. In
Section 3, we present the results and discuss them in
Section 4. We conclude with
Section 5.
3. Results
The nominal averaging count for post-processing of all sites was set to 100. The criterion for obtaining a shape as square as possible led to virtual spot diameters of ≈1.2–2.2 m for the French datasets and ≈8.8–9.2 m for the Italian datasets. The virtual diameter is the sum of of the intrinsic spot size and the diagonal’s length of the averaged area. The smaller virtual spot size of the French datasets is due to a a higher point density resulting from the slower airspeed of the uncrewed platform. A summary of parameters is given in
Table 2.
N is given by the product
and the averaged area is
. It can be seen that the averaged area always is almost square despite different point densities in flight and across directions.
The effectiveness of pre-detection averaging is demonstrated in
Figure 4 and
Figure 5.
Figure 4 showing two randomly selected pulses from the neighborhood of a shot, in thin gray, and the average of ≈100 surrounding pulses, in solid black. The averaged pulse is of smoother shape with a much lower number of local extrema. The numbers of averaged pulses, however, is only approximate for two reasons: first because the actual number is subject to the squareness criterion as described in
Section 2.5, second because outliers are removed prior to averaging.
For
Figure 5, two profiles from France 2 dataset were taken at the same crosscut location. The upper profile is without averaging, while the lower profile was created using an effective averaging number of
. The maximum resolvable depth in the not averaged case is 19 m and 24 m with averaging. The improvement due to averaging thus is 5 m.
Although turbidity in the Venice area was high, underwater points could be determined from all three areas surveyed.
Figure 6 shows a map of Porto di Lido-Canale di Nicolò superimposed with bathymetric points mapped to color. The maximum penetration depth of the laser below the water surface was about 6 m on the day of recording. The underwater structures of the MOSE (Modulo Sperimentale Elettromeccanicao), the Venetian flood protection system, are not visible in the bathymetry because when it is not in operation its parts lie below the penetration depth of the laser.
Maximum penetration depth within the Venetian Lagoon was similar to that of the PdLCdN dataset; the color map, however, was set to cover 3 m only to emphasize the high capability of structure resolution in the shallow water.
Figure 7 exhibits a map of underwater bathymetry of the surveyed area. Some voids visible in the north–south direction are due to insufficient overlapping of flight strips. The map consists of bathymetric underwater points only and shows noisy gaps in places where land cover would be expected, as the model of the SVB algorithm is only suitable for underwater analysis. No maps for comparison purposes could be created without prior averaging of the datasets, because classification of bathymetric points was severely hindered by the large amount of spurious points in that case.
Turbidity was very low at the French Mediterranean Coast and the data showed that a depth penetration of 24 m could be attained.
Figure 8 shows a detail view of the water bottom in perspective projection, taken from the France 1 G1 dataset, highlighting that it is possible to obtain structures in high resolution despite a high averaging number provided flight conditions are appropriate. The rocks in the centre axis of the picture are of diameter approximately 6 m. The visual appearance of the rocks was further enhanced by applying a shadowing effect. Maximum depth could not be achieved in the France 1 dataset, because of a storage limitation that applies for long waveform recordings taken with a setting of 200 kHz for the PRR.
4. Discussion
The effect of summation of a group of closely spaced laser beams with small footprint can be understood as being equivalent to a single pulse of higher energy but larger footprint. Consequently, it can be expected that the measurement properties of the presented method with respect to environmental conditions in terms of wind speed or water currents are comparable to a system with correspondingly higher laser power and larger beam divergence. It is known, e.g., that proper selection of beam divergence, and thus footprint size, can be used to compensate for the influence of surface waves using the averaging effect over the footprint size, see, e.g., [
14]. Although the purpose of averaging is different in that case, i.e., not to increase the SNR, the principle is the same, which in our opinion is a further justification for the term virtual footprint caused by the group of laser beams subjected to averaging.
As indicated in Equation (
6) the SNR should increase in proportion to the square root of the number of waveforms averaged. Instead of showing the quantitative improvement of the SNR in a direct manner, we use the reasoning from
Section 2.4 to show the improvement indirectly. If both the increase of SNR by the square root of averaged waveforms and the exponential law of power decay are true, then the improvement in depth penetration, following from Equation (
11) must be logarithmic in the number of averaged waveforms:
From experience [
20], we learned that typically two times the Secchi depth can be reached with the particular sensor type with SVB waveform processing. In an attempt to compare the theoretical relations with the empirical findings without having available on-site measurements of Secchi depth, we use the top profile in
Figure 5, where a maximum depth penetration of 19 m is achieved, to estimate
to be 9.5 m. Furthermore, a value of 2.15 for
is taken in accordance with ISO 7027-2 [
11]. With these values, relation (
19) is met quite closely.
One of the adverse effects of averaging is, as must be expected, a reduction in resolution of the target scene due to the low pass behavior of averaging. This effect, however, depends on the sampling parameters of the un-averaged dataset as can bee seen in
Figure 8. The strong oversampling in the France 1 G1 dataset still allows a high averaging rate without compromising resolution. But even if resolution is reduced, a certain degree of averaging will be highly beneficial because robustness against the influence of surface waves is increased, thus improving derivation of the water surface. Theoretically, a low pass filtered dataset should be able to be represented with less data without loosing information, as stated by the Nyquist-Shannon sampling theorem. We took, however, the approach of not reducing the amount of data, since often the higher resolution dataset is required for further processing anyways. On the other hand, since the storage and full waveform processing requirements can be demanding the potential for optional reduction may be interesting for some projects.
Even if a reduction in resolution is undesirable, a small averaging degree can still significantly reduce spurious targets, which are more likely to occur in turbid water. While these targets may be of interest for determining water properties in general, they make it difficult to efficiently classify the ground. Applying the proposed method during post-processing allows it to be adapted to different problem definitions.
Although the virtual footprint size of the Italian datasets is already half the size of satellite based systems, e.g., for ICESat-2 the footprint size is 17 m [
21], the airborne system benefits from tunability of the size and can thus be adapted during post-processing. Similar to the approach of Mader et al. [
4] one can also combine bathymetry with high averaging and one without averaging to increase the resolution by considering targets close to the smoothed bathymetry only.
Another possibility is to increase depth performance without reducing the resolution by scanning faster and flying slower yielding a higher point density an approach particularly interesting for use with uncrewed platforms.
Figure 8 is an example of such a use case.
All test datasets show that the approximate Equation (
16) suffices for selecting a suitable group of waveforms for averaging. A more thorough analysis would start at a more accurate description of the laser pattern such as the Palmer Scan, i.e., an elliptical pattern, see [
22], or at an even more advanced analysis as conducted by Li [
23] taking care of the deviations from the elliptic shape.
Another question that requires consideration is the determination of the mean distance
r used in Equation (
16). We use the median value that can be obtained from online range measurements, i.e., before the pre-detection step. This means that although these values represent plausible distances above a water surface, they do not necessarily reflect the distance to a water surface in heterogeneous areas with mixed water and land cover. For the Lagoon dataset, a discrepancy between the altitude above the water surface as stated in
Table 1 and the surface as obtained from the registered pointcloud, as in
Figure 7, was observed; however, the effect was not very large. It is a simple extension to use a nominal, manually specified, altitude instead, but at the expense of non-automatic operation.
We described the method for use in post-processing, but an extension that allows averaging during the flight in real time is conceivable. From the parameters needed in Equation (
16), everything is available in real time from the IMU/GNSS unit.
Comparing the presented method to an approach where the waveforms are registered into a geographical frame before averaging, we can identify the following: Proper registration requires a sufficient number of points that can be derived before averaging is applied. The fact that two points are close to each other in geo-referenced space does not mean that the laser directions associated with them are parallel, a prerequisite for proper averaging.
It should not go unmentioned that the method allows a high degree of parallelism in processing and at the same time has a relatively small, constant memory requirement. This makes it possible to assign each of the horizontal slices that contribute to the
point in
Figure 3 to a separate CPU for simultaneous processing.
While increasing depth performance is one of the goals of averaging, within the scope of this technical note we could also demonstrate the effect of better structure recognition in shallow water. A higher SNR reduces the amount of spurious points resulting in a cleaner pointcloud which greatly assists point classification as demonstrated by the results of
Figure 6 and
Figure 7 which would not have been possible with the un-averaged datasets.
The noise model used in
Section 2.3 is the ubiquitous Gaussian model under assumption of uncorrelatedness of samples. For a more thorough analysis, however, the model should be refined. A distinction would need to be made between sources of interference in the optical and electrical regimes. In the optical case, under high turbidity, the assumption that the repeatability of the pulse shape under stationary conditions is sufficiently high becomes questionable. In the electrical regime, the effect of the detector, a magnitude detecter, is to change the Gaussian noise into a Rician distribution. Since the signals are also amplified, another source of noise that depends on the magnitude of the signal, i.e., shot noise, enters into the refined analysis.