4.1. Nondestructive vs. Destructive Analysis
Nondestructive, in-situ analysis techniques are key components of complex site investigations and have several advantages over destructive techniques. Notably, they allow samples to be analysed quickly and in a safe manner, minimising operator exposure. The most important drawback of destructive analysis techniques is that analyses cannot be repeated because parts of the sample, and particularly features of interest, are destroyed during the analysis process. This is true for traditional chemical analysis methods such as chromatography, which typically require sample fragments to be dissolved, but also for stand-off methods such as laser-induced breakdown spectroscopy (LIBS). In the specific case of LIBS, a high-power laser pulse is used to vaporise a sample and generate a light-emitting plasma, of which the spectral lines are subsequently analysed with a spectrometer. The U-associated emissions, for example, are then easily identified at 385.464 nm, 385.957 nm, and 386.592 nm respectively [
34]. Because the amount of vaporised material and the resulting plasma wavefront is slightly different every time a LIBS analysis is conducted, the obtained data tend to drift. Firing the laser at the same location multiple times leads to cavitation of the sample, affecting plasma formation. If the sample is heterogeneous (for example, if it has a surface composition that is slightly different due to oxidation effects), then subsequent LIBS measurements will show varying results. This is because the first plasma will vaporise surface oxides whereas the following measurements reflect the composition of the now exposed interior of the sample. Analysis techniques destroying samples pose a challenge for scientific falsification. This scientific research principle dictates that peer reviewers must be able to replicate research results to independently verify the research data. It can also have consequences for forensic investigations if samples later turn out to have been criminal evidence [
35] (pp. 455–456).
A third drawback of using destructive analysis methods is that, by definition, a part of the sample transforms into waste products. When analysing radioactive samples, this leads to radioactive waste products that must be disposed of, and such disposal process is regulatorily complex if the composition of the sample and its resulting waste is unknown. Radioactive waste produced during the destructive analysis of radioactive samples may also contaminate analysis instrumentation, increasing its background radiation signature, and further jeopardising repeatability of the analysis. In the case of wet chemical analysis, they are in the form of a radioactive effluent. LIBS, for example, can cause aerosolised radioactive particles found in the plasma to be dispersed over a wider area. This leads to possible contamination of the equipment and scientists performing in situ measurements, with radiologically contaminated PPE requiring additional remediation as a result. Nonetheless, techniques such as LIBS have been used for in situ nuclear industrial archaeology [
36], but when the subject is related to radioactive contamination, such techniques can thus contribute to spreading contamination or causing cross-contamination.
4.2. Geiger–Müller Counting
Geiger–Müller (GM) counters are among the oldest radiation detectors available and remain useful in field-deployed nuclear industrial archaeology as the first line of intervention on site. GM counters can be fitted with probes sensitive to alpha [
37], beta [
38], gamma [
39,
40,
41], positron [
42], or a combination of those ionising radiation types [
43]. GM counters have several advantages in comparison to more sophisticated scintillators or semiconductor detectors that can perform gamma energy spectrometry: they are low-cost [
44], [
45] (p. 103), can be tuned to different radiation types easily by fitting filters, and produce digital pulses that are processed electronically in a straightforward manner [
46,
47].
GM tubes are often wired straight to a buzzer, which produces characteristic ticks when ionising radiation strikes the tube. The ticking rate (per minute or second) is proportional to the number of detected events in a given period. It can also be calibrated over a known distance to allow for initial triaging of samples and areas. The instrument’s ease of use, simplicity, and resulting reliability make it a standard instrument in the nuclear industrial archaeology toolkit.
What makes a GM counter particularly useful in the field is the instant response of the tick rate, as opposed to analogue or digital displays, which make use of moving averages over a set period to calculate and display a count rate in counts per second (cps) or counts per minute (cpm). Because GM counters with a mica window are sensitive to , , and radiation and have such a fast response, they are the best method to track down the location of radioactive contamination of unknown origin. This is typically done by sweeping a pancake-type GM tube with a large window over a surface and listening to the tick rate. Using this method, a trained operator can quickly locate the source of ionising radiation. If the source is physically small, such as a single droplet of a solution containing a radioactive salt, a GM counter is often the fastest and most reliable way to track it down. An additional advantage is that operators do not need to look at a display or gauge to interpret measurements, which improves safety in the field in nuclear industrial archaeology environments such as disused industrial facilities that present numerous trip hazards.
GM counters also have two drawbacks. The first limitation is the inability of a tube to differentiate between
,
, or
radiation. Any of these radiation types will trigger the electron cascade (the Geiger–Müller effect), and the resulting current pulse contains no information on the type of radiation or its energy. This can be partially mitigated by fitting different filters, exposing the tube either unfiltered (
), through a thin Al window (
), or through a thick Al window (
only) [
48]. Some models have built-in filters, whereas others require different probes to be fitted. A second limitation of GM tubes is the low density of their fill gas, which makes them more sensitive to
and
radiation than to
radiation, of which most passes through the tube without significant attenuation, especially at energies above 100
. However, this is rarely an issue in practice because pure
-emitters are of synthetic origin and usually sufficiently concentrated to produce detectable changes in tick rates even at low concentrations and short distances.
Because GM counters are low-cost, compact, and lightweight, they should be among the first instruments deployed in an unknown field environment to assess the radiological situation. First and foremost, they are used to establish whether there is a radiological anomaly present—that is, whether the area of interest shows count rates that are significantly above natural background levels. Secondly, if ionising radiation is detected, it should be used to qualitatively assess the radiation levels in preparation for further field nuclear industrial archaeology analysis. This takes advantage of the fast impulse response and relative directionality of the sensor to measure whether radiation levels are low enough to accommodate safe further inspections or too high (thereby putting operators at risk). For most instruments, the saturation limit—the point at which subsequent pulses overlap—is between – cps. As a general rule of thumb, if the GM counter saturates during the survey, radiation levels are potentially hazardous, and the survey should be aborted as a precaution. In such cases, a robot may need to be deployed instead.
The efficiency and speed with which GM counters can help locate radiation sources in the field means that further analysis methods can be deployed more quickly and effectively (at specific points of interest rather than needlessly sampling large areas). The resulting reduction in time required to perform analyses also reduces the radiation dose of workers.
A consequence of the inability of GM counters to differentiate between ionising radiation types or their respective energies is that personal electronic dosimeters (PED), which are often based on internal GM tubes, are intrinsically unable to measure actual dose rates. This is because the radiation dose is a function of radiation type and energy, neither of which the GM tube can measure [
49]. PEDs are typically calibrated under the assumption of specific isotopic composition (such as
137Cs or
90Sr in a nuclear power station environment), which leads to drastically different results when surveying a processing facility for naturally occurring radioactive materials (NORM), for example. If accurate dosimetry is desired, a gamma spectrometer must be used instead.
The final field utilisation of GM counters is during postdeployment decontamination. Regardless of how careful nuclear industrial archaeology field analysts are during their work, there is always a risk of contamination of protective equipment (PPE) or instrumentation. For overshoes, gloves, or hard hats, such contamination is even expected. GM counters are the ideal instrument for contamination assessments on personnel and equipment, facilitating targeted decontamination if needed.
4.4. X-ray Fluorescence
X-rays are well suited to probe the composition of matter because they can penetrate the surface of a sample to a depth inversely proportional to the density of that sample. The most popular analysis methods using X-rays are X-ray diffraction (XRD) and X-ray fluorescence (XRF). Both are available in a portable form for in-situ (handheld) analysis: portable XRF analysers are in use to determine alloy compositions in metal recycling industry, and portable XRD analysers are in use for metal alloy phase identification in junkyards. Combined XRD/XRF instruments have been developed for space missions, such as the CheMin instrument onboard the
Curiosity Mars rover [
54,
55]. XRF gives information on the sample’s elemental composition and is the most simple of the X-ray analysis techniques.
For analysis of samples in nuclear industrial archaeology applications, XRF as an analysis technique works well for samples containing heavy elements because these elements have electrons in high-energy orbits, which thus produce high-energy fluorescence. Light elements (LE; H to Na) have loosely bound electrons and give rise to high Compton scattering rather than fluorescence. The fluorescence effect is maximal when the incident X-rays are only marginally higher in energy than the electron’s binding energy. If the photons have an energy that is too high, there is a higher chance of scattering than fluorescence. Likewise, if the photons have an energy that is under the binding energy, electrons cannot be ejected, and no fluorescence can occur. Because binding energy is a function of the electron shell, and these, in turn, are a function of the number of electrons the atom has, the ideal X-ray excitation energy depends on the sample’s elemental composition. This is an essential consideration in the design of an XRF instrument. Even when fluorescence occurs in LE, the emitted secondary photons are of such low energy (in the order of a few eV) that they are either absorbed by the sample, the air, or the window of the spectrometer itself. The lightest identifiable element using XRF can vary from Be to Ti depending on spectrometer sensitivity. All heavier elements can be identified up to Am [
56,
57] in concentrations ranging from ppm to percentage levels, although noble gases are rarely calibrated for.
The capability of a pXRF to identify elemental presence from ppm to per cent concentrations makes it an attractive third nuclear industrial archaeology analysis tool in the field—after gamma spectroscopy and GM-counting. Depending on configuration and energy range, the preliminary elemental composition can be obtained after a few seconds, and quantitative analysis results after 1–3 min. Because operators can potentially accumulate a substantial and potentially harmful radiation dose in 1–3 min if the radiation flux is high enough, a preliminary sweep with other dosimetry, such as a GM counter, is a necessary first step to ensure dose rates are sufficiently low to allow further investigation. Handheld instruments such as the Olympus Vanta XRF or ThermoScientific Niton XL3T XRF are sealed instruments with an IP rating of 54 or higher, making them resistant to dust ingress and splashing liquids from all directions. This enables their use within contaminated environments with high dust or humidity levels. The Olympus Vanta XRF used in this study weighs 1.70 kg with battery installed and can be operated with one hand by an operator, as shown in
Figure 2. Miniature X-ray tubes mean these instruments no longer contain radioisotopes such as
55Fe,
109Cd, or
241Am like their predecessors [
58] (p. 94), [
59]. Therefore, the produced excitation beam is an order of magnitude higher, reducing the time needed for analysis by an order of magnitude, but reducing the residual dose to 0 when the instrument is turned off. During operation, the dose is quite substantial, at ca.
at 10
from the aperture, as opposed to only 267
for the TN 9000 XRF analyzer from 1995 [
58] (p. 94). The relatively high dose rate is accompanied by numerous safety features, such as a password-protected user interface, X-ray ON lights, proximity sensors, quick battery release, and dead man switch [
58] (p. 95). The X-ray tube itself is capped in power to 4
[
60] (p. 2).
As in-situ measurements typically do not require sample preparation, there is no setup time needed before NIA measurements can commence. There are, however, several limitations. First, the fluorescence effect peaks at excitation energies slightly higher than electron binding energy levels. If the X-ray energy is either too high or too low, Compton or Rayleigh scattering are the dominant interaction modes for the X-rays [
61]. In the case of Compton scattering, a lower energy X-ray will be scattered into the detector, and in the case of Rayleigh scattering, the scattered X-ray will have identical energy. These produce additional counts in the detector, swamping weak signals and lowering measurement results. If the density of the sample is unknown, or the sample contains both high- and low-density fractions, it is necessary to sweep the X-ray source between ca. 8
and 50
to enhance the measurement accuracy over a range of material densities.
Secondly, as the penetration power of the source X-rays is proportional to their energy and the sample’s composition, acquiring accurate data from in situ measurements requires samples of sufficient thickness to either absorb, fluoresce, or scatter the source X-rays completely. The minimum sample thickness to produce homogeneous results ranges from 55
for solid Pb to
for sand [
61] (p. 16, Table 1). When samples are insufficiently thick or heterogeneous in depth in comparison to the materials with which the instrument has been calibrated, the elemental composition reported by the instrument will be inaccurate. When measuring samples of insufficient thickness on a substrate, the elemental composition of the substrate will also be included in the measurement results, which is a drawback to the technique. This is inevitable for in situ measurements of crusts or oxidation on pipes or tanks, or thin layers of spilt substances in solid or liquid form. There are two ways around this issue that use postprocessing of data. One option is to measure the sample on its substrate or matrix, then measure the “clean” substrate or matrix separately and subtract them from each other. Doing so invalidates concentration data but allows for quick identification of foreign elements on common industrial surfaces such as 304L or 316L stainless steel [
62]. The drawback is those sample elements present (including a variety of application-specific trace elements) in stainless steel are also removed. Another option is to measure samples on different substrates or matrix materials and use subtraction to find the elements both datasets have in common. The drawbacks of this method are identical.
As EDXRF instruments cannot detect LE, a way around the substrate problem is by measuring samples on a substrate that exclusively contains LE. An example of these is disposable cardboard plates or scoops, which can retain a small amount of material for analysis. As neither cardboard nor air induces X-ray fluorescence, measurements made this way will only show the sample’s elemental composition. Alternatively, Kapton tape [
57], or mylar pouches and carry-straps can also be used [
60]. The practical feasibility of each approach depends on the type of material and its specific activity. Tools such as files, paint knives, or scalpels may be needed to scrape off a small amount of material onto a cardboard holder for analysis.
Thirdly, a consequence of analysing radioactive materials that undergo alpha decay is that X-rays produced due to the alpha decay process can interfere with the EDXRF measurements. For isotopes undergoing gamma decay, this is usually not a problem because the gamma energies are beyond those of X-rays (over 100
). However, for isotopes undergoing alpha decay, this may lead to a more complex spectrum and therefore complicate the interpretation of the measurement results [
57]. For example,
Pu exhibits a prominent uranium L X-ray emission peak series as a result of alpha decay into
U. The
U daughter nucleus has a probability of being left in an unstable state, transferring excess energy to an electron through internal conversion. This could lead to the ejection of the electron, just like it would have been ejected by a higher-energy X-ray originating from the XRF analyser source [
63]. The secondary X-rays emitted by higher shell electrons filling the vacancy are therefore indistinguishable, and the XRF detector would interpret the additional counts as higher concentrations of the corresponding elements in the sample. This is a problem for radioactive samples containing Pu and Am, such as nuclear industrial archaeology of radioisotope thermoelectric generators, for which XRF analysis cannot accurately determine the elemental composition in a sample. High-resolution gamma-ray spectroscopy (HRGS) is the preferred technique for studying actinides [
64]. This also allows the use of coincidence methods to enhance the signal-to-noise ratio. This is discussed in further detail in the Gamma Spectroscopy techniques section above.
Finally, as portable XRF measurements for in situ nuclear industrial archaeology are made in the air rather than in vacuum or helium, absorption of low-energy fluorescence is inevitable, and elements with exclusively low electron binding energy can therefore not be measured. The XRF analyser instrument should ideally be brought in physical contact with the sample surface to eliminate as much air absorption as possible and, at the same time, maximise the detection of fluorescent X-rays. When measuring samples with both light and heavy elements, measurement results can be misleading. For example, an organic ionic exchange resin for separation of Th from REEs (shown in
Figure 2) will typically be composed of over 95% resin, and only a few per cent heavy ions. However, because the XRF does not obtain a signal from the LE in the resin, it may falsely indicate disproportionally large concentrations of Th or REEs in such a sample. Some portable XRF analysers such as the Olympus Vanta mitigate this effect by comparing received X-ray intensity with emitted X-ray intensity, and assuming the difference is due to LE it cannot detect. In nuclear environments, it is not always possible or desirable to establish physical contact with a sample, as this could result in radiological contamination of the instrument itself. If prior GM-counter sweeps have established that a surface is contaminated with radioactive compounds or otherwise emitting ionising radiation, it is left to the operator’s judgement to decide whether or not physical contact with the sample is justified. Instruments such as the Olympus Vanta XRF have Kapton windows that can be replaced if they become contaminated, if necessary. The instrument is equipped with a proximity sensor that cuts off the X-ray beam if the aperture is further than 15
away from the sample. Standoff measurements take considerable practice from the operator to keep the instrument’s aperture suspended and steady for the measurement duration, which can be as long as 3 min.
Although it is possible to operate a portable XRF instrument as a single operator, it is instead advisable to opt for an operator pair instead for a nuclear deployment scenario. In such a team, one member can focus on identifying sample locations for in situ measurements or collecting samples on cardboard plates, whereas the other member can focus on performing the XRF measurements with the portable XRF analyser. To avoid contaminating the instrument, pouches are available to carry it hands-free. If a large number of samples are analysed, mounts are also available [
65] (p. 4, Figure 1). It is possible to make up to 15–20 measurements per hour in handheld mode. A hard hat-mounted camera is recommended to facilitate documentation of the measurement conditions.
Figure 2 is an example of a photograph made automatically with a GoPro Hero 8 camera mounted on a hard hat. The GoPro was configured in “photo” mode and set to take a photograph every
.
Decontamination of the portable XRF analyser instrument is necessary between measurements in contact with samples. As XRF is a sensitive nuclear industrial archaeology technique, minor quantities of material in the order of a few grains of sand could result in the instrument falsely reporting elemental composition in the next sample due to contamination. Avoiding cross-contamination is therefore of extreme importance. The best way to avoid cross-contamination is to avoid physical contact with samples completely; if this is not possible or accidental contact was made with any surface, the instrument must be decontaminated meticulously. This can be done inside an active environment by wiping the instrument’s front with industrial wet wipes. Surfactants such as Decon-90 [
66] may be used to remove more persistent contamination. After the survey, the portable XRF analyser must be inspected with a handheld GM counter for any residual contamination and decontaminated as deemed necessary.
4.6. Photogrammetric Survey
Photogrammetry is the trigonometric stitching together of photographs taken from one or multiple sources to recreate a 2D or 3D environment in digital form (coordinates and derived geometric elements) or graphical form (images, drawings and maps) [
68]. Multiple research fields are involved, including optics, projective geometry and cloud computing.
One of the many benefits of photogrammetric models is that they allow 2D photo stills to be compiled into a 3D rendering, sometimes referred to as a digital twin, which enables researchers, workers, and other interested parties to easily visualise the layout and conditions within a contaminated building as a virtual walkthrough and without requiring PPE or exposure to a high radiation dose. It also allows precise measurements to be made of structures that would otherwise be difficult to reach or access. Anything that can be photographed can be rendered into a photogrammetric representation, from sand grains to landscapes.
Photogrammetry straddles the line between qualitative and quantitative data. Although its primary use is in recreating environments for users to explore safely, it can also determine some length measurements such as the height and width of infrastructure features such as tanks, furniture, and walkways. These are not as accurate as using a properly calibrated measurement tool but can be very helpful in obtaining approximate dimensions of internal and external structures. These digital twins, and the photographic records from which they are constructed, can be used to compare changes such as structural degradation, damage, and destabilisation over time in a quantifiable manner. They also enable the study of details which may have been missed in the field due to time constraints, poor lighting, or investigative priority decisions.
This has broad implications for inspection protocol and informed decision-making within hazardous facilities.
4.6.1. Analysis Principle
Image coordinates from n views, n exterior orientations, 3D coordinates, inner orientations of the camera, camera models, and other further observations such as measurements and pattern recognition all combine to produce a photogrammetric model.
Image coordinates define the locations of the object points’ images on the camera. The exterior orientation of a camera defines both its location in space and its view direction. The 3D coordinates define the locations of object points within the 3D space.
The inner orientation defines the geometric parameters of the imaging process. This is primarily the focal length and lens distortions.
These are especially seen in wide-angle lenses such as those used by consumer-grade “action cameras”, used commonly in the field due to their rugged build and ease of use, but every camera lens will have flaws.
These sets of points are then processed within histogram software and analysed for similarities and combined into a point cloud showing the likely position of all photographs. This is then further analysed to produce a wireframe mesh, as seen in
Figure 4. This mesh is then turned into a solid model, as seen in
Figure 5. The texture is then mapped onto this mesh to create an image that a viewer can interpret, as seen in
Figure 6.
Further additional observations play an essential role in adding value to photogrammetric models. For example, scale bars, annotated hazards, overlain radiation maps and marked sample locations vastly increase the amount of information shown in a model.
Photogrammetry is an excellent technique for use in the field because it requires no specialist equipment. Any modern smartphone with a camera can capture photos suitable for photogrammetry, as can any basic point-and-shoot/compact digital camera, bridge camera, digital single-lens reflex (DSLR) camera or image-capable robot or drone. It is easy for researchers to carry or pilot at least one of these devices into all but the most secure or sensitive environments, and minimal training is required to take photos in an efficient grid. Multiple heights, orientations, and positions add complexity and depth to a survey. Still, valuable results can be produced even when taken from a singular position at multiple heights and angles, such as on a gantry or accessway.
Photogrammetric models can be used to create animated “flythroughs” of hazardous sites, used to train new personnel on the layout and dangers ahead of them before undertaking an entry, thereby improving safety and reducing the amount of time personnel need to spend inside and therefore the dose they are exposed to.
Minor specialist training is required to establish a sampling grid method that increases the utility of photographs taken. Once trained, any member of a field-deployed nuclear industrial archaeology team wearing sufficient PPE and personal dosimetry can be entrusted with creating a photogrammetric record. The decision to carry it out should only be taken when dose rates are relatively well understood. It is a form of data collection that can be undertaken between other resource-heavy investigative methods—for example, while waiting for batteries to charge between robot deployments. Operator fatigue is minimal, although can become an issue during time-consuming investigations that involve large, multistorey buildings with complex internal features.
Photogrammetry can also yield evidence that was missed in the field due to poor lighting, time constraints, and other confounding factors. For example,
Figure 7 shows a poster that was spotted during an examination of the 3D reconstruction of the site. This allowed a photo to be found retrospectively, which had sufficient clarity for some translation.
4.6.2. Limitations of Photogrammetric Surveys
Barriers to accurate photogrammetric processing include inconsistencies between frames. Within well-lit scenes, this can come from reflective surfaces and mobile surfaces (such as vegetation moving in the wind). Surfaces with limited texture cause reconstruction problems unless there are other observable features in the photographs. Within poorly lit scenes, low light, moving shadows, and variable lighting produced by head torches and camera flashes all have an impact. Camera blur, image exposure, and other internal inconsistencies also prevent appropriate matching on the ability of photogrammetry software to match images and place them accurately within the 3D model. The algorithms used can average out occasional photographs containing team members. Ideally, this will be eliminated by using clear communication and taking surveys while other members of the team are working elsewhere.
To create as static an environment as possible, floors were photographed by two analysts “back-to-back” around noon whenever sunlight can reach the scene. This ensured that lighting was as full and consistent as possible, facilitating the use of two different camera models. This method also enabled analysts to stay out of photographs.
The point cloud in photogrammetry derives from extra processing done after collection. In contrast, a LiDAR point cloud is a direct product, available in seconds when the position of the sensor is understood.
4.7. LiDAR Mapping
Using 3D scanning light detection and ranging (LiDAR) is increasingly commonplace in providing rapid measurements of complex 3D spaces such as building sites or civil engineering projects. In a nuclear industrial archaeology context, LiDAR provides an enhanced capability over photogrammetry in terms of accuracy (mm versus cm) and the ability to operate in very low (or nonexistent) external light conditions.
Photogrammetry can generate high-quality 3D maps, but one limiting factor is the computing power requirements for stitching together the images. This can be a time-consuming process and is difficult to perform accurately in real time; it is normally done in post-processing. A LiDAR unit uses the reflection of lasers to find the distance to the sensor’s surroundings and can use this data to perform simultaneous localisation and mapping (SLAM), where successive scans are compared to work out the location of the unit in space and build up a map. In this way, the position of radiation measurements can be ascertained in real time. A LiDAR unit can be carried by a human operator, mounted on an unmanned aerial vehicle (UAV) [
69] or unmanned ground vehicle (UGV) [
70], mounted on a robotic manipulator, or mounted to a human operator, and information from other coincident or adjoined sensors, e.g., gamma spectrometers, can be position stamped by the LiDAR [
70]. An example of an UGV is shown in
Figure 8.
If such a sensor is not used, the location of measurements would need to be manually recorded, increasing burdens on the operator in a hazardous environment and potential error. As the LiDAR mapping requires no input from the operator, they are free to perform other tasks if the LiDAR unit can be mounted appropriately on an operator or UAV/UGV, with the position of each task being accurately recorded.
Two-dimensional LiDARs are more cost effective, but a 3D LiDAR, which consists of multiple sets of lasers scanning at different angles, can generate a 3D point cloud of a space. This point cloud can be used to take accurate measurements, and sensor measurements can be expressed within it. As this is a real-time approach, it can be used to influence strategy and decision making immediately in the field, as compared to photogrammetry, which requires some postprocessing.
A LiDAR mapping unit can also be uploaded with existing maps, which the unit can then use to attempt to localise itself within its current environment. This means that measurements taken at multiple different times can share the same reference frame, and thus be directly compared to each other. Photogrammetry models will not have a consistent reference frame, meaning any comparisons must be done manually.
4.8. Unoccupied Aerial Vehicles and Unoccupied Ground Vehicles
Consumer drones are relatively cheap compared to some scientific instruments and can be modified for use in the field. Unoccupied vehicles, whether they are aerial or ground based, provide the possibility of delivering sensors into areas that would be unfavourable or hazardous to send a human operator. Depending on the variant of unoccupied vehicle chosen and its associated payload capacity, multiple sensor packages can be included to collect multiple data types simultaneously, allowing for real-time data fusion and near-immediate interpretation of the results. This can include sensors that are unsafe for human operators. Remote surveys and inspections carried out by aerial or ground-based robotic systems have become significantly more prevalent since 2005, partly due to the significant reduction in the cost of the robotic platforms themselves [
71].
UAVs provide a terrain-independent method of collecting standoff measurements, allowing operators to access areas that would otherwise be inaccessible. Three primary types of UAV exist, consisting of multirotor, single-rotor, and fixed-wing, although a further subdivision can be made for vertical takeoff and landing (VTOL) vehicles that takeoff and land like a multirotor, but operate as a fixed wing in flight [
72]. Most UAVs operated within a commercial environment are of the multirotor type, which is generally the easiest to pilot and offers the most stable platform for data collection, and so the following section will primarily focus on these systems [
72].
There exists a large variation in total size and weight within the commercial offerings of UAVs on the market. These can range from as little as 250
to more than 20
. However, the operation of vehicles with a takeoff weight (including batteries and payloads) above 20–25 kg are often more heavily restricted within aviation regulations throughout the world, making their operation more heavily regulated in many environments. Commercial UAVs with an overall mass at the lower end of this scale are generally equipped with a single, integrated camera as a payload, allowing for the collection of visual information through still images or video. As described above, these can be used for simple visual inspection or for being processed into photorealistic 3D models through photogrammetry [
69]. Generally, as the UAV increases in size or mass, the weight of the payload that it can also carry increases. Larger UAVs, therefore, provide an increased capacity to collect data from multiple payloads simultaneously [
73]. Common combinations within radiological and wider nuclear UAV surveys include the combination of a visual camera, some form of gamma spectrometer system and, often, a LiDAR scanning system [
73,
74,
75,
76].
UGVs, while more limited than UAVs in the areas they can cover, can detect low-level or diffuse sources that would have been invisible to a UAV due to the standoff distance and the inverse square law [
70].
Although a large focus in nuclear industrial archaeology is placed on identifying the chemical or radiological nature of the material within a site, acquiring a contextual visual representation of the site and ensuring the continuing safety of operators should be one of the first objectives for operating workforces. UAVs and UGVs are crucial in this regard, as they can more easily reach and operate in locations that are physically problematic or dangerous for humans, and at the same time, offer greater repeatability and measurement reliability. When coincident radiometric and 3D data is recorded by using adequate methodologies, there is also an opportunity to postprocess the recorded data by using inversion techniques, which often utilise Kaczmarz methods [
77] to provide a localisation calculation of possible “hot” emitters in the survey zone. Although this capability is currently performed in postprocessing, it is expected that near real-time localisations may be achieved in the near future in concert with advances in the computing power of microprocessors used in handheld field instrumentation.
Passing through airport security and customs is more straightforward when there is a basic understanding of the transported equipment, which is nowhere more acutely felt than in the recent rise of consumer-grade drones. Ten years ago, transporting a drone battery involved a discussion with security personnel regarding drones and why they need to be transported; these days, drone batteries and parts in hand luggage do not provoke much interest.
4.9. Discussion
In this work, we have delineated a toolkit of portable analysis technologies for NIA applications. As with any toolkit, tools can be added and taken out depending on the application scenario. The expectation is that the technologies we have outlined will, as a function of time, evolve to increase their effectiveness and useability, and decrease their size, energy consumption, and mass.
We expect that more portable analytical tools will become available within the next decade, for example, Raman, LIBS, and standoff alpha imagers.
In this work, we have put a series of NIA tools into a case study context. The buildings assessed in this work presented a series of challenges and hazards, ranging from the obvious (holes in the floor) to the subtle (doses of up to ). One of the reasons that this investigation is important and relevant is that many buildings on the PChP site are still in use, with complex ownership status leading to previously unknown issues being uncovered in recent years. For example, the designated outdoor smoking area for one of the factories was for many years situated next to a highly active, abandoned uranium leachate storage tank from the building next door. As a direct result of the NIA investigation, this situation has been resolved by management and the smoking area moved.
For this work at PChP, data from individual instruments was recorded and then later combined in postmortem analysis. This meant that there was a delay in the fusion of the data, and therefore a delay in fully understanding the disposition of materials and hazards within the facility. It is expected that within this decade, technological advancements will enable a combination of data streams in near real time and with advances in data transmission technology (5G), the ability to relay large sets of information to a central repository as it is collected. This will ensure that all data recovered is protected and archived. With advances in computing power, techniques such as photogrammetry will be compiled and outputted onsite in minutes, allowing more advanced visualisations of 3D datasets in real time. Advancements in real-time data visualisation will enhance the relay and explanation of important arising information to site stakeholders, decision makers, and workers operating in these environments. Importantly, it will enhance communal understanding of the whereabouts and significance of any residual radiological risks in a visually intuitive manner, which is independent of written or spoken language.
Over the next decade, with the continued sensorisation of the industrial world, it is expected that the field of NIA stands to continually benefit from advances that will be driven by technology development within other industrial fields of application. Equally, NIA methodologies could be beneficially applied to other relevant industrial areas, for example verification of NORM contamination of oil and gas/mining infrastructure.
Technological advances validated in the field of NIA will have downstream benefits for nuclear forensics applications of the same tools. Accordingly, as NIA technology advances, so does NF.
At present, there are many nuclear facilities worldwide reaching the end of their operational lifetimes, and many countries are planning substantial new nuclear energy infrastructure to contribute toward net zero carbon emissions targets by the middle of the century. This means that the field of NIA will apply to an increasing number of sites. These sites are known to be diverse and contain unique challenges and materials, as no two sites are the same. It is our assertion that by understanding and cataloguing the nuclear challenges of the past, it is possible to avoid the repetition of these issues in the development of this planned nuclear infrastructure. At the same time, NIA enables objective assessment of the historic value of nuclear industrial heritage, and NIA research provides the scientific foundations to initiate often much-needed protection and conservation work.