Next Article in Journal
Snow Cover in the Three Stable Snow Cover Areas of China and Spatio-Temporal Patterns of the Future
Previous Article in Journal
Demand for Ecosystem Services Drive Large-Scale Shifts in Land-Use in Tropical Mountainous Watersheds Prone to Landslides
Previous Article in Special Issue
RETRACTED: Robot Path Planning Method Based on Indoor Spacetime Grid Model
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Review

The Hitchhiker’s Guide to Fused Twins: A Review of Access to Digital Twins In Situ in Smart Cities

1
Cognitive Science, ETH Zurich, 8092 Zurich, Switzerland
2
Game Technology Center, ETH Zurich, 8092 Zurich, Switzerland
3
Visual Computing Group, Harvard University, Cambridge, MA 02138, USA
4
Center for Sustainable Future Mobility, ETH Zurich, 8092 Zurich, Switzerland
5
Geoinformation Engineering, ETH Zurich, 8092 Zurich, Switzerland
6
Department of Biology, Saint Louis University, St. Louis, MO 63103, USA
7
Data Science, Systems and Services Group, ETH Zurich, 8092 Zurich, Switzerland
8
Department of Architecture, University of Cambridge, Cambridge CB2 1PX, UK
9
Bartlett School of Architecture, University College London, London WC1H 0AY, UK
10
Learning Sciences and Higher Education, ETH Zurich, 8092 Zurich, Switzerland
11
Department of Psychology, Bond University, Robina, QLD 4226, Australia
*
Author to whom correspondence should be addressed.
Current address: Visual Computing Group, Harvard University, Cambridge, MA 02138, USA.
Remote Sens. 2022, 14(13), 3095; https://doi.org/10.3390/rs14133095
Submission received: 20 April 2022 / Revised: 8 June 2022 / Accepted: 21 June 2022 / Published: 27 June 2022
(This article belongs to the Special Issue Digital Twins for Sustainable and Smart Cities)

Abstract

:
Smart Cities already surround us, and yet they are still incomprehensibly far from directly impacting everyday life. While current Smart Cities are often inaccessible, the experience of everyday citizens may be enhanced with a combination of the emerging technologies Digital Twins (DTs) and Situated Analytics. DTs represent their Physical Twin (PT) in the real world via models, simulations, (remotely) sensed data, context awareness, and interactions. However, interaction requires appropriate interfaces to address the complexity of the city. Ultimately, leveraging the potential of Smart Cities requires going beyond assembling the DT to be comprehensive and accessible. Situated Analytics allows for the anchoring of city information in its spatial context. We advance the concept of embedding the DT into the PT through Situated Analytics to form Fused Twins (FTs). This fusion allows access to data in the location that it is generated in in an embodied context that can make the data more understandable. Prototypes of FTs are rapidly emerging from different domains, but Smart Cities represent the context with the most potential for FTs in the future. This paper reviews DTs, Situated Analytics, and Smart Cities as the foundations of FTs. Regarding DTs, we define five components (physical, data, analytical, virtual, and Connection Environments) that we relate to several cognates (i.e., similar but different terms) from existing literature. Regarding Situated Analytics, we review the effects of user embodiment on cognition and cognitive load. Finally, we classify existing partial examples of FTs from the literature and address their construction from Augmented Reality, Geographic Information Systems, Building/City Information Models, and DTs and provide an overview of future directions.

1. Introduction

Can Digital Twins (DTs) [1,2] really make cities smart? The pressure to transform cities into Smart Cities is mounting amongst urban policymakers worldwide [3]. Initially, the hope was that smartification would be instilled in cities by simply creating citywide Internet of Things (IoT) installations and remote sensing applications [4,5,6]. The idea behind this thinking is simple yet alluring: to have everything in the city attached to the Internet and make it proclaim its own state as if it were a smart object [7] remotely observing everything [4,6]. Once observed and connected, everything should become measurable. In such a Smart City, every minute detail, if measured, can be included in the analysis. This would allow for the uncovering of new insights [8] and give stakeholders a deeper understanding of processes that were previously thought to be fully understood.
However, the complexity of the collected data, the observed processes, and the city itself requires a unifying view. A DT [9] provides a platform on which cities can become really smart by shifting attention from data collection to data use. Here, a digital representation of the city allows for modelling and predicting its ever-changing state via simulations [10,11,12,13,14]. Unfortunately, until now, the virtual and Physical Environments of cities remain insufficiently interconnected [15]. At best, DTs are remotely displayed in city administration offices to obtain insights [9,16]. However, DTs have the potential beyond administration to inform decision-making on all levels from citizens to politicians and planners [17]. To realise such a wide-ranging implementation of a DT for a Smart City, its data must be accessible in situ within its physical context. Situated visualisations [18,19] enable sensing and actuating in the environment with which it is combined.
We introduce the Fused Twins (FTs) (on a grammatical side note, FTs are always plural because they consist of both the Physical Twin and the Digital Twin and their interaction) paradigm (see Figure 1) to enable the conceptualisation of an accessible Smart City. We argue that there is a need to clarify the definitions of DTs, Situated Analytics, and Smart Cities with an emphasis on the visualisation and development of a practical interface. The FTs concept aims to embed the DT in its Physical Twin (PT), fusing them via Extended Reality (XR) applications [17]. Throughout this paper, we follow the working hypothesis that DTs combined with Situated Analytics enable lower cognitive load, enhanced cognition, and easier embodiment; and therefore, FTs can make Smart Cities more accessible to users ranging from managers and policy makers to citizens. Ultimately, only when data are accessible where it is most needed and understandable to those who consult it, can DTs fulfil the promise that IoT alone could not: to power Smart City developments by interlinking the physical and Virtual Environments.
This review is structured in four main sections. To define the term FTs, we need a thorough understanding of Smart Cities, Situated Analytics, and DTs. The first three sections cover the necessary background to develop, understand, and apply the FTs paradigm. We dedicate Section 2 to gaining an understanding of how Smart Cities arose as a research field and why they converge towards DT research. Section 3 covers Situated Analytics as a novel data exploration approach for Smart Cities. The similarity in the implementation structure leads us to the following section on DTs. Section 4 elaborates on DTs. We provide a definition following [23], where a DT is constructed using five components (see Figure 2). DTs is a vague term that requires multiple clarifications in terms of composability (Section 4.1), servicisation (Section 4.2), and differences among cognate terms (i.e., functionally and historically related but slightly different terms [24]; Section 4.3). Finally, in Section 5, we cover the FTs concept in depth including the required steps to implement FTs and classify examples of de facto FTs in the recent literature. Section 6 provides an in-depth discussion on Smart Cities (Section 6.1), Situated Analytics (Section 6.2), DTs (Section 6.3), and the newly introduced FTs paradigm (Section 6.4).
Beyond the definition for FTs, we also provide several new points on a wide range of topics. These topics have often been covered shallowly in shorter research papers, which has left a great potential for confusion. In this review, we try to discern some of these, especially in the context of DTs.

2. Smart Cities

Smart Cities are an inherently technological approach to resolving urban issues that spawn economic, social, environmental, and governmental aspects [25] (see Figure 3). Smart Cities are constituted using bench-marking indicators that can be constructed from technological sources [26]. This technologist approach only provides a narrow perspective [27] that is often divided along disciplinary boundaries [28]. The weakness in the Smart City concept is juxtaposed by its necessity because technological advances are one of the few options to manage the complexity of modern cities [25]. However, at the same time, it must be acknowledged that there are other perspectives or “urban imaginaries” [29] beyond Smart Cities such as resilient cities [30], intelligent cities [26], and sustainable cities [31]. In the remainder of this section, we will expand on the evolution of “Smart Cities” from the nineteenth century all the way to its portrayal in contemporary urban discourse.
Throughout their evolution, advances in ICT have repeatedly helped cities to cope with growing complexity [32,33,34]. During the nineteenth century, industrial cities were faced with the management challenges of large industrial enterprises as well as new administration challenges due to unprecedented population growth. The need to process information and communicate it quickly over great distances led to the invention of the telegraph, which enabled the industrial city to keep growing [32,35].
This transformation, from a society in which communication was impeded by distance to a society in which distance is addressed using ICTs, has forever changed humanity and cities [35]. On the one hand, critical obstacles to the growth of cities were removed with the integration of technologies that enabled businesses to flourish and municipalities to govern more effectively. On the other hand, futurists such as Toffler [36] and Negroponte [37] predicted that these technologies will lead to the “death of distance” because it is possible that improved telecommunications technologies will make cities, and space itself, obsolete. Their basic proposition was that electronics will eliminate the need for face-to-face interactions and that the cities at their core facilitate those interactions and thus would themselves become obsolete [36,37]. Clearly, this is not yet the case. Major cities have become points of intense investment in ICT because the production, exchange, and dissemination of information were and still are critical to their function. Paraphrasing [38], the shift from goods handling to service and information industries has made cities into communication systems that are central to information flow in the global economy [38].
Today, the integration of ICTs into Smart Cities has become a complex man-made creation that is unparalleled in its cascading effects on every aspect of the city and its citizens [32,33]. Two innovations in the 20th and 21st centuries have enabled this evolution: the rise of the Internet infrastructure for networked computers and mobile personal computing. In the words of [33], “The democrati[s]ation of computing power that started with the PC in the 1970s and leap[t] onto the Internet in the 1990s is now spilling out into the streets [33]” which can be interpreted as the rise of IoT as part of the Smart City landscape.
Contemporary ICTs have restored the importance of space to the previous discourse on urban telecommunication [39]. Everything and everyone in cities generate geo-tagable information which can reveal the position of a multitude of stationary or moving objects in real time [39,40]. Geo-tags enrich space with contextualised electronic content and are a fundamental dimension of a sort of return to space, or spatial turn, to use the expression coined by urban geographer Soja [41]. This spatial turn of digital technologies has been reinforced by the proliferation of electronic interfaces and wireless communications, which realise the notion of Ubiquitous Computing or Ambient Intelligence [33,39,42,43,44,45] and ultimately underpin a DT.
Sometimes visible, but more often hidden, countless chips and sensors allow objects, people, and vehicles to be located; consumption levels and transactions to be recorded; and temperatures, pollution levels, population densities and flows to be measured [46,47,48]. The possibility of transforming every Thing in the built environment into something more “smart” appears more feasible than ever before. Low cost, miniature sensors enable everything to sense and be sensed. The ability to perform analytics over the data collected has become ubiquitous, enabling Things to “think”, while wireless communication technology enables Things to “talk” to one another as part of the Internet of Things (IoT) [33,37,39,40]. However, as noted earlier, this dream of smartification has not been realised due to the increasing complexity of cities. It increasingly looks as if a DT is required to provide the necessary abstractions to enable the Smart City.
Ambient Intelligence has taken on many forms that resemble the functionality of DTs, but they remain disconnected and often disjoint. This leaves it up to the observer to connect the dots, forming the meta-analysis on the city level. Examples of Ambient Intelligence range from trash cans and parking lots to streetlights. Smart waste disposal containers in Barcelona are able to know when they have reached their maximum capacity and must be serviced [49]. Smart parking uses sensors to know if a parking lot is free or not. Citizens can then be directed to available parking in the London borough of Westminster [50]. In the Future City Glasgow project [51], streetlights can be equipped to sense and analyse what is happening below them. Sensors collect data on temperature, humidity, noise, and movement, and interpret the data into a functional “information-scape”. The streetlights can then respond in real time by adapting their light intensity or by communicating the interpreted data to the nearby police station, reporting any ongoing crime.
There are also even more ambitious approaches towards constructing a DT. The Spanish city of Santander with 180,000 inhabitants installed 20,000 sensors to monitor temperature, luminosity, ground humidity, and available parking slots [52]. Paris’s 120,000 trees have each been equipped with a radio-frequency identification (RFID) chip [53]. In a large number of towns, vehicular traffic is monitored in real time through technology ranging from induction loops, to wireless sensors, to video cameras. In Singapore, this monitoring enabled the introduction of one of the first dynamic pricing systems, which consists of making drivers paying in proportion to the desirability and, more importantly, the congestion levels of thoroughfares [39]. This new capacity of cities to detect, measure, and record what is happening within them is truly unprecedented in their history [33,39,40] and yet requires new infrastructure to manage the sheer complexity [54,55]. In the words of [56]: “If aerial photography revealed the skeletal structure of the city during the nineteenth century, the revolution brought about by contemporary ICTs is likely to reveal its circulatory and nervous systems—as it reveals the “real-time city”. For the first time in history, cities can be understood as a whole, the way biologists see a cell—instantaneously and in excruciating detail, but also alive [56]”.
This new understanding of real-time urban dynamics is not (yet) centralised, nor is it ensured to be the property of city governments and planners. If servicised properly and arranged as DT components, it may remain distributed in nature with a low barrier to entrance and can be accessed from nearly everywhere at nearly any time. The ongoing rise in personal mobile computing, especially smartphones, is transforming how individuals perceive and interact with the urban space that surrounds them. Situated Analytics powered by DT infrastructure would provide citizens, tourists, and other users unprecedented insights into the city to help them in their every day life and in planning how to use the city in the future. Using DTs to power Smart Cities will make analytical tools available to everybody and will make them accessible in a spatial context that will democratise the understanding and use of the city. To realise this vision of a Smart City and deal with the complexity that is already plaguing current implementations of Smart Cities, DTs offer a bedrock. The composability of separate DTs into a city-wide DT, underpinning the Smart City, is a promising approach (see Figure 4).

3. Situated Analytics

Augmented Reality (AR) combines computer-generated visual content in 3D with the real world by superimposing virtual content onto a view of the real world [57,58]. The principles of an AR system are three-fold, combining real and virtual elements, registering in three dimensions, and, importantly, providing interactions in real-time [57]. There are at least two categories of approaches to produce AR. First, screen-based approaches, or “see-through AR”, can be divided into the optical combination of a screen with the user’s perception of the real world or 2D displays that replay a live recording of the real world and superimpose virtual content [57,59]. Second, projector-based approaches, or “spatial AR”, consist of projecting the virtual content directly onto the real world, benefiting from projection mapping algorithms and hardware [59,60,61]. The last decades have refined see-through AR and, today, the main lines of AR systems are either based on the HoloLens for optical systems and ARKit or ARCore for display-based AR on smartphones and tablets. With the systems maturing to the point where it is both easy to use and easy to generate content, AR applications are going mainstream [62] and find applications across all industries [63], commerce [64], entertainment [65], education [66], and medicine [67].
As AR becomes more than a gadget, it opens the door for immersive analytics [68] of data and information either in situ or remotely (see Figure 5). Whereas the systems to display AR have been developed over the last decades, the visualisation of information in this new media is still in its infancy and yet rapidly evolving [69]. For example, traditional visualisation on 2D screens and print is a well established field with many well known representations and toolboxes, and high literacy among producers and consumers. However, most spatial 3D visualisations are still hand-coded and often imitate 2D visualisations. 3D visualisations also offer the opportunity for interaction and embedding in the augmented space [70,71], possibly easing cognitive load and enhancing cognition [72] through externalisation [73].
The development of AR technologies also enables new types of interaction with digital content. Specifically, embodied interaction is a concept that reconnects the fields of tangible and social computing and argues that interaction is fundamentally physical and social. Thus, failing to consider these aspects for design interaction would be a fallacy [74,75]. Embodied interaction is defined as “the creation, manipulation, and sharing of meaning through engaged interaction with artifacts” and places the responsibility of meaning-making on users rather than designers [74]. The level of responsibility for meaning-making at the object’s level is driven by different definitions of affordances across research fields that have yet to be reconciled [76]. There is an epistemological difference between affordances as properties of the object or interaction [77] and affordances as properties of the users’ mental representation [78]. From both Normanian [78] and Gibsonian [77] perspectives, AR can be viewed as deeply anchored in the physical world and therefore demonstrate strong affordances for embodied interaction [57]. Moreover, AR supports a higher amount of embodiment compared to 2D displays by enabling high levels of sensorimotor engagement and immersion [79].
For example, anchoring and embedding visualisations in the physical 3D world can offer tangible ways to explore one’s physiological data [80] and, to some extent, one’s own “personal” DT. For example, Tobe is an AR tangible humanoid DT to explore one’s physiological data such as heart rate and breathing patterns and was used in a social context to let two users relax together via data observation [81]. Similarly, Teegi is an AR tangible humanoid DT used to explore one’s brain activity via electroencephalogram visualisations [82]. Non-humanoid representations have also been explored. Inner Garden is an augmented sandbox embodying one’s physiological state, representing breathing patterns as waves and cardiac coherence as weather [83]. In the wider context of embodied interaction, users’ bodies have also been explored as canvases for digital content [84], for example, by co-locating display and interaction on the users’ hands [85] or utilising embodied metaphors, such as mimicking scissors with one’s hand to cut through digital content [86]. Tests of embodied AR approaches have also been initiated for data exploration in the context of geographical data for visually impaired users [87,88,89].
The integration of users’ bodies and their physical context has been conceptualised during interaction with digital content. Towards that end, different physicalities of embodiment have been described: body-centred, object-centred, and environment-centred embodiment [90,91]. We offer examples of embodied DTs and FTs following this categorisation in Figure 6. These categories themselves include direct-embodied, enacted, manipulated, surrogate, and augmented approaches [90,91]. The direct-embodied approach considers the users’ bodies as primary constituents of cognition and therefore the body state is used to convey information. In the enacted approach, bodily actions are considered a form of enacting knowledge through movement. The manipulated approach uses the body to directly manipulate a representative object. The surrogate approach uses a physical avatar that performs an interaction with the object. Finally, in the augmented approach, the body is used to manipulate an avatar to interact with the larger environment. The impact of the physicality of embodiment on interaction and cognition is not yet clear. However, these considerations are of importance because a mismatch between activity design and type of embodiment can result in decreased movements, persistence, and learning outcomes [92].
When discussing the spatial context of data visualisation and data embodiment, we also ought to consider the role of the users’ bodies in the interaction. Often, the users’ bodies are solely considered as physical entities (i.e., Körper) utilised to push buttons and perform actions, rather than feeling entities (i.e., Leib) [93]. Novel approaches such as somaesthetic appreciation design [94,95] describe the manner in which users’ bodies may be integrated in the design process as well as in the experience with digital content and data. Empirical evidence suggests that bodies’ representations impact the sense of embodiment [96,97], as well as the types of gestures performed to interact with a system and the time spent observing and interacting with data [98]. In this regard, the challenge of FTs extends beyond meaningfully anchoring virtual data onto a physical context to include the design of the interaction with the augmentation from an embodied, somaesthetic perspective [74,94,95].
Two further problems arise when visualising embodied information in augmented space. First, most traditional authoring tools are focused on 2D visualisations, and new toolkits such as DXR [99] and RagRug [100] have only provided the basis to display 2D visualisations of data in 3D. Second, the benefit of the third dimension is difficult to grasp and even more difficult to implement. Most modern immersive analytics actually rely on embedding 2D displays into the real world [17,99,100,101]. Visualisations in 3D are typically derived from the previous 2D media, but there is a trend towards exploring new possibilities. For example, multiple 2D visualisations can be arranged to use the third dimension to filter information and visualise this filtration [102].
Situated visualisations are less represented in the literature. Recently, there has been some defining work that characterised the types of enhancements that immersive analytics may provide through enhanced vision and visual synesthesia [72]. Addressing the question of how situated visualisations might improve cognition is an ongoing research area. The supposition has been put forth that cognition may be improved by enhanced attention, enhanced numeracy, enhanced recall, enhanced comparison, and enhanced prediction [72]. Moreover, theoretical frameworks such as embodied, situated, and socially situated cognition support the idea that cognition is deeply rooted in its physical and social context [103,104,105]. However, there are no empirical evaluations yet because, so far, researchers have not focused on the embodiment of the user in the analytics or understanding how the spatial context can reduce cognitive load or enhance cognition.
Some research is starting to address the impact of the third immersive dimension on visualisation possibilities [106]. For example, one small study demonstrated that a toolkit to create new immersive visualisations helped to keep users engaged for longer and produce more elaborate immersed visualisations. We have a clear understanding of how 2D visualisation improves cognitive load through externalisation [73], but for AR, the evidence is less clear [107,108,109,110,111]. Previous work has proposed the following mechanisms for how 3D visualisations can improve cognition by reducing task switching with visual cues [112,113] and context-sensitive displays [114]. Moreover, recent work also explores the potential of AR to reduce negative split-attention effects in multimedia activities [85,115,116]. Split-attention effects [117] occur when two related sources of information are spatially or temporally distant. However, split attention is not always negative. For example, indexing (i.e., connecting multiple data representations through pointing gestures) is investigated as a mechanism for sense-making [118]. The distance between information sources can also be reduced using AR, in particular the FTs paradigm, because it co-locates physical and virtual entities in real time [57,115,116,119,120]. In addition, overwhelming displays of information have been shown to decrease performance [121]. There is no definite answer as to how immersive analytics ultimately improves users’ performance, but theoretic arguments keep this research area active.
Furthermore, the spatial context in the augmented space has often been used as an empty canvas [99] instead of a source of information [122]. Most commonly, this spatial context has been used for geographic representations to show the built environment either in the past through remote sensing or as planned in the future [122]. The subfield of Situated Analytics [19] combines data with its spatial context to achieve a form of virtual data physicalisation [123] and embodied interaction [70,74,75,85,86] that enables humans to maximise the use of all senses to perceive the data (see Figure 7). Indeed, information on cities can be embedded in the situated context with meaningful 3D representations that users can explore. For example, situated AR has been used to teach historical events by anchoring the relevant data in the actual physical context of these events [124]. The real world can become the canvas to help users make sense of the information in situ [125].

4. The Digital Twin

Before we explore how FTs can make DTs more accessible, we will review DTs, their historical background, and their relations to other adjacent technologies and the wide range of definitions that accompany them. Historically, the conceptualisation of DTs originated from the industrial branch of IoT developments in 2003 [1]. The concept was loosely defined at the time but already included the most important components: a physical object, a virtual object, and their connections [2]. Throughout this section, we will expand this understanding to reach our definition of a DT based on composability and the five components, the Physical Environment, the Data Environment, the Analytical Environment, the Virtual Environment, and the Connection Environment (see Figure 2).
Historically, there was a lag in the adoption of DTs because the technologies necessary to effectively implement a DT were not yet developed [8]. In the 2010s, the DT concept outgrew its industrial roots by capturing cyber–physical integration as well as representing the physical world virtually [127,128]. As DTs grow in size from factories and cities [129] to the globe [130], remote sensing with drones and satellites plays an increasingly important role to keep DT representations up to date but also offer an opportunity for interactions with the city through interactive mobile display technology [61]. For a full review on current DT applications, we refer to [54]. Here, we will elaborate the definitions of DTs to form a common basis from where we can develop the concept of FTs.
DTs have been “hyped” according to Gartner as one of the top ten most promising technological trends of the previous decade [8,131,132,133,134]. Beyond the hype, across many different definitions, DTs fuzzily describe any software aiming to closely represent something in reality [135]. These qualities are in line with common and older definitions such as Cyber–Physical Systems (CPSs) [136], Ubiquitous Computing (UC) [46], and Mirror Worlds [137], which begs the question of whether we can (and should) draw a line between DTs and these concepts. Here, we argue that DTs should be approached as an analogy to a mental model that helps to structure information and actions and to attach them to the real world through appropriate mechanisms such as the FTs concept.
In this review, we next situate DTs in the context of their predecessors. DTs are a specialised form of Cyber–Physical System (CPS) [136] because they focus on the accurate representation and manipulation of processes in the real world with a model. DTs are similar to Ubiquitous Computing [46] because they attach computations to real-world objects, but without the FTs paradigm, DTs are ultimately not Ubiquitous Computing because the computations are detached from the embodiment of the represented object. Fusing the PT and DT [17,101] joins CPS properties with the PT to produce truly Ubiquitous Computing as envisioned by Weiser [46].
Whereas Ubiquitous Computing has no clear requirements for the kind of computations to be performed, there is a common theme across different definitions of DTs with respect to the need for simulations to predict a system’s response to critical events. The aim to predict often goes beyond simulating known events towards understanding deviations between the actual observations and the predictions to indicate unknown issues [138,139]. Furthermore, DTs are often expected not only to report on the state of the system, but to also actively manipulate the system through actuators [2,54,139]. The tension between the accuracy of DT predictions and computational capacity is captured by the idea of Mirror Worlds [137], which perfectly imitate reality similar to a mirror. Early ideations of the DT therefore required the underlying model to support similitude [138], while similitude requires the implementation of a multi-physics, multi-scale, probabilistic, ultra-fidelity simulation that mirrors the PT based on historical data, real-time sensor data, and a physical model. The “ideal DT” requires similitude to become as close as physically possible to mirroring the real world. Indeed, DTs need to model the mutual dependency of hardware and software to tightly couple the PT and the DT, producing high-quality simulations. Until recently, most simulations were feasible only for certain sub-components of complex systems that relied on scientific knowledge and real observations to produce a more accurate offline simulation. Therefore, the requirements of similitude have only been satisfied in specialised cases such as with the components of flying vehicles [138]. New computational methods increase simulation performance such that events in the DT can be simulated fast enough for predictive purposes in the real world [140]. However, even with respect to aircraft, there are only plans to produce a DT of a whole vehicle [141,142] which are expected to be completed in the coming decade. Regarding the sheer scale of cities, the Mirror Worlds that would underlie a Smart City seem to be a distant dream. Due to the simulation complexity of large-scale systems such as cities, researchers have since called for a less rigorous model definition to make DTs relevant to real-world applications [9].
In other words, stepping away from ideal DTs to “practical DTs” can allow for more practical applications. Easing any of the core aspects of similitude allows for the implementation of many useful concepts such as analytical assessment, predictive diagnosis, and performance optimisation [8], even in large-scale systems. Limiting the functionality of DTs to a synchronous representation of reality can also allow for the monitoring of complex production processes, adjust these processes, and possibly optimise them [143]. Ideally, the model underlying a DT is complex enough to enable a system to continuously react to dynamic changes in physical space [144], including unforeseen/unexpected changes [145]. DTs also open the door to continuous design improvement through data feedback loops [146] enabling data assimilation into a model [147,148]. These incomplete but practical DTs can become especially useful if they are composable to cover more features of the PT. Next, we will introduce specialised, practical DTs which answer questions on particular topics and can be composed to resemble ideal DTs.
In our definition, we side-step the discussions surrounding ideal DTs by focusing on structural features of practical DTs. We refine the definition of a DT based on earlier work [8] as a five component system: the Physical Environment, the Data Environment, the Analytical Environment, the Virtual Environment, and the Connection Environment (see Figure 2). In the Physical Environment, IoT systems [47] and remote sensing units [149] try to capture the PT and provide interactions with the PT through actuators. The Data Environment contains the DT’s memory and is usually located somewhere in the cloud but may also be stored locally or at the edge [150,151]. For the Analytical Environment, the processes and models of the DT are computed [152], often in the cloud, and insights and predictions are communicated back to the Data Environment. The Virtual Environment makes the DT accessible to human users and possibly other machines [17,101,153]. Finally, the Connection Environment provides an Application Programming Interface (API) to transfer data as required between the components. We consider two ways of designing the Connection Environment, (1) from a service perspective, where the Connection Environment is not an independent component but refers to the APIs exposed by the different components, and (2) from a compatibility/hypervisor layer perspective where the Connection Environment is an independent component that mediates between the other different components in the system. Each component could be implemented in many different ways, and making a sensible choice may be overwhelming. This review gives an overview to guide newcomers through the most important components and decisions that are required to implement the FTs concept with a DT.
To understand DTs, it is necessary to comprehend the function and goal of the core components in relation to the literature. Specifically, we connect our definition of the Physical Environment to research on IoT and remote sensing, as well as our definitions of Analytical and Virtual Environments to research on simulation and interaction with DTs. Our definitions of the Data and Connection Environments are not broadly covered in previous research because previous implementations have used case-specific data solutions and were not focused on interoperability.
The Physical Environment component captures the PT and associated interactions via IoT and remote sensing infrastructure. The composition of IoT stack is regularly repeated but rarely analytically decomposed with respect to its trade-offs. For most applications, the stack is designed with the purpose of underlining where and why a specific new technology is needed [154,155,156,157]. Technically, this is a useful approach to help experts situate a technology in the literature. However, these stacks typically make many unspecified assumptions that we disentangle here. We start with the notion that the DT perspective can be mapped to aspects of IoT and remote sensing. Following the semantic decomposition of the term IoT [158], Things themselves are not actually in the Internet, but information about them is perceived either with attached sensing units or with remote sensing and then shared via the Internet. The DT can be used as the necessary representation of the perceived Thing in the Internet.
Previous resource-oriented and service-oriented architectures have represented IoT hardware as practical DTs without using this particular nomenclature [157,159,160,161,162,163,164]. In contrast to IoT definitions, the DT perspective does not necessitate the digital information to be shared globally as may be implied by the term Internet. The potential local scope of DTs is reflected in their origins in cyber–physical systems [2,136] where the application scope is usually limited to an industrial production site. Nonetheless, the DT perspective provides us with a useful definition of the properties that a Thing should have when represented digitally. A broad range of example IoT applications have been addressed in previous publications [7,45,47,165,166,167] and can be generalised to DT solutions.
The Analytical and Virtual Environments in DTs are difficult to precisely define because their development is multi-disciplinary and reaches beyond the simulation perspective. These paradigms have been advanced by independent research in the areas of information science, production engineering, data science, and computer science [8] as well as electrical engineering [168]. A DT is a highly complex system that combines infrastructure, simulation, data processing, human–computer interaction, and services [8]. These approaches give rise to another understanding of DTs as an advanced digitalisation of a system that extends representation of the PT with operations. This produces a rich digital model that has many desired properties such as scalability, interoperability, extensibility, and high fidelity [169]. To arrive at a digital model, digitalisation in a space can also be understood as a virtualisation process [170]. Here, the system is represented in 3D space via modelling and remote sensing. First, behaviours are extracted and knowledge representations are created. Second, the interactions between its constituent elements are modelled and integrated into the knowledge representations. Third, operations and their effects are modelled. Lastly, everything is assembled in a simulation to make the system match the PT as close as possible. Because digitalisation/virtualisation only creates the Analytical Environment (see Figure 2), most research has clearly focused on the analysis and not interaction with the DT. Most research on DTs also follows the simple physical–digital split into PT and DT components [2,135], but this split often fails to capture the difference between the Analytical Environment for computation and the Virtual Environment for interaction (but see [23]). The FTs paradigm offers a new perspective that highlights the unique contribution of the Virtual Environment, resulting in a more intuitive use of the DT in nearly every application domain.
There may also be issues with many definitions of DTs because of the unrecognised difference between a universal ideal DT and a partial practical DT. Ultimately, universal ideal DTs are unachievable because of the need to integrate the complex requirements of similitude and to comply with the many definitions across different literatures. Therefore, any practical DT is only partially implemented compared to ideal DTs. Confusion regarding this distinction has led to a heated debate and has resulted in the differentiation of DTs according to their degree of completeness from digital models and Digital Shadows to (ideal) DTs [152,171], see Figure 8.
A digital model is an independent representation of a real-world system that may be used in simulations but that is generally not fed real-time data. According to our interpretation, a digital model reflects the implementation of the Analytical Environment. A Digital Shadow is the assembly of data from the real-world system for descriptive purposes. By our interpretation, a Digital Shadow reflects the Physical Environment of the DT. Thus, our definitions can shift the discussion on Digital Shadows and digital models from whether something is a DT to what component of a DT is implemented. While a completely implemented, ideal DT at the urban scale is probably out of scope for now, it is feasible to implement and compose partial practical DTs that are capable of performing important tasks for Smart Cities. To elucidate IoT, DTs, and Smart Cities, the present paper adopts the FTs perspective because it unites three views on DTs that are often confounded: Why to generate the DT (i.e., business case for the Analytical Environment), how to generate the DT (i.e., technological implementation of the Physical Environment), and how to use the DT (i.e., end-user interaction of the Virtual Environment).

4.1. Composition of Digital Twins

Composed DTs offers an integrative perspective to understand and address the complexity of the PT by using multiple abstractions and performing different tasks on each abstraction (see Figure 9). Towards this end, DTs are composable through their Connection Environment allowing both for the nesting and recombination of components. This approach allows us to re-conceptualise any sensors in the Physical Environment as virtually belonging to all associated composites of the DTs. Composites of the DTs can be produced to offer previously unavailable information via new virtual sensors [172] or virtual sensor networks [173,174].
For DTs to be composable, interoperability between components is the most important feature [169,175,176,177,178]. An example composition of the different components is shown across associated DTs in Figure 10. The Physical Environment contains sensory equipment within arbitrary boundaries to capture properties of the PT, which is immutable and equally anchors all overlapping DTs. The Data Environment is a heterogeneous composition of storage architectures that best befits the data requirements. The Analytical Environment is usually defined per task per DT. In addition, it may also have feedback loops to other DTs via the Data Environment. Outputs of the Analytical Environment are fed back into the Data Environment as virtual sensors to provide the information for display and interaction at a later time. The Virtual Environment forms a local Mirror World [137] that allows for the visual inspection of information in the DTs. The scope of the Virtual Environment is limited to the current composition of DTs and allows for access to and interaction with both the data from the Physical Environment and the data from the Analytical Environment. The Connection Environment manages access rights between components and ensures that users (both humans and other DTs) have the appropriate access rights to use the respective environments, addressing concerns regarding trust, security, and privacy [178,179,180,181,182,183,184,185,186,187,188,189,190,191,192].
In practice, DTs are applicable to all fields in which IoT has been applied [47,165,193], including industry [159,194], health care [43,44,166,195,196,197,198], urban management [48,199,200], vehicular application [201], and mobile end-user applications [202]. However, the composability of DTs suggests that, by extension, these applications should be composable, too. For example, DTs in industry, health care, and transportation could exchange vital information to improve their effectiveness, but, to the best of our knowledge, this has not been accomplished in practice. To understand how DTs of these varied systems could be composed in the context of true Smart Cities, we categorise each application by its system type and location type to form a taxonomy of DT contexts (see Table 1).
The system type of a DT is either open or closed [203]. Open systems are characterised by an exchange with the wider environment in which they are embedded, whereas closed systems do not directly exchange with the environment and are usually observed in isolation. In addition, closed systems usually supervise well defined processes, whereas open systems try to characterise ill defined processes. Obviously, no system is entirely isolated, but industrial processes or health applications can be considered relatively closed systems with usually well defined interactions with the environment. In contrast, cities and traffic [178] have no clear limit to what is entering from the surrounding environment, and congestion may be a side effect of this ill defined boundary.
The location type of a DT describes the relation between the unit of observation and the sensors as active or passive sensing [204]. In stationary systems, the sensors are fixed in the environment and observe throughput. In mobile systems, the sensors are moving around either with the observed unit or independent of it. They may also capture environmental conditions around the observed unit such as in human navigation [205] and health monitoring [206].
When composing different types of DTs, it is important to understand their characteristics. Mobile DTs may only temporarily compose with Stationary DTs, while Closed DTs often form units within Open DTs. Open Stationary DTs often form the overarching type of DT that contains other types of DTs, whereas Mobile Closed DTs usually describe atomic units in composed DTs. Typically, Mobile Closed DTs can be part of Mobile Open DTs when a person has a Body Area Network (BAN) [207] for health care applications and a Personal Area Network (PAN) [208] to interact with the environment. In this example, the PAN will provide a wide range of interactions that can be composed with the Open Stationary DT of a Smart City, whereas the BAN will only share limited information. Similarly, a Closed Stationary DT for a Smart Factory within Open Stationary DT of a Smart City will only share limited information.

4.2. Servicisation of Digital Twins

While the word DT suggests some form of unitary entity, it would be more accurate to understand DTs as context-aware representations of Things in the cloud composed of services that act upon the abstraction. Services can range from interacting with the environment to representing the sensed and forming the bedrock of ubiquitous computing [46]. Basic services again can be composed into higher-order services to address more complex issues.
With digital representation by the DT, it is possible to turn everything into a service—or servicise [209,210,211,212]. In other words, the customer buys the temporary use of a service associated with a product instead of the product itself. The cost of developing and maintaining a DT’s infrastructure justifies this approach because it is more efficient to reuse resources rather than to duplicate functionalities [212]. Therefore, DTs can be freely composed of services that conform to the requirements of the Connection Environment.
The three most common themes in “as a Service” (*aaS) are Infrastructure as a Service (IaaS), Platform as a Service (PaaS) and Software as a Service (SaaS). Newer types of services can be subsumed in these categories, including Sensing as a Services (SenaaS) [213,214], Semantics as a Service (SemaaS) [176], and Robots as a Service (RaaS) [215]. Notably, Infrastructure, Platform, and Software “as a Service” form an interesting interaction with the DT components because they can provide abstractions for particular components (see Figure 2). From an abstraction perspective, each replaces the lower layers of the stack with an interface that hides implementation complexity. In this section, we will use the example of weather services as part of a Smart City to explain how IaaS, PaaS, and SaaS can map to the components of a DT (see Figure 11).

4.2.1. Infrastructure as a Service (IaaS)

A DT requires the creation of a Physical Environment as the representation of the PT. The Physical Environment provides access to and measurement of the PT using infrastructure to collect raw data (see Figure 2). Creating and maintaining the DT infrastructure is the most physically laborious part of developing a DT. This severely limits the number of entities that can afford to create a DT on a city scale. However, provided as a service, it may be possible to finance the required infrastructure in aggregate and reuse it across multiple DTs. When creating a DT, the IaaS approach allows for the reuse of a service provider’s communication network, sensor network, and remote sensing units rather than having to build one’s own system from scratch. Therefore, IaaS enables one to have alternative competing and composable DTs based on the same or different data source(s). There are also more specialised intrastructure services that are available such as Sensing as a Service (SenaaS) [213,214], Remote Sensing as a Service (RSaaS) [216], and Robots as a Service (RaaS) [215].
In our weather service example (see Figure 11), the sensors, network, and remote sensing units in a Smart City could be servicised. The city government or a private provider could implement a policy to maintain such an IaaS to make digital services in a city more accessible, enabling the creation of different city DTs. This would allow new entrants to easily develop their own DTs that provide large-scale digital services without having to build physical hardware with redundant functionality [210].
At the same time, IaaS may raise some concerns regarding security. The IoT used for IaaS consists of billions of locatable, addressable, and readable devices [191]. There are three avenues by which security in IoT must be enforced: information, physical infrastructure, and management practices [189]. Security breaches from unexpected behaviour may arise from data collection and management via privacy protection problems, heterogeneous network authentication, and access control problems [189,190,191,217]. The fundamental challenges for all of these topics are questions of trust: how to give whom access for what purpose and for what duration. Over the last decade, the IEEE and IETF have laid the foundations to build a security infrastructure along the IEEE 802.15.4 standard [218,219] and the IETF LLN stack [208]. These efforts include sensor network security [184], wireless sensor network security [183], and dedicated secure protocols such as CoAP/Lithe [220,221], 6LowPSec [222], and RPL [223]. Ultimately, such as with the Internet, security will be an ongoing process that must be attended to at all times.

4.2.2. Platform as a Service (PaaS)

PaaS extends access to hardware by simplifying interfaces. PaaS usually packages digital representations into a Data Environment and thus reduces the complexity of accessing the data. Platforms vary largely and are usually tied to one of the five IoT application areas, industry [159,194], health care [43,44,166,195,196,197,198], urban management [48,199,200], vehicular applications [201], and end-user applications [202]. The concept of PaaS can be vague, and its use may sometimes be difficult to differentiate from IaaS or SaaS. In the context of a DT, a PaaS ought to provide only access to data as a Data Environment in contrast to an IaaS that collects raw data from the physical world as a Physical Environment and a SaaS that derives higher-order data from previously collected data as a Analytical Environment.
A DT probably cannot rely on a single PaaS to source all of the required data. More often, a DT will mix and match PaaS sources [224] to fulfil its task. The Data Environment is more complex than simply making data available and needs to be curated, indexed, and integrated into a knowledge graph based on context-awareness [156], ontologies [225,226,227], and middleware [228,229,230]. In reality, IaaS and PaaS are often mixed and sometimes offer a subset of services that should be associated with SaaS [231].
In our weather service example, the PaaS stores the information of the weather-related sensor data from sensing units throughout the city, including their spatiotemporal context. Specifically, the PaaS does not include any analytical components but offers a raw data source upon which higher-order data services may be built. A clear distinction is necessary between services to disambiguate their different responsibilities in the DT and ensure composability.

4.2.3. Software as a Service (SaaS)

According to the classical SaaS literature, SaaS is often exposed directly to an end user [202] and ought to follow a singular use case. However, in the context of the DT, SaaS could be mapped to both the Analytical Environment and the Virtual Environment. Following the “as a service” paradigm, we suggest limiting the use of SaaS to the Analytical Environment in DTs to reduce redundant functionality and to enable more servicisation in end-user applications (see Figure 2). Conventionally, one may consider SaaS as built upon PaaS, but the relationship between the Data and Analytical Environments may actually be bidirectional. The results of the analyses may be written back to the Data Environment before being forwarded to the Virtual Environment for user interaction. Thus, there may be an ambiguity that makes it difficult to separate SaaS from PaaS. The difference lies in whether one considers the results of an analysis to be newly computed as a SaaS or retrieved from memory as a PaaS.
In our weather service example, the classification of the predictive weather data to an end user could be qualified as SaaS if it is being calculated directly. However, if the predictive weather data are provided as a component of a Smart City environmental observation system, it would be qualified as a PaaS. In general, a SaaS conducts some analytical tasks and represents the Analytical Environment of DTs because it generates the models to mirror reality in the DT.

4.3. The Cognates of Digital Twins

DTs are a relatively new perspective that shares a common but amorphous root with a variety of other perspectives. While this root may be difficult to trace precisely and may differ from the intent of the author, it may have originated with various thinkers such as Wiener [136], Weiser [46], or even Tesla ([232], p. 122). Given this complexity, we borrowed the term “cognate” from linguistics [24], which refers to words with a formal resemblance or a common historical root. The confusing overlap among DT cognates is driven by developments across a diversity of research backgrounds such as information science, production engineering, data science, and computer science [8]. The different components of the DT are prioritised differently in each field, which results in similar yet slightly different meanings. For instance, the cognate term IoT is dominating the public sphere according to search word trends (see Figure 12). However, as hype gives way to pragmatism, simply having the physical infrastructure necessary for the creation of DTs is insufficient by itself. The two cognates IoT and DTs are often confused, but IoT may form the basis of DTs in a similar manner that the Internet forms the basis of the Web. Although it is too early to determine, search word trends may eventually place DTs or another cognate as the ultimate concept to describe these systems. The following sections clarify the nuances associated with DTs and their cognates across different disciplines.

4.3.1. Cyber–Physical Systems

A Cyber–Physical System (CPS) represents physical objects as virtual objects. Users can interact with either physical or virtual objects to change the state of both. A DT is a special instance of a CPS that includes a model of the PT for predictive operations.
The term DT is relatively new because the original idea of digital representation was first described in detail by Wiener [136] in the field of control theory. Later, the term Cyber–Physical Systems (CPSs) Theory [233,234] was coined to describe a physical object, homologous to a “Thing” [158] in the IoT context and the PT in the DT context. This physical object has a cyber correspondence, which is homologous to the “Digital Twin” [2]. A priori, CPSs have no predetermined (physical) scope and are an idealised form to describe possible interactions with entities [45,235]. Therefore, the CPS literature is rarely linked directly to DTs, remote sensing, or the IoT (but see [236,237]). CPSs are increasingly used to capture social data as well [238]. Their high level of abstraction makes CPSs a suitable paradigm to theorise concerning DTs, and findings in the CPS literature are very relevant to DTs.

4.3.2. (Wireless) Sensor (and Actuator) Networks

(Wireless) Sensor (and Actuator) Networks (WSN/SN/SAN/WSAN) provide sensing and actuating infrastructure that processes the signals to and from the physical world to enable a CPS or DT.
A DT cannot be conceptualised without having first digitised the PT. While the DT is usually presented as agnostic with respect to its underlying hardware, the DT cannot be constructed without it. Creating this representation requires sensing to comprehend the physical world and actuating to change the physical world [168,193]. Sensor networks ground any Cyber–Physical System in the physical world by providing readings thereof. For each device, questions of how to implement “autonomicity, scalability, energy efficiency, heterogeneity in terms of user equipment capabilities, complexity and environments” [239] need to be answered. Sensor networks had been researched long before the origination of the terms IoT or DT. However, contemporary research focuses more on how sensor networks are seen as a component of an IoT system [240].
In a sensor network context, there are active and passive devices that can be placed in the environment. Passive devices called Radio Frequency Identification (RFID) harvest energy from the transmission signal via induction to return a signal [241]. The signal can be used as an identification of an object in order to trigger a software-based service or to determine its location depending on the RFID reader [242]. Typically, RFID cannot return environmental information (but see [243]). Active devices are usually battery powered and have actual computational infrastructure, sensory equipment, and possibly actuators to interact with the environment. Whether active or passive devices are required depends on the task. A DT will often require a combination of both active and passive devices to efficiently represent the PT and the processes therein. A comparison of tasks and the technology that best captures them can be found in Table 2.
Without using the term DT, RFID implements a similar abstraction because a “physical object is accompanied by a rich, globally accessible virtual object that contains both current and historical information on that object’s physical properties, origin, ownership, and sensory context [241]”. RFID helps to embed access to knowledge in the environment and can make access to the DT context-driven [241]. For example, if users and environment are both tagged, the appropriate DT services can be called to facilitate the users’ task. At the same time, the system can track such interactions to help model activity for the DT. However, RFID tags could reveal sensitive information, and it is important to manage access control [244].
In contrast to passive devices, active devices usually have local computational power but may be constrained, which should be classified according to usability. Resource constraints on active devices can be grouped into six categories: battery, communication range, bandwidth, memory storage, processing capability, half duplex communication [178]. The IETF has classified devices based on hardware capabilities [245], and each class requires different operating systems (OSs) to function optimally [246,247]. They range from Class 0 devices to Class 2 devices. Class 0 devices have a tailored OS that provides no functionality beyond the required features. In contrast, Class 1 devices and above may already run a generalised if bare OS fulfilling the criteria of a small memory footprint, support for heterogeneous hardware, network connectivity, energy efficiency, real-time capabilities, security, and opportune scheduling policies [246,247]. Active devices can be fully integrated into DTs and can be managed from within the DT, whereas passive devices currently only provide anchor points for DTs.

4.3.3. Smart Objects

Smart Objects are physical objects that have embedded computational power to provide interactions for its virtual object to make it more interesting and may represent the fusion of the PT and DT.
A DT can be understood as a highly accurate model of a physical system [2,9,135], whereas a Smart Object (SO) offers interesting interactions with a particular object through embedded or remote computing power [7]. Both DTs and SOs represent a CPS [234]. It could be argued that the concept of an SO either incorporates or encapsulates the concept of a DT. However, a DT is usually required to mirror the role of an actual object in a larger context, whereas SOs only provide end users with interactions with the particular object [7]. Nonetheless, many SOs will represent the underlying physical object sufficiently closely to use them interchangeably with a DT. Assuming that an SO qualifies as a DT and that the SO can offer physical access to its digital representation, it can be considered a dual embodiment for both the object (i.e., PT) and the DT, providing a physical instance of the FTs paradigm (see Figure 9).
However, a DT may also describe an entity that is composed of multiple SOs, other Things, processes, and high-order entities. These entities may entail buildings, cities, or regions [4,248,249,250] modelled through remote sensing [54,129,251]. Moreover, a DT may also be formed of Things that are immaterial or at least incomprehensible in the physical world such as phantom traffic jams [252]. These can be considered purely virtual objects in contrast to virtualised objects [253], which would have a mapping back to a physical object. Composition allows for the formation of even-higher-order virtual objects that form the basis of a DT in a complex system such as the processes or aggregates representing a city [10,254], traffic flow [255], or production line [23]. Furthermore, a DT may also be produced exclusively by remote sensing without having local sensors attached to the PT [54]. A DT is thus a more encompassing concept than an SO.

4.3.4. Ubiquitous and Pervasive Computing/Ambient Intelligence

Ubiquitous Computing is a paradigm that expects computation to be available for all objects to form a CPS, whereas Pervasive Computing specifies the computation to occur at the object or close to it. SO and Ambient Intelligence are instances of Ubiquitous Computing in CPSs in which the sensing and actuating capacity is directly embedded into the object that is described by the corresponding DT.
Ubiquitous and Pervasive Computing are usually considered interchangeable. The former is an older term to link computations to any object in the physical world [46]. The latter is a younger term and is associated with placing computing power within (or close to) any object in the physical world [256]. Without loss of generality, we will only refer to Ubiquitous Computing in the remainder of the present paper. For a DT, any data collected in the physical world must be represented in “cyberspace” and thus requires computation. Many Ubiquitous Computing applications pre-date the IoT paradigm but have been subsumed by it (see [257]).
The closeness between the Thing and computations performed on sensor data from the Thing has often resulted in them being regarded as equivalent to each other [47]. Similarly, the concept of a DT is often used as a synonym for both the PT (homologous to the Thing) and the Physical Environment with WSAN (homologous to computations on sensor data from the Thing) from which the DT is created. Ubiquitous Computing offers us a lens to distinguish between PT and DT by considering the difference between computations on the state of an object and the actual object. SOs [7,258,259,260] are a typical example where the distinctions between PT, DT, sensors, actuators, and computations are blurred. SOs are expected to have awareness of their environment and have some model of the ongoing activities, policies, and processes around them. At the same time, SOs should represent functions, rules, and flows that allow for interactivity [7,258]. This formulation clearly indicates that the SOs have a Physical Environment in which they are embedded, an Analytical Environment to model their behaviour, a Data Environment to maintain their state, and a Virtual Environment in which their interactivity can be triggered. When the computations on the object or Thing become unrecognisable to humans because they are “woven into the background”, as predicted by Weiser [46], the term Ambient Intelligence becomes more relevant [42,43,44,45]. Making a clear distinction between a Thing, an SO, and Ambient Intelligence is probably neither possible nor desirable.

4.3.5. Internet of Things

The IoT is an amalgam of competing technologies with the goal of digitally representing the physical world. It interconnects WSANs to be combined into local CPSs that in turn compose larger DTs.
The idea behind IoT was first hyped [261,262,263], then overestimated, and then slowly became practical [264]. The term IoT was probably coined towards the end of the last millennia [265], but its roots are older, albeit unnamed. The functional similarity of CPSs [136] and Ubiquitous Computing [46], as well as stipulations from early thinkers such as Nikolai Tesla ([232], p. 122), make its historical provenance a question of choosing a remotely related cognate. Following Atzori and colleagues [232], we take the probable origin of the term “Internet of Things” by Ashton in 1999 [265] as a starting point to limit ourselves to cases where the general idea of IoT was already the goal of research.
Nearly anything remotely related to Information and Communication Technologies (ICT) [264] is included in modern IoT research [47,165,247,266,267,268,269,270,271,272,273,274,275,276,277] because IoT promises to be the invisible glue to digitally represent the world and offers new means of analysing and understanding [46]. In part, this may be due to the well known term Internet that immediately allows people to associate it with global information infrastructure [158]. On the other side, a Thing is vague enough to label any entity that can be distinguished by some measure [158]. The Intranet of Things in a WSAN has been transformed through ICT [277,278,279] and cloud-centric processing [268,280,281] to the IoT. This loose description is what allowed the IoT to soak up many kinds of ICT developments for the last decade (see Figure 12).
While the term IoT has recently become the most common term for the linkage between the virtual and the real, that was not always the case. Previously, a set of competing ideas (e.g., CPS, WSAN, and Ubiquitous Computing) co-created the ecosystem that today is generously labelled IoT and which laid the hardware foundation for DTs. These ideas were closely related but ultimately based on different assumptions because they originated from different fields. According to public perception, the terminology has settled on IoT even as the underlying technology is still quickly diversifying and expanding. With current trends, DTs appear to be a concept that is set to overtake the IoT in terms of public perception (see Figure 12).
Another issue is that IoT is a buzzword that is difficult to grasp because it is overloaded with meaning. Simply tapping into the literature is overwhelming and, at the same time, highly specific. The many technologies peddled under the umbrella-term IoT would be more distinctive if labelled specifically according to their function. While the technological umbrella-term is now well established, if ill defined, there is still a lingering sense that different fields of application are uniquely distinct and require special attention. Research and development is thus heavily directed towards five distinct real-world applications that promise the biggest gains: industrial application, urban management, health care, vehicular applications, and end-user applications [47,165,202]. This ingrained belief regarding the distinction leads to difficulties in synthesising common knowledge across applications. Overcoming this barrier could help the interdisciplinary fields of IoT and DTs in many ways. Not only would it enable learning from each other to clarify their theoretical backgrounds, but it would also make the use of IoT as a basis for DTs more accessible. Indeed, a clearly communicated and unified approach to IoT would lower the barriers of entry tremendously.
Weiser [46] postulated the goal of Ubiquitous Computing to make the computations disappear in the mundane of the environment. To obtain services that hide in plain sight, three categories of design considerations have been brought forward that need to be considered for IoT to work: fundamental properties of a Thing, network communication properties, and real-time interaction properties. The fundamental properties of a Thing include existence, sense of self, connectivity, interactivity, dynamicity, and sometimes environmental awareness [178,191,239]. Much research has also been conducted on properties to make network communication more efficient, including energy, latency, throughput, modular design for improved scalability, topology, security of information, safety [42,178,239], deployment, mobility, cost, efficient resource usage, heterogeneity, communication modality, infrastructure, coverage, network size, and lifetime [178,239]. Real-time requirements in industrial and healthcare settings have brought forth another set of properties that need to be considered to enable real-time interaction with IoT networks: predictable delays and latencies, congestion-free communication, accurate time synchronisation, real-time assurance, interoperability with existing infrastructure, and prioritised communication [178].
The historic development IoT and the evolution of definitions has been addressed in detail by [232]. The work identified three generations of IoT development with different foci that have resulted in the current diffusive definitions (see Table 3). An additional new generation is identified in this review paper that exemplifies the most current trends of DT research in IoT that were not reported in [232]. Notably, while these four generations focus on distinctive features, they are effectively contemporary because these generations are still actively researched. To make sense of these concurrent generations, it is best to understand each generation as a necessary condition to start work on the following generations.
The first generation is driven by the rise of RFID tags [265] to digitally distinguish Things. At this time, IoT was often taken to be synonymous with RFID. This generation is tightly coupled to Ubiquitous Computing [46], Machine-to-Machine (M2M), and CPSs as the RFID technologies ultimately derive from these research fields. However, by the late 2000s, most of these technologies had been tagged IoT. The second generation is driven by interconnecting Things instead of just centralising the information flow. It focuses on the network component of IoT devices and realistically brings the Internet into the IoT [282]. Control of information becomes more central as well, reinforcing the connections to its roots in CPSs [136]. The third generation is driven by a re-centralisation around cloud technologies that matured meanwhile [268,280,281]. This generation focuses on processing data and generating knowledge. The fourth generation is driven by a second wave of decentralisation due to time sensitivity of applications, fog computing [283], and the rise of DTs [2] and servicisation [212]. As questions of processing have been addressed, the newest generation brings more efficiency to IoT. IoT is often mentioned as a multiplier that makes other technologies more attractive [232], contributing to the diffusion of the term.

4.3.6. M2M

Machine-to-Machine (M2M) communication allows Things and network components to interact automatically with each other without human actions, thus enabling automated formation of CPSs or DTs on different scales.
The complexity of an IoT system underpinning a DT requires much communication between each involved Thing and the network infrastructure to support a control system [136,234,285]. These communication protocols have been aggregated and named M2M communication. M2M is an ongoing research field and is usually required for all other IoT-related and DT-related activities [286,287,288,289]. Essentially, M2M describes how IoT components interact with each other and can be understood as part of the implementation of the Connection Environment of a DT. As such, M2M, IoT, and DTs cannot be meaningfully separated, although they do not constitute the same concept. Before IoT became the defining word for ICT, M2M was a contender for being the umbrella-term for these technologies.
The number of devices connected to IoT today surpasses the number of devices manually connected to the Internet. Thus, it is impossible to manually manage the communication between devices by connecting them or maintaining them [178,290,291,292]. This aspect triggered the development of the M2M concept in the early 2000s [288,293] followed by standardisations such as oneM2M [153] in the 2010s. At the centre of M2M activities are the protocols that are used to exchange information between devices. Protocols describe predefined procedures of how to initiate, maintain, and reestablish communication [153]. While some protocols are high-level, they are often bound by the physical limitations of the communication capabilities of the constrained devices. The complexity of the classical standard Internet Protocol (IP) was quickly disqualified as a platform [294] because of the assumptions that the computing infrastructure was powerful, wired, and connected and that communication was a negligible overhead. However, constrained devices have been highly optimised in terms of energy consumption, size, and costs such that communication has become the major bottleneck [295].
The physical infrastructure to connect constrained devices to the network is usually referred to as physical/PHY [181], and the data link/MAC address is used to uniquely identify single units [296,297,298]. These terms originate from the OSI model and refer to the physical and data link layers between devices [299]. Originally, a link would be a copper or fibre cable. However, wireless transmission has adopted the same terminology, so PHY usually refers to some kind of antenna that uses a pattern of signals to encode information [251]. To enable a more efficient use of bandwidth and potentially allow for more devices, Cognitive Radios allow secondary users to use communication channels while primary users are idle [287,300,301]. The main trade-off for the PHY connection is between data rates, connection distances, and energy consumption. Currently, technology can be approximately grouped into four application areas: Body Area Networks (BAN), Personal Area Networks (PAN), Local Area Networks (LAN), and Wide Area Networks (WAN) (see Table 4 for an overview).
The large standardising bodies, IEEE and IETF, approach protocols in constrained environments based on expansion of the IEEE Standard 802.15 [315,316] and the IETF LLN protocol [294,295] to handle IP. In addition, there are technological niches that have not been fully covered by those standardising organisations such as the Low Power Wide Area Networks (LPWAN) where protocols vary and are highly incompatible, representing a low level of industry cooperation [309,310,311,312]. However, several consortia have developed around the different core technologies associated with each protocol, and only subsequent years will show how prevalent each WAN technology has become.
The commonality among all protocols is that they are simplified to move data from the constrained devices to some kind of translation unit that transfers the data into the IP protocol, making it widely available on the Internet [317,318,319]. This translation is necessary to cope with the reduced computing capacity of constrained devices as well as questions of how to transmit data and scale local IoT networks for use in a DT. A difficulty arises for these protocols because the translation unit of constrained devices not only sends data (such as with sensors) but may also receive data (such as with actuators). This aspect requires a bidirectional translation protocol that can uniquely address every node from the outside [320].

4.3.7. Computation-on-the-Wire/Dew/Mist/Fog/Cloud and Edge Computing

Computation-on-the-wire computes all necessary transformations on data at the best possible point in the network according to the subsidiarity principle to deliver Ubiquitous Computing for CPSs such as DTs.
The trend to locally compute, distribute, and store data in end-user devices and local clusters rather than centralised data centres [283,321] is a reaction to the third generation of IoT’s cloudification [232]. At the same time, the trend away from cloudification is the foundational moment for the fourth generation of IoT. Furthermore, sensors generally pose a trade-off between computational power, latency, cost, accuracy, energy consumption, and security [42,191,322,323]. To balance these trade-offs, additional computational power was introduced physically closer to the sensors, which also limited data transmission and reduced the load on the network. The placement of computational units in the physical world to provide the necessary computing capacity for Ubiquitous Computing produced a set of hydrologically inspired terms to describe different trade-offs. The division of remote computing into cloud [186,236,324,325,326,327,328], fog (also edge or grid) [150,151,283,322,327,328,329,330,331,332,333,334,335,336,337,338,339,340], mist [341,342], and dew [341,343,344,345,346] is an attempt to follow this principle (see Figure 13). While the analogy may have been taken a bit too far, this simile is quite adept at equating the distance from the observer to the water droplet and the distance from the device to the computation location. At the same time, the high level of differentiation and the unclear boundaries between these concepts—sometimes used interchangeably—makes the discussion more complicated. For example, fog, mist, and dew computing are often used interchangeably across papers, and edge and grid computing often fit partial definitions of the hydrological terminology.
Another, less metaphorically laden, view focuses on computation-on-the-wire to describe the same phenomena [347]. This perspective assumes that data need to be manipulated between the data generation point (i.e., a Thing at a physical location) and the data consumption point (i.e., a user inquiring about something related to the PT via its DT). To be the most cost-effective, the computation is required to take place as close to the generation and consumption points as sufficient and as far away as necessary. Therefore, the computation-on-the-wire concept is essentially the application of the subsidiarity principle. Taken from political science, subsidiarity implies that a central authority should only perform tasks that cannot be performed more locally [348]. The principle is most famously applied in the European Union and should be taken as the guiding principle for IoT as well. On a more abstract level, it implies that actions regarding an entity should be taken (a) as close as possible to an entity and at the same time (b) as far as necessary to successfully execute the action. Applied to computation-on-the-wire and DTs, this principle implies that computations for DTs should occur (a) as close to a device, a requester, or the PT as possible and (b) as far into the cloud as necessary to successfully compute the output. The concept elegantly joins the hydrological simile and the underlying trade-offs by applying the subsidiarity principle [348].

4.3.8. Remote Sensing

Remote sensing has two meanings. First, it provides observations recorded by a distant measurement device with some form of camera. Second, remote sensing refers to distant access to a local measurement device. Both can be processed to represent the Physical Environment around a CPS (such as a DT) and to possibly interact with the Physical Environment.
Remote sensing can be easily misunderstood as either exclusively meaning distant measurements based on some kind of camera [349] or using remote access to sensing units [350,351] to measure local properties from afar such as in agriculture [352], healthcare [353], Smart Cities [354], or everywhere [355]. Traditionally, remote sensing exclusively refers to camera-based, often airborne, approaches [349]. However, the secondary meaning of remote sensing as a shorthand for “remote access to local sensing units” has become prevalent across nearly all applications that do not apply traditional remote sensing. Whereas there is a clear difference in the meanings of remote and sensing in both interpretations, recent applications combining the two such as sensor networks on Unmanned Aerial Vehicles (UAV) often linked to ground stations and the Internet [356,357] may make the difference irrelevant in the long term. More recently, for the context of the DT, the precise sourcing of information from the Physical Environment is not important. Instead, the accuracy of the measurements must be reported to allow for a decision-making process that accounts for uncertainty. Throughout this review, we understand remote sensing in both its meanings as the main source for information from the Physical Environment. Furthermore, we assume that any form of traditional remote sensing can be made available in real-time as a virtual sensor in an IoT network [149,358,359].

5. The Fused Twins Concept

DTs require visualisation to unfold their potential. Naturally, it is the Virtual Environment [20,21,101] of a DT through which information be accessed. To make sense of the different forms of visualisations, we use the Reality–Virtuality Continuum [360], see Figure 14 to classify a broad range of human interface devices.
Commonly used implementations of DTs make use of visualisation types on both ends of the continuum which includes the Dashboard and Virtual Reality (VR)-based digital models. Dashboards [361] reside on the reality end of the spectrum because they display information independent of the physical world on screens which display numeric visualisations (note that reality here is the continuum instance in which the user is located and not the medium through which the application is experienced). The visualisation is not embedded directly in reality and is accessed in a non-immersive context. On the other end of the spectrum, remote monitoring [362,363] is located in the virtuality end and can be implemented both as VR and non-local AR. The information usually has a digital model that is displayed in the virtual space with no direct relations to the physical space in which the data are consumed.
The FTs concept captures the situation where the DT is queried from within the PT allowing for a fused in situ representation. A medium is used to connect the PT with the DT through the display of Situated Analytics [100]. Possible media range from tablet-based (AR) and HoloLens (MR) to SOs (AV). AR and MR devices can be used to overlay immersive analytics of the PT with the DT. In contrast, SOs [7] are interactables in the environment that allow users to manipulate the state of both the PT and the DT.
Despite the similiarites, FTs are different from situated visualisations [19]. FTs require a fully fleshed DT behind the situated visualisation that includes all five types of environments and the ability to compose the underlying DT with other DTs that may be spatially or logically connected. FTs use the situated visualisation as an entry point to interact with all systems that are connected to that particular point in the PT to provide access to the whole PT through the DT. In contrast, many situated visualisations focus on local applications such as smart switches for IoT devices [364]. However, at first sight, the situated visualisation is often indistinguishable from FTs for end users. Nonetheless, continuous interaction with the tool eventually reveals the limited functionality. Thus, any situated visualisation can be expanded to implement the FTs paradigm by integrating it into a DT model.
This section will outline the DT Requirements to implement the FTs paradigm. The particular features of the main environments are discussed, but the Data Environment and the Connection Environment are not discussed in detail. The Connection Environment has not been implemented as interoperability between DTs or their components is as of yet uncharted territory. Consequently, the Data Environment is usually a proprietary implementation that requires no further discussion here. An open standard for DTs would open up the research on the Connection Environment and the Data Environment. Lastly, while discussing the Virtual Environment, we reflect upon current situated visualisations that are close to implementing the full FTs concept.

5.1. The Physical Environment of the Digital Twin

The Physical Environment encapsulates the infrastructure to cast the Digital Shadow [152] of the PT (see Figure 8) into the Data Environment. When assembling an IoT or remote-sensing system for a DT, it is important to be able to determine the dimensions which matter most for the functioning of the DT and choose an architecture accordingly. Most research on architecture has been done in the context of IoT. However, as time progressed, architectures focused more on how to deal with the data and consequently today, these architectures are also used in DT cognates and DTs themselves.
International institutions have developed IoT reference architectures [365,366,367]. Architectures are usually presented as layers. However, multiple architectures exist with a similar number of layers but different functional goals. Going into detail of how each layer is supposed to work sometimes contradicts other architectures with the same number of layers (see Figure 15). On the one hand, there is no unified architecture due to the sheer endless number of application scenarios across different branches of research. On the other hand, architectures are sometimes compatible, and we provide a mapping of the different components in Figure 15. These architectures may appear as fully fleshed DTs at first glance but on closer inspection will reveal themselves to be mostly concerned with the Physical Environment and the Data Environment. However, depending on the overall goal of the architecture, the upper layers may actually contain features of the Analytical Environment and the Virtual Environment. IoT architectures are the basis for most current DT applications and will probably morph into fully fleshed DTs in the coming decade.
Architectures aim to be somewhat self-explanatory and ideally each layer covers exactly one functionality on transforming the data. Realistically, layers help conceptually to assign tasks to different parts of the real system. However, most implementations rely on a messier, less clean interaction that takes place between the components. To choose the layer, it is typical to consider aspects of security, energy consumption, data collection density (spatial and temporal), maintenance, availability (real time vs long-term evaluation), interactivity (explicit reaction versus implicit reaction), and heterogeneity [42,178,191,239]. Lastly, management of the infrastructure and security concerns relates to every layer and is often represented as vertical columns to the side of the layers [47,165,247], but they are not shown in Figure 15 for brevity. When comparing this to our DT design, it is those management layers that map to the Connection Environment that would ultimately allow to compose DTs given appropriate permissions.
While the number of architectures is infinite, we will focus on a few of the most common approaches in more detail here. The simplest architecture divides the world into two layers, the Physical layer and some abstract layer such as Digital [2] or Cyber [234,236]. Mapping these realms to our DT model, the Physical covers the Physical Environments and the Digital/Cyber covers the Data, Analytical, and Virtual Environments. The Connection Environment maps most closely to the Management and Security columns in previous architectures but the focus is more on interoperability while also satisfying management and security requirements. This architecture has no representation of how data are collected, transported, transformed, and attributed in the DT. Nonetheless, it is a powerful vision that is easily understood despite the technological intricacies necessary to make such a vision possible. This perspective strengthens ontological approaches to describe processes and qualities of objects in such a way that allows for interaction. This concept is often applied in Industrial IoT (IIoT) where it is usually used as an end-user perspective on a Service-Oriented Architecture (SOA) [154].
For practical implementations, often a three-layer architecture is used that describes the detailed work in the middle [165,368]. These architectures are already able to address how the DT arises and as such a third layer is brought to attention: the middleware. The physical world is sensed in a perception layer mapping to the Physical Environment. Then, the data are structured into semantically meaningful units in the middleware layer mapping to the Data Environment and the Analytical Environment and finally presented to end users in an application layer mapping to the Virtual Environment. This architecture belongs to the earlier models and is focused more on the process of how to create basic IoT infrastructure. The perception layer describes the process of data acquisition usually provided from sensors, actuators, and tags. The middleware layer aggregates the data and provides APIs to access necessary information. When comparing the three-layer model to the two-layer model, middleware is the hidden layer that converts the real world’s PT into its DT. However, most real systems stop short of complex analysis based on simulation and often limit themselves to presenting raw data or aggregates.
The application layer offers useful services on top of the data. In a sense, those services are offered in the Virtual Environment of the DT (see Figure 2), but they heavily rely on the Analytical Environment to produce interesting services. Naming conventions for the three layers vary, but their basic functionality remains the same such as in ETSI M2M [165], SENSEI [368]. Additional layers always focus on very specific features and mostly subdivide the Application layer for Service-oriented Architecture [154], Resource-oriented Architecture [162] business applications [155], or Context Awareness [156].

5.2. The Analytical Environment of the Digital Twin—Models and Services

The Analytical Environment is often considered to be the DT itself [152] because it simulates the PT behaviour. However, throughout this review paper, we have outlined more specific tasks for the Analytical Environment such as running simulations, producing aggregations, and providing services. We also have elaborated on why a DT is more than just a simple simulation with real-world data. Nonetheless, it is the Analytical Environment that usually gives DTs and their cognates the “smart” label such as in Smart City.
The word smart generally alludes to the superiority of a (connected, digitised) Thing compared to its dumb (unconnected, analogue) counterpart. Since connectivity and digitalisation on their own do not make anything smart but only provide accessibility, the smartness implies easy interaction with the Thing, new insights into the Thing, and new applications of the Thing. To attain any of these goals, models are required to represent reality, predict the future, and monitor ongoing processes. For DTs, this means that services [369,370] are provided on top of the data through the Analytical Environment to enrich a PT’s functionality and simplify interactions. In general, most Things will require a DT to become smart.
However, the term smart has been misused. The previous description of smart is applicable to any meaningful service. The use of “smart” as a prefix for services has proliferated to the point where it has become superfluous. Following the above argument, the prefix has been attached to anything that could have a DT or has implemented features that are indistinguishable from a DT. In the simplest case, a Smart Object provides awareness, representation, interactivity, and identity [7,258]. Smart Objects are concrete implementations of a DT (and even the FTs paradigm), but they limit the Thing to be a small-scale object with a fixed use. Smart Objects are also a form of data physicalisation [123] that provides the opportunity to interact with data beyond a visual medium.
As objects increase in size and become places, the terminology adapts to Smart Environment [321], Smart Home [371,372], Smart Office [372], Smart Museum [373], Smart City [4,10,48,199,200,254,324,371,374,375,376], and even Smart Regions [377] and Smart Communities [248,378]. Conceptually, these Smart Places [379] do not differ from Smart Objects [7] but only generalise them to sets of Things. This scaling aligns with the composability of DTs and allows for these concepts to be nested.
However, there is a difference to be made between physical Things and non-physical Things such as processes. The Thing in this context is a non-physical entity but interacts with physical entities in a defined way allowing to define smart processes such as Smart Governance [380], Smart Manufacturing [381,382], Smart Construction [383], Smart Health Care Services [384], Smart Mobility [385], and Smart Experiences (for touristic activities) [386].
With everything labelled smart, it becomes more meaningful to talk about the kind of actions a Smart Thing can provide. Interestingly, it is these interactions that are also sought after in the immersive analytics context [68]. We locate actions on a scale of how independently they can react to external input:
  • Sensing: Only produce data;
  • Reacting: Use threshold on collected data to trigger action;
  • Interacting: Use additional input beyond collected data to change behaviour;
  • Analysing: Use a model to interpret collected data when choosing actions;
  • Proposing: Use a decision model to propose actions based on data models;
  • Deciding: Autonomously perform all necessary action to interact with data.
The action space of Smart Things is constantly expanded and we cannot enumerate everything here. However, all of these developments can be understood as the Analytical Environment of a DT. Research has covered many approaches to provide the Analytical Environment needed to transform data into insights (see Table 5).

5.3. The Virtual Environment of the Digital Twin—Access, Interaction, and Representation through the Fused Twins Paradigm

The last environment we discuss provides the human–computer interaction [422,423] with the DT. When implementing the FTs paradigm for a DT, the main focus lies on AR interactions with the DT. AR is only beginning to arrive in the world of SOs, the IoT, remote sensing, and DTs but has already been accepted as a crucial feature in the near future [424]. AR is of interest for FTs as it explicitly supports embodied interaction: interaction that is not agnostic to its physical and social context [74,80]. AR can also reduce negative split-attention effects [117,119], and appropriate visualisations of data can lower cognitive load [73] and enhance cognition [72]. The expectation is that AR will offer a new avenue to simplify interaction with SOs, the IoT, and DTs [17]. The FTs paradigm is a formalisation of the interaction between AR and DTs in a situated context.
Because the combination of technologies is only emergent now, most published material towards this end is either a prototype [17,101,364,425,426] or a proposal for future research directions [427,428,429,430]. The main reason is that applying AR on top of the already complex combination of SO, IoT, remote sensing, and DT stacks requires many software systems to run smoothly that are currently only being prototyped.
The fusion of the the digital representation and the physical entity lies at the core of the FTs paradigm [17] but also of the more general Situated Analytics [19]. Interestingly, other research on the topic of Situated Analytics has often implemented FTs to some degree but without the explicit mentioning of the DT aspect [364,425,426]. A notable exception is the “Corsican Twins” [101] implementation of an FTs platform with an explicit DT powering the Situated Analytics. Short papers on future research directions in Situated Analytics clearly point towards FTs-like implementations [427,428,429,430]. We summarise ongoing and future research that is already covering a wide variety of topics (see Table 6). Some of the research explicitly mentions the DT approach or a Situated Analytics approach but often simply uses the available technology to demonstrate that their problem can be addressed [428,430,431], implicitly developing a DT and applying Situated Analytics.
Since the FTs concept in particular and DT concept in general is not well defined yet, we analysed the literature for traces of it. To construct Table 6, we analysed the papers for the presence of implementations of any of the five components. For the Physical Environment, we considered it as fully implemented (i.e., we gave it a check-mark (✓) in Table 6) if active sensor data collection was reported, or we considered it as partially implemented (i.e., we gave it a star (*) in Table 6) if passive sensors or markers were reported (e.g., QR code stickers). For the Data Environment, we considered it fully implemented when explicit data collection and storage was reported (e.g., SQL databases or BIM models), and we accepted partial implementation if data were streamed but not permanently stored. For the Analytical Environment, we considered it fully implemented if there was an element of only analysis, and we considered it partially implemented if simple data transformation was performed. For the Virtual Environment, we considered it fully implemented if some digital model was reported (e.g., 3D models of rooms or machines) or partially implemented if a digital representation could be inferred but was not explicitly reported (e.g., spatial representation of labels). For the Connection Environment, we considered it fully implemented if it was possible to add or compose elements in different environments such as adding new sensors, changing the Virtual Environment, or using different analysis and partially implemented if it was theoretically extendable or if the reports hinted at extendability of a single component (e.g., adding more sensors).
Across all partial implementations of the FTs paradigm, we find the recurrent theme of how to tightly couple the PT and the DT. All DTs require an effective representation of the PT to convey meaningful information to the user. For the built environment, this commonly involves Building Information Models (BIM) [17,428,430,437,446], or City Information Models (CIM), and the Geographic Information System (GIS) [428,433]. In a city context, usually, BIM/CIM is used for the local model, whereas GIS provides a global reference frame. However, if BIM/CIM is not available, remote sensing offers methods to produce placeholder models [440]. Both GIS and BIM/CIM have a long research tradition and have been used in a variety of ways in industry, academia, and governance. It could be argued that they provided DTs before the word was created. However, in the DT literature, a consensus has built up that they rather provide a digital model [152,171] and are at best the basis for a DT.
The reason that these opinions have developed is that BIM/CIM and GIS are often seen as a rather static representation of DTs and that, for dynamism, it is required to continuously measure from sources such as IoT [17,425,428,430]. These measurements are often called the Digital Shadow [152,171]. Additionally, the raw data provided by sensors is often classified as Big Data [447,448]. Visualising the data becomes nearly impossible without analytical tools. This highlights the strong interaction between the Analytical Environment and the Virtual Environment. At the same time, a DT is characterised by the bi-directional interaction between the PT and DT. Therefore, there is an expectation for actuators [449] to be able to be triggered from within the Virtual Environment to implement the full functionality of a DT.
Beyond an accurate model, data acquisition, and actions, FTs require a mapping between the virtual world and the physical world. Many approaches are available, and they commonly identify features in the physical world that can be used as anchors for the virtual world. These location services are based on different technologies such as Bluetooth-based [364], image-marker-based [101], vision-based [17,425], or signal-based [426]. Consequently, it is possible to display content in the Virtual Environment as an overlay in physical world. Depending on the medium, it is possible to interact with the augmented content through poses, gestures, mid-air interfaces, tangibles, touch, voice, face, or gaze [450,451,452]. Immersive and Situated Analytics [19,68] provide the research field in which most of the visualisations that will be used in FTs will be developed.
Implementing FTs requires some expansion of an existing DT. In the Virtual Environment, services for the digital model, the human–computer interaction, and the location model must be integrated to be able to effectively display the measurements from the IoT and remote sensing in situ. The services first locate data in relation to the BIM/CIM [437,446] and then locate the BIM/CIM in the physical world through Computer Vision and GIS [453]. The result is displayed in AR (see Figure 16) and offers Situated Analytics to the observer.

6. Discussion

To understand FTs, we have reviewed three distinct but connected literatures. First, this review has focused heavily on how to understand DTs today. The fuzziness around their definition has contributed to the continued vagueness of nearly any ICT being labelled or associated with IoT and DTs. However, with the rising practicability of DTs, no aspect of life will be beyond digitalisation for good and bad. Second, we looked at AR technology as a means to interact with information and Situated Analytics in particular to cover the interaction with a DT in situ of its PT. This combination is forming the basis to fuse the twins. Third, we explored the tightly connected topic of Smart Cities which often has been a canvas for testing ICT, IoT, DT, and AR applications. We expect the FTs concept to be applied often in this context, even under a different name or no name at all. We have enumerated a large number of FTs applications that partially qualify already today to give an idea of what is to come.

6.1. Smart Cities Discussion

Smart Cities are still a far way from being smart. In particular, because most “smart services” are not accessible. DTs offer an organisational paradigm that could make the complex Analytical Environment of cities accessible by flattening the hierarchy from the user perspective. The DT abstracts the heterogeneous technological structure of a city into semantically meaningful units that can be queried for information extracted from the underlying networks. At the same time, the DT also offers an opportunity to overcome the heterogeneity of sources from IoT to remote sensing without loss of accuracy. Lastly, DTs building on top of heterogeneous sources can represent abstract concepts that have no physical counterpart such as the livability and sustainability of a city [454] or the spatial movement of traffic jams [252].
A fully implemented DT would impact the management, planning, and use of a city. It would allow its economy to become more efficient, markets to be more transparent, processes to be explorable and problems possibly to be more addressable. If a DT of a city is made accessible to the general public, it could even serve as a participatory tool and integrate the city’s population. At the same time, concerns of privacy and security will not be easily resolved and a delicate balance between gains and losses has to be struck to realise DTs of cities.

6.2. Situated Analytics Discussion

AR combines real and virtual elements registered in three dimensions and provides interaction in real time. Despite decades of development, immersive analytics are still in their infancy, and no standard user interfaces have been established [125,455]. Recent reviews have further found that, for Situated Analytics, there is still a lack of an overarching definition, data representation is not unified, and the target audience is not defined [125]. Broad adoption is only now within reach, and the future will bring many new developments.
Interestingly, the description of how to assemble a Situated Analytics system maps closely to our layout of a DT (see Figure 7). There is a synergy between AR and the FTs concept: the FTs concept offers one of the most promising applications of AR, while AR supports an implementation of FTs that reduces split-attention effects and cognitive load. In this context, the FTs paradigm combining a DT of Smart Cities with Situated Analytics in AR integrates the physical and social aspects of the DT and thus aligns with the theories of embodied interaction, embodied cognition, situated cognition, and socially situated cognition [74,103,104,105]. Whether Situated Analytics live up to the expectation of our working hypothesis will require more rigorous research on the impact of embodiment, cognitive load and enhanced cognition. Nonetheless, we argue that the FTs concept may become the largest application of Situated Analytics to be seen, used, and even embodied by most people on the planet regardless of formal verification.

6.3. Digital Twins Discussion

We have reviewed the history of the DT to disambiguate the concept from its cognates. There is a fine line between DTs and CPSs, M2M, WSANs, SOs, Ubiquitous Computing, IoT, computation-on-the-wire, and Remote Sensing. In the context of this review, we have shown that most cognates are actually underlying technological approaches or perspectives that constitute components of the DT. The sole exception is a CPS which is a more general approach with the DT being a specialised instance thereof. We believe that DTs have the potential to subsume current cognates similar to how the IoT has been previously considered the most encompassing term in this area. At the same time, each cognate carries an important meaning that is worth maintaining, and we strongly recommend to understand these differences in order to gain a deeper understanding of DTs.
Most DTs “in the wild” are prototypes that were created as novel approaches to particular problems but whose theoretical background is underdeveloped beyond the definition of an ideal DT with similitude that is impractical for nearly all applications. Indeed, most research on practical DTs lacks theoretisation beyond the assembly of the system. Consequently, a debate arose on the differences between digital models, Digital Shadows, and DTs. This is beginning to change with the theoretical work of [8] which first developed a more sound differentiation of DT components. We follow the initial ideas from [23] but diverge in some details. We generalise the theoretisation into five general components (see Figure 2) that we rename the Physical, Data, Analytical, Virtual, and Connection Environments of the DT. Our theoretisation allows us to resolve the differences between models, shadows, and twins to partial implementations of the components of DTs.
Ideally, these environments work in isolation and are interoperable to enable the composition of DTs on the level of components as well as nesting. Furthermore, our definition covers most DTs and DT-like applications by assuming partial or implicit implementations of its components. Throughout this review, we have gathered evidence that cognates, other DT models, Smart Cities, and even Situated Analytics have theoretical correspondences to the components in our DT model.
At the same time, it is difficult to assemble the different environments. The servicisation of IoT has provided an interesting foundation upon which DTs could be quickly assembled with very little overhead. However, Iaas, PaaS, and SaaS are not sufficiently separate in the real world and, under current implementations, would require a Connection Environment that leans towards a hypervisor to compensate for the unreliable interfaces between services. It would be preferable to develop DTs and their components in a more standardised way via isolated and interoperable environments. If Iaas, PaaS, and SaaS were optimised to strictly map to DT components, then DT components “as a service” could democratise the use of DTs by enabling people to quickly bring DT services together. Two open questions that remain are whether the services should be provided in a centralised or decentralised manner and whether the services are provided as a government service or a market product. These services could be more accessible with an open standard for DTs that clearly define how the different components interact.

6.4. Fused Twins Discussion

Without appropriate human–computer interaction, the data in the DT will remain insurmountably complex and thus incomprehensible. The composability of DTs underpinned by interoperable IoT and remote sensing will enable a smoother interaction between people and the environment. In this context, the FTs concept will allow for an intuitive interface to access data, information, knowledge, and understanding where it is most needed and useful. The FTs Paradigm attains this with the AR/MR/AV interaction between the DT and the PT to identify opportunities for simple interactions in the complex data in situ where the data are generated and collected. It is paramount to align the user interfaces, data formats, general interoperability, and creative services that eventually converge towards real Smart Cities for its users.
The idea underlying FTs has been around for some time, albeit under a different or no specific name. DTs and their cognates Ambient Intelligence and Smart Objects, as well as sometimes the more general Things, can all be understood as different embodied FTs. At the same time, FTs and Situated Analytics are very similar, which opens the question of whether a new term adds clarification. We do believe that a new term is necessary because we are currently observing the confluence of two topics, and the new terms expand beyond each individual conceptualisation. These two topics are the aforementioned research in AR that does not necessarily rely on DTs (but may possibly implement FTs without mentioning it) and the expanding applications of DTs in nearly every research area and field. The necessity of visualising the DT, and the capability to do it in augmented space rather than on a display, ushers in a new era in which data can be made visually and spatially accessible at the time of creation and across the consumption. FTs play a special role in this context where the data are fused to its spatial context which differs from classical AR which merely uses the real-world space as an empty canvas devoid of any context (in most cases).
FTs are not unique to Smart Cities, but they are especially applicable in Smart Cities. Some application areas for FTs include healthcare and industry. For example, in healthcare, human DTs have been proposed to improve individual diagnosis and care. Monitoring patients with FTs allows practitioners to locate medically relevant information in the human body. This may be especially helpful in the context of surgery during which representations of bones and other tissues may be superimposed to facilitate the surgeon’s decision-making. In an industrial context, the inner workings of machines can be overlaid onto the real machines and augmented with analytical information. While healthcare and industrial application are most useful for specialists, FTs in a Smart City context may be an accessible form of FTs for the average citizen. For example, the Augmented Reality maps from Google already embed navigation information in the physical city and arguably employ a DT in the background, making it a simple and accessible example of FTs. Similarly, in the context of tourism, museums have created DTs of specific exhibits but could be expanded to include the entire exhibition to create an FTs. In the future, FTs could also be expended to communicate city planning initiatives to the public.

7. Conclusions

In this review paper, the relevant technologies and terminology behind the FTs concept for interacting with the DT of Smart Cities have been presented. We argue that FTs constitute a new phenomena beyond Situated Analytics because they explicitly use a DT and therefore stand for a richer interaction. Suitable examples of FTs applications across different literatures have been highlighted (see Table 6). Nonetheless, a stable and applicable out-of-the-box solution to fulfil the promised FTs for Smart Cities remains to be developed and validated to answer our working hypotheses. DTs also remain to be well specified, which may cause delays in developing FTs. Here, we provided a thorough review of DTs and their cognates to address open points and a concrete definition that can be applied in retrospect to nearly all DTs. Compared to previous research that employed Dashboards and other interfaces for Non-Situated Analytics, FTs represent a special application of the Virtual Environment of a DT to provide Situated Analytics. The key technologies behind the FTs Paradigm for DTs are IoT, remote sensing, BIM/CIM, GIS, and location services. The IoT, remote sensing, BIM/CIM, and GIS are used to assemble a DT and provide analytics, and location services are used to place the Situated Analytics in the PT. Each of these technologies on its own is well developed, but only now are they being combined into prototypes that can be classified as FTs. While many researchers have theorised about similar systems that could be built, few FTs full-fledged systems have been implemented. Given the plurality of fields that have converged on this general idea, we expect that the FTs paradigm will be a common theme in future technologies and may eventually be as common as a smartphone today.

Author Contributions

Conceptualisation, J.G.; methodology, J.G. and L.A.; validation, J.G. and T.T.; investigation, J.G., L.A., M.G.-M. and J.C.; resources, J.G., C.H., V.R.S. and R.W.S.; writing—original draft preparation, J.G.; writing—review and editing, J.G., T.T., L.A., J.C., M.G.-M., V.R.S., R.W.S. and C.H.; visualisation, J.G., M.G.-M. and J.C.; supervision, V.R.S., C.H. and R.W.S.; project administration, J.G., V.R.S. and C.H.; funding acquisition, T.T., V.R.S., C.H. and J.G. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by ETH Zurich grant number ETH-15 16-2 and J.G. was supported by an ETH Zurich Doc.Mobility Fellowship.

Data Availability Statement

Not applicable.

Acknowledgments

The authors thank Johanna Beyer and Zhu-Tian Chen for the feedback on the AR section. All diagrams were generated with templates from http://draw.io. And don’t forget your towel.

Conflicts of Interest

The authors declare no conflict of interest.

Abbreviations

The following abbreviations are used in this manuscript:
IoTInternet of Things
DTDigital Twin
PTPhysical Twin
FTsFused Twins
XRExtended Reality
CPSCyber–Physical System
UCUbiquitous Computing
SOSmart Object
WSNWireless Sensor Network
SNSensor Network
SANSensor and Actuator Network
WSANWireless Sensor and Actuator Networks
MRMixed Reality
ARAugmented Reality
VRVirtual Reality
AVAugmented Virtuality
IaaSInfrastructure as a Service
PaaSPlatform as a Service
SaaSSoftware as a Service
SenaaSSensing as a Services
SemaaSSemantics as a Service
RaaSRobots as a Service
RSaaSRemote Sensing as a Service
RFIDRadio Frequency Identification
OSOperating System
ICTInformation and Communication Technologies
M2MMachine-to-Machine
BANBody Area Networks
PANPersonal Area Networks
LANLocal Area Networks
WANWide Area Networks
LPWANLow Power Wide Area Networks
BIMBuilding Information Model
CIMCity Information Model
GISGeographic Information System
UAVUnmanned Aerial Vehicles

References

  1. Grieves, M. Digital Twin: Manufacturing excellence through virtual factory replication. White Paper 2014, 1, 1–7. [Google Scholar]
  2. Grieves, M.; Vickers, J. Digital Twin: Mitigating unpredictable, undesirable emergent behavior in complex systems. In Transdisciplinary Perspectives on Complex Systems; Springer: Berlin/Heidelberg, Germany, 2017; pp. 85–113. [Google Scholar]
  3. Deakin, M.; Al Waer, H. From intelligent to Smart Cities. Intell. Build. Int. 2011, 3, 140–152. [Google Scholar] [CrossRef]
  4. Su, K.; Li, J.; Fu, H. Smart City and the applications. In Proceedings of the 2011 International Conference on Electronics, Communications and Control (ICECC), Ningbo, China, 9–11 September 2011; IEEE: Piscataway, NJ, USA, 2011; pp. 1028–1031. [Google Scholar]
  5. Hung, M. Leading the iot, gartner insights on how to lead in a connected world, Technical report. Gart. Res. 2017, 1, 1–5. [Google Scholar]
  6. Deng, T.; Zhang, K.; Shen, Z.J.M. A systematic review of a Digital Twin city: A new pattern of urban governance toward Smart Cities. J. Manag. Sci. Eng. 2021, 6, 125–134. [Google Scholar] [CrossRef]
  7. Sánchez López, T.; Ranasinghe, D.C.; Harrison, M.; McFarlane, D. Adding sense to the Internet of Things. Pers. Ubiquitous Comput. 2012, 16, 291–308. [Google Scholar] [CrossRef]
  8. Tao, F.; Zhang, H.; Liu, A.; Nee, A. Digital Twin in Industry: State-of-the-Art. IEEE Trans. Ind. Inform. 2018, 15, 2405–2415. [Google Scholar] [CrossRef]
  9. Batty, M. Digital Twins. Environ. Plan. B 2018, 45, 817–820. [Google Scholar] [CrossRef]
  10. Kaur, M.J.; Mishra, V.P.; Maheshwari, P. The Convergence of Digital Twin, IoT, and Machine Learning: Transforming Data into Action. In Digital Twin Technologies and Smart Cities; Springer International Publishing: Cham, Switzerland, 2020; pp. 3–17. [Google Scholar] [CrossRef]
  11. Melgar, L.E.A.; Lalith, M.; Hori, M.; Ichimura, T.; Tanaka, S. A scalable workbench for large urban area simulations, comprised of resources for behavioural models, interactions and dynamic environments. In International Conference on Principles and Practice of Multi-Agent Systems; Springer: Cham, Switzerland, 2014; pp. 166–181. [Google Scholar]
  12. Jacob, S.; Aguilar, L.; Wijerathne, L.; Hori, M.; Ichimura, T.; Tanaka, S. Agent based modeling and simulation of tsunami triggered mass evacuation considering changes of environment due to earthquake and inundation. Proc. Civ. Soc. A2 (Appl. Mech.) 2014, 70, I_671–I_680. [Google Scholar] [CrossRef] [Green Version]
  13. Aguilar, L.; Lalith, M.; Ichimura, T.; Hori, M. On the performance and scalability of an HPC enhanced multi agent system based evacuation simulator. Procedia Comput. Sci. 2017, 108, 937–947. [Google Scholar] [CrossRef]
  14. Aguilar, L.; Wijerathne, L.; Jacob, S.; Hori, M.; Ichimura, T. Mass evacuation simulation considering detailed models: Behavioral profiles, environmental effects, and mixed-mode evacuation. Asia Pac. Manag. Rev. 2019, 24, 114–123. [Google Scholar] [CrossRef]
  15. Are Digital Twins the Future of Urban Planning? Available online: https://www.smartcitiesdive.com/news/are-digital-twins-the-future-of-urban-planning/609232/ (accessed on 20 September 2021).
  16. Lock, O.; Bednarz, T.; Pettit, C. HoloCity—Exploring the Use of Augmented Reality Cityscapes for Collaborative Understanding of High-Volume Urban Sensor Data. In Proceedings of the VRCAI ’19: The 17th International Conference on Virtual-Reality Continuum and Its Applications in Industry, Brisbane, QLD, Australia, 14–16 November 2019; Association for Computing Machinery: New York, NY, USA, 2019. [Google Scholar] [CrossRef]
  17. Grübel, J.; Gath-Morad, M.; Aguilar, L.; Thrash, T.; Sumner, R.W.; Hölscher, C.; Schinazi, V.R. Fused Twins: A Cognitive Approach to Augmented Reality Media Architecture. In Proceedings of the 5th MAB ’20: Media Architecture Biennale Conference, Amsterdam and Utrecht, The Netherlands, 28 June–2 July 2021. [Google Scholar]
  18. White, S.; Feiner, S. SiteLens: Situated visualization techniques for urban site visits. In Proceedings of the SIGCHI Conference on Human Factors in Computing Systems, Boston, MA, USA, 4–9 April 2009; Association for Computing Machinery: New York, NY, USA, 2009; pp. 1117–1120. [Google Scholar]
  19. Thomas, B.H.; Welch, G.F.; Dragicevic, P.; Elmqvist, N.; Irani, P.; Jansen, Y.; Schmalstieg, D.; Tabard, A.; ElSayed, N.A.; Smith, R.T.; et al. Situated Analytics. In Immersive Anal.; Springer: Heidelberg, Germany, 2018; Volume 11190, pp. 185–220. [Google Scholar]
  20. Grübel, J.; Thrash, T.; Hélal, D.; Sumner, R.W.; Hölscher, C.; Schinazi, V.R. The Feasibility of Dense Indoor LoRaWAN Towards Passively Sensing Human Presence. In Proceedings of the 2021 IEEE International Conference on Pervasive Computing and Communications (PerCom), Kassel, Germany, 22–26 March 2021; pp. 1–11. [Google Scholar] [CrossRef]
  21. Grübel, J.; Weibel, R.; Jiang, M.H.; Hölscher, C.; Hackman, D.A.; Schinazi, V.R. EVE: A Framework for Experiments in Virtual Environments. In Spatial Cogn. X; Springer: Heidelberg, Germany, 2016; pp. 159–176. [Google Scholar]
  22. Speicher, M.; Hall, B.D.; Nebeling, M. What is mixed reality? In Proceedings of the 2019 CHI Conference on Human Factors in Computing Systems, Glasgow, UK, 4–9 May 2019; pp. 1–15. [Google Scholar]
  23. Tao, F.; Zhang, M. Digital Twin shop-floor: A new shop-floor paradigm towards smart manufacturing. IEEE Access 2017, 5, 20418–20427. [Google Scholar] [CrossRef]
  24. Carroll, S.E. On cognates. Interlang. Stud. Bull. (Utrecht) 1992, 8, 93–119. [Google Scholar] [CrossRef]
  25. Tompson, T. Understanding the contextual development of Smart City initiatives: A pragmatist methodology. She Ji J. Des. Econ. Innov. 2017, 3, 210–228. [Google Scholar] [CrossRef]
  26. Vanolo, A. Smartmentality: The Smart City as disciplinary strategy. Urban Stud. 2014, 51, 883–898. [Google Scholar] [CrossRef]
  27. Mattern, S. A city is not a computer. Places J. 2017. [Google Scholar] [CrossRef]
  28. Luque-Ayala, A.; Marvin, S. Developing a critical understanding of smart urbanism. In Handbook of Urban Geography; Edward Elgar Publishing: Cheltenham, UK, 2019. [Google Scholar]
  29. Huyssen, A. Other Cities, Other Worlds: Urban Imaginaries in a Globalizing Age; Duke University Press: Durham, NC, USA, 2008. [Google Scholar]
  30. Newman, P.; Beatley, T.; Boyer, H. Resilient Cities: Responsing to Peak Oil and Climate Change; Taylor & Francis: London, UK, 2009. [Google Scholar]
  31. Haughton, G.; Hunter, C. Sustainable Cities; Routledge: Abingdon, UK, 2004. [Google Scholar]
  32. Castells, M. The Rise of the Network Society: The Information Age: Economy, Society, and Culture; John Wiley & Sons: Hoboken, NJ, USA, 2011; Volume 1. [Google Scholar]
  33. Townsend, A.M. Smart Cities: Big Data, Civic Hackers, and the Quest for a New Utopia; WW Norton & Company: New York, NY, USA, 2013. [Google Scholar]
  34. Gath-Morad, M.; Schaumann, D.; Zinger, E.; Plaut, P.O.; Kalay, Y.E. How smart is the Smart City? Assessing the impact of ICT on cities. In International Workshop on Agent Based Modelling of Urban Systems; Springer: Cham, Switzerland, 2016; pp. 189–207. [Google Scholar]
  35. Gottmann, J. Megalopolis and antipolis: The telephone and the structure of the city. In The Social Impact of the Telephone; MIT Press: Cambrige, UK, 1977; pp. 303–317. [Google Scholar]
  36. Toffler, A.; Alvin, T. The Third Wave; Bantam Books: New York, NY, USA, 1981. [Google Scholar]
  37. Negroponte, N. Being Digital. In The Information Technology; Forester, T., Ed.; Vintage Books: New York, NY, USA, 1995. [Google Scholar]
  38. Abler, R. What makes cities important. Bell Teleph. Mag. 1970, 49, 10–15. [Google Scholar]
  39. Picon, A. Smart Cities: A Spatialised Intelligence; John Wiley & Sons: Hoboken, NJ, USA, 2015. [Google Scholar]
  40. Goodchild, M.F. Citizens as sensors: The world of volunteered geography. GeoJournal 2007, 69, 211–221. [Google Scholar] [CrossRef] [Green Version]
  41. Soja, E.W. Postmodern Geographies: The Reassertion of Space in Critical Social Theory; Verso: London, UK, 1989. [Google Scholar]
  42. Flügel, C.; Gehrmann, V. Scientific Workshop 4: Intelligent Objects for the Internet of Things: Internet of Things–Application of Sensor Networks in Logistics. In Constructing Ambient Intelligence; Springer: Berlin/Heidelberg, Germany, 2009; pp. 16–26. [Google Scholar]
  43. Dohr, A.; Modre-Opsrian, R.; Drobics, M.; Hayn, D.; Schreier, G. The Internet of things for ambient assisted living. In Proceedings of the 2010 Seventh International Conference on Information Technology: New Generations, Las Vegas, NV, USA, 12–14 April 2010; IEEE: Piscataway, NJ, USA, 2010; pp. 804–809. [Google Scholar]
  44. Cubo, J.; Nieto, A.; Pimentel, E. A cloud-based Internet of Things platform for ambient assisted living. Sensors 2014, 14, 14070–14105. [Google Scholar] [CrossRef]
  45. Darwish, A.; Hassanien, A.E. Cyber physical systems design, methodology, and integration: The current status and future outlook. J. Ambient. Intell. Humaniz. Comput. 2018, 9, 1541–1556. [Google Scholar] [CrossRef]
  46. Weiser, M. The Computer for the 21st Century. Sci. Am. 1991, 265, 94–105. [Google Scholar] [CrossRef]
  47. Atzori, L.; Iera, A.; Morabito, G. The Internet of things: A survey. Comput. Netw. 2010, 54, 2787–2805. [Google Scholar] [CrossRef]
  48. Puiu, D.; Barnaghi, P.; Tönjes, R.; Kümper, D.; Ali, M.I.; Mileo, A.; Parreira, J.X.; Fischer, M.; Kolozali, S.; Farajidavar, N.; et al. Citypulse: Large scale data analytics framework for Smart Cities. IEEE Access 2016, 4, 1086–1108. [Google Scholar] [CrossRef]
  49. Nikitin, K.; Lantsev, N.; Nugaev, A.; Yakovleva, A. Data-Driven Cities: From Concept to Applied Solutions. PricewaterhouseCoopers (PwC). 2016. Available online: http://docplayer.net/50140321-From-concept-to-applied-solutions-data-driven-cities.html (accessed on 1 June 2022).
  50. Westminster City Council. Westminster City Council Annual Parking Review 2013/2014; Technical Report; Westminster City Council: Westminster, UK, 2014. [Google Scholar]
  51. Leleux, C.; Webster, W. Delivering smart governance in a future city: The case of Glasgow. Media Commun. 2018, 6, 163–174. [Google Scholar] [CrossRef]
  52. Sánchez, L.; Gutiérrez, V.; Galache, J.A.; Sotres, P.; Santana, J.R.; Casanueva, J.; Muñoz, L. SmartSantander: Experimentation and service provision in the smart city. In Proceedings of the 2013 16th International Symposium on Wireless Personal Multimedia Communications (WPMC), Atlantic City, NJ, USA, 24–27 June 2013; IEEE: Piscataway, NJ, USA, 2013; pp. 1–6. [Google Scholar]
  53. ITR Manager La Ville de Paris gère ses arbres avec des puces RFID. ITR. 2006. Available online: https://itrmanager.com/articles/59758/59758.html (accessed on 1 June 2022).
  54. Botín-Sanabria, D.M.; Mihaita, A.S.; Peimbert-García, R.E.; Ramírez-Moreno, M.A.; Ramírez-Mendoza, R.A.; de J. Lozoya-Santos, J. Digital Twin Technology Challenges and Applications: A Comprehensive Review. Remote Sens. 2022, 14, 1335. [Google Scholar] [CrossRef]
  55. Bilal, M.; Usmani, R.S.A.; Tayyab, M.; Mahmoud, A.A.; Abdalla, R.M.; Marjani, M.; Pillai, T.R.; Targio Hashem, I.A. Smart Cities data: Framework, applications, and challenges. In Handbook of Smart Cities; Springer: Cham, Switzerland, 2020; pp. 1–29. [Google Scholar]
  56. Foth, M. Handbook of Research on Urban Informatics: The Practice and Promise of the Real-Time City: The Practice and Promise of the Real-Time City; IGI Global: Hershey, PA, USA, 2008. [Google Scholar]
  57. Azuma, R.T. A survey of Augmented Reality. Presence Teleoperators Virtual Environ. 1997, 6, 355–385. [Google Scholar] [CrossRef]
  58. Milgram, P.; Takemura, H.; Utsumi, A.; Kishino, F. Augmented Reality: A class of displays on the reality-virtuality continuum. In Telemanipulator and Telepresence Technologies; International Society for Optics and Photonics: Bellingham, WA, USA, 1995; Volume 2351, pp. 282–292. [Google Scholar]
  59. Van Krevelen, D.; Poelman, R. Augmented Reality: Technologies, Applications, and Limitations Vrije Univ; Department of Computer Science: Amsterdam, The Netherlands, 2007. [Google Scholar]
  60. Grundhöfer, A.; Iwai, D. Recent advances in projection mapping algorithms, hardware and applications. In Computer Graphics Forum; Wiley Online Library: Hoboken, NJ, USA, 2018; Volume 37, pp. 653–675. [Google Scholar]
  61. Brock, A.M.; Chatain, J.; Park, M.; Fang, T.; Hachet, M.; Landay, J.A.; Cauchard, J.R. Flymap: Interacting with maps projected from a drone. In Proceedings of the 7th ACM International Symposium on Pervasive Displays, Munich, Germany, 6–8 June 2018; pp. 1–9. [Google Scholar]
  62. Carmigniani, J.; Furht, B.; Anisetti, M.; Ceravolo, P.; Damiani, E.; Ivkovic, M. Augmented Reality technologies, systems and applications. Multimed. Tools Appl. 2011, 51, 341–377. [Google Scholar] [CrossRef]
  63. Lavingia, K.; Tanwar, S. Augmented Reality and industry 4.0. In A Roadmap to Industry 4.0: Smart Production, Sharp Business and Sustainable Development; Springer: Berlin/Heidelberg, Germany, 2020; pp. 143–155. [Google Scholar]
  64. Scholz, J.; Smith, A.N. Augmented Reality: Designing immersive experiences that maximize consumer engagement. Bus. Horiz. 2016, 59, 149–161. [Google Scholar] [CrossRef]
  65. Stapleton, C.; Hughes, C.; Moshell, M.; Micikevicius, P.; Altman, M. Applying mixed reality to entertainment. Computer 2002, 35, 122–124. [Google Scholar] [CrossRef] [Green Version]
  66. Elmqaddem, N. Augmented Reality and Virtual Reality in education. Myth or reality? Int. J. Emerg. Technol. Learn. 2019, 14, 234–241. [Google Scholar] [CrossRef] [Green Version]
  67. Eckert, M.; Volmerg, J.S.; Friedrich, C.M. Augmented Reality in medicine: Systematic and bibliographic review. JMIR mHealth uHealth 2019, 7, e10967. [Google Scholar] [CrossRef]
  68. Marriott, K.; Schreiber, F.; Dwyer, T.; Klein, K.; Riche, N.H.; Itoh, T.; Stuerzlinger, W.; Thomas, B.H. Immersive Analytics; Springer: Berlin/Heidelberg, Germany, 2018; Volume 11190. [Google Scholar]
  69. Dwyer, T.; Marriott, K.; Isenberg, T.; Klein, K.; Riche, N.; Schreiber, F.; Stuerzlinger, W.; Thomas, B.H. Immersive analytics: An introduction. In Immersive Analytics; Springer: Berlin/Heidelberg, Germany, 2018; pp. 1–23. [Google Scholar]
  70. Genay, A.C.S.; Lécuyer, A.; Hachet, M. Being an Avatar for “Real”: A Survey on Virtual Embodiment in Augmented Reality. IEEE Trans. Vis. Comput. Graph. 2021. [Google Scholar] [CrossRef] [PubMed]
  71. Gervais, R.; Roo, J.S.; Hachet, M. Tangible viewports: Getting out of flatland in desktop environments. In Proceedings of the TEI’16: Tenth International Conference on Tangible, Embedded, and Embodied Interaction, Eindhoven, The Netherlands, 14–17 February 2016; pp. 176–184. [Google Scholar]
  72. Willett, W.; Aseniero, B.A.; Carpendale, S.; Dragicevic, P.; Jansen, Y.; Oehlberg, L.; Isenberg, P. Superpowers as Inspiration for Visualization. IEEE TVCG 2021, 2021. [Google Scholar] [CrossRef]
  73. Scaife, M.; Rogers, Y. External cognition: How do graphical representations work? Int. J. Hum. Comput. Stud. 1996, 45, 185–213. [Google Scholar] [CrossRef] [Green Version]
  74. Dourish, P. Where the Action Is: The Foundations of Embodied Interaction; MIT Press: Cambridge, MA, USA; London, UK, 2004. [Google Scholar]
  75. Kirsh, D. Embodied cognition and the magical future of interaction design. ACM Trans. Comput. Hum. Interact. (TOCHI) 2013, 20, 1–30. [Google Scholar] [CrossRef] [Green Version]
  76. Michaels, C.F. Affordances: Four points of debate. Ecol. Psychol. 2003, 15, 135–148. [Google Scholar] [CrossRef]
  77. Gibson, J.J. The Ecological Approach to Visual Perception: Classic Edition; Psychology Press: London, UK, 2014. [Google Scholar]
  78. Norman, D. The Design of Everyday Things: Revised and Expanded Edition; Basic Books: New York, NY, USA, 2013. [Google Scholar]
  79. Johnson-Glenberg, M.C.; Megowan-Romanowicz, C. Embodied science and mixed reality: How gesture and motion capture affect physics education. Cogn. Res. Princ. Implic. 2017, 2, 24. [Google Scholar] [CrossRef] [Green Version]
  80. Gervais, R.; Roo, J.S.; Frey, J.; Hachet, M. Introspectibles: Tangible interaction to foster introspection. arXiv 2016, arXiv:1603.04581. [Google Scholar]
  81. Gervais, R.; Frey, J.; Gay, A.; Lotte, F.; Hachet, M. Tobe: Tangible out-of-body experience. In Proceedings of the TEI’16: Tenth International Conference on Tangible, Embedded, and Embodied Interaction, Eindhoven, The Netherlands, 14–17 February 2016; pp. 227–235. [Google Scholar]
  82. Frey, J.; Gervais, R.; Fleck, S.; Lotte, F.; Hachet, M. Teegi: Tangible EEG interface. In Proceedings of the 27th Annual ACM Symposium on User Interface Software and Technology, Honolulu, HI, USA, 5–8 October 2014; pp. 301–308. [Google Scholar]
  83. Roo, J.S.; Gervais, R.; Frey, J.; Hachet, M. Inner garden: Connecting inner states to a mixed reality sandbox for mindfulness. In Proceedings of the 2017 CHI Conference on Human Factors in Computing Systems, Denver, CO, USA, 6–11 May 2017; pp. 1459–1470. [Google Scholar]
  84. Spiel, K. The bodies of tei–investigating norms and assumptions in the design of embodied interaction. In Proceedings of the Fifteenth International Conference on Tangible, Embedded, and Embodied Interaction, Salzburg, Austria, 14–17 February 2021; pp. 1–19. [Google Scholar]
  85. Chatain, J.; Sisserman, D.M.; Reichardt, L.; Fayolle, V.; Kapur, M.; Sumner, R.W.; Zünd, F.; Bermano, A.H. DigiGlo: Exploring the Palm as an Input and Display Mechanism through Digital Gloves. In Proceedings of the Annual Symposium on Computer-Human Interaction in Play, Virtual Event, Canada, 2–4 November 2020; pp. 374–385. [Google Scholar]
  86. Pei, S.; Chen, A.; Lee, J.; Zhang, Y. Hand Interfaces: Using Hands to Imitate Objects in AR/VR for Expressive Interactions. In Proceedings of the CHI Conference on Human Factors in Computing Systems, New Orleans, LA, USA, 30 April–5 May 2022; pp. 1–16. [Google Scholar]
  87. Thevin, L.; Brock, A.M. Augmented Reality for people with visual impairments: Designing and creating audio-tactile content from existing objects. In International Conference on Computers Helping People with Special Needs; Springer: Berlin/Heidelberg, Germany, 2018; pp. 193–200. [Google Scholar]
  88. Brulé, E.; Bailly, G.; Brock, A.; Gentès, A.; Jouffrais, C. An exploratory study of the uses of a multisensory map—With visually impaired children. Multimodal Technol. Interact. 2018, 2, 36. [Google Scholar] [CrossRef] [Green Version]
  89. Albouys-Perrois, J.; Laviole, J.; Briant, C.; Brock, A.M. Towards a multisensory Augmented Reality map for blind and low vision people: A participatory design approach. In Proceedings of the 2018 CHI Conference on Human Factors in Computing Systems, Montreal, QC, Canada, 21–26 April 2018; pp. 1–14. [Google Scholar]
  90. Melcer, E.F.; Isbister, K. Bridging the Physical Divide: A Design Framework for Embodied Learning Games and Simulations. In Proceedings of the 2016 CHI Conference Extended Abstracts on Human Factors in Computing Systems, CHI EA ’16, San Jose, CA, USA, 7–12 May 2016; Association for Computing Machinery: New York, NY, USA, 2016; pp. 2225–2233. [Google Scholar] [CrossRef]
  91. Ottmar, E.R.; Walkington, C.; Abrahamson, D.; Nathan, M.J.; Harrison, A.; Smith, C.; Embodied Mathematical Imagination and Cognition (EMIC) Working Group. North American Chapter of the International Group for the Psychology of Mathematics Education. In Proceedings of the 41st Annual Meeting of PME-NA, St. Louis, MO, USA, 17–19 November 2019. [Google Scholar]
  92. Chatain, J.; Ramp, V.; Gashaj, V.; Fayolle, V.; Kapur, M.; Sumner, B.; Magnenat, S. Grasping Derivatives: Teaching Mathematics through Embodied Interactions using Tablets and Virtual Reality. In Proceedings of the Interaction Design and Children (IDC’22), Braga, Portugal, 27–30 June 2022. [Google Scholar]
  93. Mueller, F.; Byrne, R.; Andres, J.; Patibanda, R. Experiencing the body as play. In Proceedings of the 2018 CHI Conference on Human Factors in Computing Systems, Montréal, QC, Canada, 21–26 April 2018; pp. 1–13. [Google Scholar]
  94. Höök, K.; Jonsson, M.P.; Ståhl, A.; Mercurio, J. Somaesthetic appreciation design. In Proceedings of the 2016 CHI Conference on Human Factors in Computing Systems, San Jose, CA, USA, 7–12 May 2016; pp. 3131–3142. [Google Scholar]
  95. Hook, K. Designing with the Body: Somaesthetic Interaction Design; MIT Press: Cambridge, MA, USA, 2018. [Google Scholar]
  96. Kilteni, K.; Groten, R.; Slater, M. The sense of embodiment in Virtual Reality. Presence Teleoperators Virtual Environ. 2012, 21, 373–387. [Google Scholar] [CrossRef] [Green Version]
  97. Nimcharoen, C.; Zollmann, S.; Collins, J.; Regenbrecht, H. Is that me?—Embodiment and body perception with an augmented reality mirror. In Proceedings of the 2018 IEEE International Symposium on Mixed and Augmented Reality Adjunct (ISMAR-Adjunct), Munich, Germany, 16–20 October 2018; IEEE: Piscataway, NJ, USA, 2018; pp. 158–163. [Google Scholar]
  98. Trajkova, M.; Alhakamy, A.; Cafaro, F.; Mallappa, R.; Kankara, S.R. Move your body: Engaging museum visitors with human-data interaction. In Proceedings of the 2020 CHI Conference on Human Factors in Computing Systems, Honolulu, HI, USA, 25–30 April 2020; Association for Computing Machinery: New York, NY, USA, 2020; pp. 1–13. [Google Scholar]
  99. Sicat, R.; Li, J.; Choi, J.; Cordeil, M.; Jeong, W.K.; Bach, B.; Pfister, H. DXR: A toolkit for building immersive data visualizations. IEEE Trans. Vis. Comput. Graph. 2018, 25, 715–725. [Google Scholar] [CrossRef] [Green Version]
  100. Fleck, P.; Calepso, A.S.; Hubenschmid, S.; Sedlmair, M.; Schmalstieg, D. RagRug: A Toolkit for Situated Analytics. IEEE Trans. Vis. Comput. Graph. 2022. [Google Scholar] [CrossRef]
  101. Prouzeau, A.; Wang, Y.; Ens, B.; Willett, W.; Dwyer, T. Corsican twin: Authoring in situ Augmented Reality visualisations in Virtual Reality. In Proceedings of the International Conference on Advanced Visual Interfaces, Salerno, Italy, 28 September–2 October 2020; pp. 1–9. [Google Scholar]
  102. Hubenschmid, S.; Zagermann, J.; Butscher, S.; Reiterer, H. Stream: Exploring the combination of spatially aware tablets with Augmented Reality head-mounted displays for immersive analytics. In Proceedings of the 2021 CHI Conference on Human Factors in Computing Systems, Yokohama, Japan, 8–13 May 2021; Association for Computing Machinery: New York, NY, USA, 2021; pp. 1–14. [Google Scholar]
  103. Wilson, A.D.; Golonka, S. Embodied cognition is not what you think it is. Front. Psychol. 2013, 4, 58. [Google Scholar] [CrossRef] [Green Version]
  104. Brown, J.S.; Collins, A.; Duguid, P. Situated cognition and the culture of learning. Educ. Res. 1989, 18, 32–42. [Google Scholar] [CrossRef]
  105. Smith, E.R.; Semin, G.R. Socially Situated Cognition: Cognition in its Social Context. In Advances in Experimental Social Psychology; Elsevier Academic Press: San Diego, CA, USA, 2004; Volume 36, pp. 53–117. [Google Scholar] [CrossRef]
  106. Merino, L.; Sotomayor-Gómez, B.; Yu, X.; Salgado, R.; Bergel, A.; Sedlmair, M.; Weiskopf, D. Toward agile situated visualization: An exploratory user study. In Extended Abstracts of the 2020 CHI Conference on Human Factors in Computing Systems, Honolulu, HI, USA, 25–30 April 2020; Association for Computing Machinery: New York, NY, USA, 2020; pp. 1–7. [Google Scholar]
  107. Turan, Z.; Meral, E.; Sahin, I.F. The impact of mobile Augmented Reality in geography education: Achievements, cognitive loads and views of university students. J. Geogr. High. Educ. 2018, 42, 427–441. [Google Scholar] [CrossRef]
  108. İbili, E. Effect of Augmented Reality Environments on Cognitive Load: Pedagogical Effect, Instructional Design, Motivation and Interaction Interfaces. Int. J. Progress. Educ. 2019, 15, 42–57. [Google Scholar] [CrossRef]
  109. Thees, M.; Kapp, S.; Strzys, M.P.; Beil, F.; Lukowicz, P.; Kuhn, J. Effects of Augmented Reality on learning and cognitive load in university physics laboratory courses. Comput. Hum. Behav. 2020, 108, 106316. [Google Scholar] [CrossRef]
  110. Buchner, J.; Buntins, K.; Kerres, M. A systematic map of research characteristics in studies on augmented reality and cognitive load. Comput. Educ. Open 2021, 2, 100036. [Google Scholar] [CrossRef]
  111. Buchner, J.; Buntins, K.; Kerres, M. The impact of Augmented Reality on cognitive load and performance: A systematic review. J. Comput. Assist. Learn. 2022, 38, 285–303. [Google Scholar] [CrossRef]
  112. Neumann, U.; Majoros, A. Cognitive, performance, and systems issues for Augmented Reality applications in manufacturing and maintenance. In Proceedings of the IEEE 1998 Virtual Reality Annual International Symposium (Cat. No. 98CB36180), Atlanta, GA, USA, 14–18 March 1998; IEEE: Piscataway, NJ, USA, 1998; pp. 4–11. [Google Scholar]
  113. Bonanni, L.; Lee, C.H.; Selker, T. Attention-based design of Augmented Reality interfaces. In CHI’05 Extended Abstracts on Human Factors in Computing Systems, Portland, OR, USA, 2–7 April 2005; Association for Computing Machinery: New York, NY, USA, 2005; pp. 1228–1231. [Google Scholar]
  114. Kim, S.; Dey, A.K. Simulated Augmented Reality windshield display as a cognitive mapping aid for elder driver navigation. In Proceedings of the SIGCHI Conference on Human Factors in Computing Systems, Boston, MA, USA, 4–9 April 2009; pp. 133–142. [Google Scholar]
  115. Altmeyer, K.; Kapp, S.; Thees, M.; Malone, S.; Kuhn, J.; Brünken, R. The use of Augmented Reality to foster conceptual knowledge acquisition in STEM laboratory courses—Theoretical background and empirical results. Br. J. Educ. Technol. 2020, 51, 611–628. [Google Scholar] [CrossRef]
  116. Altmeyer, K.; Malone, S.; Kapp, S.; Barz, M.; Lauer, L.; Thees, M.; Kuhn, J.; Peschel, M.; Sonntag, D.; Brünken, R. The effect of Augmented Reality on global coherence formation processes during STEM laboratory work in elementary school children. In Proceedings of the International Cognitive Load Theory Conference, Kingston, ON, Canada, 20–22 September 2021. [Google Scholar]
  117. Mayer, R.E.; Moreno, R. A split-attention effect in multimedia learning: Evidence for dual processing systems in working memory. J. Educ. Psychol. 1998, 90, 312. [Google Scholar] [CrossRef]
  118. Hornecker, E. The to-and-fro of sense making: Supporting users’ active indexing in museums. ACM Trans. Comput. Hum. Interact. (TOCHI) 2016, 23, 1–48. [Google Scholar] [CrossRef]
  119. Sweller, J.; Ayres, P.; Kalyuga, S. The split-attention effect. In Cognitive Load Theory; Springer: Berlin/Heidelberg, Germany, 2011; pp. 111–128. [Google Scholar]
  120. Kapp, S.; Thees, M.; Beil, F.; Weatherby, T.; Burde, J.P.; Wilhelm, T.; Kuhn, J. The Effects of Augmented Reality: A Comparative Study in an Undergraduate Physics Laboratory Course. In Proceedings of the CSEDU12: 12th International Conference on Computer Supported Education, Online, 2–4 May 2020; pp. 197–206. [Google Scholar]
  121. Tang, A.; Owen, C.; Biocca, F.; Mou, W. Comparative effectiveness of Augmented Reality in object assembly. In Proceedings of the SIGCHI Conference on Human Factors in Computing Systems, Ft. Lauderdale, FL, USA, 5–10 April 2003; pp. 73–80. [Google Scholar]
  122. Lobo, M.J.; Christophe, S. Opportunities and challenges for Augmented Reality Situated geographical visualization. ISPRS Ann. Photogramm. Remote. Sens. Spat. Inf. Sci. 2020, 4, 163–170. [Google Scholar] [CrossRef]
  123. Dragicevic, P.; Jansen, Y.; Vande Moere, A. Data physicalization. In Handbook of Human Computer Interaction; Springer: Cham, Switzerland, 2020; pp. 1–51. [Google Scholar]
  124. Wang, W.; Ge, Y.; Yu, H.; Lu, X.; Li, X.; Cheng, Y.; Sun, X. Intertwining History and Places: The Design of TongSEE Location-Based Augmented Reality Application for History Learning. In International Conference on Human-Computer Interaction; Springer: Berlin/Heidelberg, Germany, 2022; pp. 153–169. [Google Scholar]
  125. Bressa, N.; Korsgaard, H.; Tabard, A.; Houben, S.; Vermeulen, J. What’s the Situation with Situated Visualization? A Survey and Perspectives on Situatedness. IEEE Trans. Vis. Comput. Graph. 2021, 28, 107–117. [Google Scholar] [CrossRef]
  126. Willett, W.; Jansen, Y.; Dragicevic, P. Embedded data representations. IEEE Trans. Vis. Comput. Graph. 2016, 23, 461–470. [Google Scholar] [CrossRef] [Green Version]
  127. Lee, E.A. Cyber physical systems: Design challenges. In Proceedings of the 11th IEEE Symposium on Object Oriented Real-Time Distributed Computing (ISORC), Orlando, FL, USA, 5–7 May 2008; IEEE: Piscataway, NJ, USA, 2008; pp. 363–369. [Google Scholar]
  128. Rajkumar, R.; Lee, I.; Sha, L.; Stankovic, J. Cyber–physical systems: The next computing revolution. In Proceedings of the Design Automation Conference (DAC), 2010 47th ACM/IEEE, Anaheim, CA, USA, 13–18 June 2010; IEEE: Piscataway, NJ, USA, 2010; pp. 731–736. [Google Scholar]
  129. Lee, A.; Lee, K.W.; Kim, K.H.; Shin, S.W. A Geospatial Platform to Manage Large-Scale Individual Mobility for an Urban Digital Twin Platform. Remote Sens. 2022, 14, 723. [Google Scholar] [CrossRef]
  130. Nativi, S.; Mazzetti, P.; Craglia, M. Digital ecosystems for developing Digital Twins of the earth: The destination earth case. Remote Sens. 2021, 13, 2119. [Google Scholar] [CrossRef]
  131. Pettey, C. Prepare for the Impact of Digital Twins; Gartner: Stamford, CT, USA, 2017. [Google Scholar]
  132. Cearley, D.; Velosa, A.; Walker, M. Top 10 Strategic Technology Trends for 2017: Digital Twins; Gartner: Stamford, CT, USA, 2017. [Google Scholar]
  133. Cearley, D.; Velosa, A.; Kerremans, M.; Walker, M.; Burke, B. Top 10 Strategic Technology Trends for 2018: Digital Twins; Gartner: Stamford, CT, USA, 2018. [Google Scholar]
  134. Panetta, K. Gartner Top 10 Strategic Technology Trends for 2019; Gartner: Stamford, CT, USA, 2018. [Google Scholar]
  135. Boschert, S.; Rosen, R. Digital Twin—The simulation aspect. In Mechatronic Futures; Springer: Berlin/Heidelberg, Germany, 2016; pp. 59–74. [Google Scholar]
  136. Wiener, N. Cybernetics or Control and Communication in the Animal and the Machine; MIT Press: Cambridge, MA, USA, 1961; Volume 25. [Google Scholar]
  137. Gelernter, D. Mirror Worlds: Or the Day Software Puts the Universe in a Shoebox… How it Will Happen and What it Will Mean; Oxford University Press: Oxford, UK, 1993. [Google Scholar]
  138. Glaessgen, E.; Stargel, D. The Digital Twin paradigm for future NASA and US Air Force vehicles. In Proceedings of the 53rd AIAA/ASME/ASCE/AHS/ASC Structures, Structural Dynamics and Materials Conference 20th AIAA/ASME/AHS Adaptive Structures Conference 14th AIAA, Honolulu, HI, USA, 23–26 April 2012; p. 1818. [Google Scholar]
  139. Aguilar, L.; Lalith, M.; Ichimura, T.; Hori, M. Automatic evacuation management using a multi agent system and parallel meta-heuristic search. In International Conference on Principles and Practice of Multi-Agent Systems; Springer: Berlin/Heidelberg, Germany, 2016; pp. 387–396. [Google Scholar]
  140. Gabor, T.; Belzner, L.; Kiermeier, M.; Beck, M.T.; Neitz, A. A simulation-based architecture for smart cyber–physical systems. In Proceedings of the 2016 IEEE International Conference on Autonomic Computing (ICAC), Wuerzburg, Germany, 17–22 July 2016; IEEE: Piscataway, NJ, USA, 2016; pp. 374–379. [Google Scholar]
  141. Chowdhury, S.H.; Ali, F.; Jennions, I.K. A methodology for the experimental validation of an aircraft ECS Digital Twin targeting system level diagnostics. In Proceedings of the 11th Annual Conference of the PHM Society, Scottsdale, AZ, USA, 21–26 September 2019; pp. 21–26. [Google Scholar]
  142. The Economist. Digital Twins in Cockpits will Help Planes Look after Themselves. 2022. Available online: https://www.economist.com/science-and-technology/digital-twins-in-cockpits-will-help-planes-look-after-themselves/21809110 (accessed on 1 June 2022).
  143. Weyer, S.; Meyer, T.; Ohmer, M.; Gorecky, D.; Zühlke, D. Future modeling and simulation of CPS-based factories: An example from the automotive industry. IFAC-PapersOnLine 2016, 49, 97–102. [Google Scholar] [CrossRef]
  144. Rosen, R.; Von Wichert, G.; Lo, G.; Bettenhausen, K.D. About the importance of autonomy and Digital Twins for the future of manufacturing. IFAC-PapersOnLine 2015, 48, 567–572. [Google Scholar] [CrossRef]
  145. Vachálek, J.; Bartalskỳ, L.; Rovnỳ, O.; Šišmišová, D.; Morháč, M.; Lokšík, M. The Digital Twin of an industrial production line within the industry 4.0 concept. In Proceedings of the 2017 21st International Conference on Process Control (PC), Strbske Pleso, Slovakia, 6–9 June 2017; IEEE: Piscataway, NJ, USA, 2017; pp. 258–262. [Google Scholar]
  146. Canedo, A. Industrial IoT lifecycle via Digital Twins. In Proceedings of the 2016 International Conference on Hardware/Software Codesign and System Synthesis (CODES+ ISSS), Pittsburg, PA, USA, 2–7 October 2016; IEEE: Piscataway, NJ, USA, 2016; p. 1. [Google Scholar]
  147. Wikle, C.K.; Berliner, L.M. A Bayesian tutorial for data assimilation. Phys. D Nonlinear Phenom. 2007, 230, 1–16. [Google Scholar] [CrossRef]
  148. Reichle, R.H. Data assimilation methods in the Earth sciences. Adv. Water Resour. 2008, 31, 1411–1418. [Google Scholar] [CrossRef]
  149. Xiao, J.; Zhu, R.; Hu, R.; Wang, M.; Zhu, Y.; Chen, D.; Li, D. Towards real-time service from remote sensing: Compression of earth observatory video data via long-term background referencing. Remote Sens. 2018, 10, 876. [Google Scholar] [CrossRef] [Green Version]
  150. Shi, W.; Cao, J.; Zhang, Q.; Li, Y.; Xu, L. Edge computing: Vision and challenges. IEEE Internet Things J. 2016, 3, 637–646. [Google Scholar] [CrossRef]
  151. Hossain, S.K.A.; Rahman, M.A.; Hossain, M.A. Edge computing framework for enabling situation awareness in IoT based Smart City. J. Parallel Distrib. Comput. 2018, 122, 226–237. [Google Scholar] [CrossRef]
  152. Schluse, M.; Priggemeyer, M.; Atorf, L.; Rossmann, J. Experimentable Digital Twins—Streamlining simulation-based systems engineering for industry 4.0. IEEE Trans. Ind. Inform. 2018, 14, 1722–1731. [Google Scholar] [CrossRef]
  153. Swetina, J.; Lu, G.; Jacobs, P.; Ennesser, F.; Song, J. Toward a standardized common M2M service layer platform: Introduction to oneM2M. IEEE Wirel. Commun. 2014, 21, 20–26. [Google Scholar] [CrossRef]
  154. Xu, D.L.; He, W.; Li, S. Internet of things in industries: A survey. IEEE Trans. Ind. Inform. 2014, 10, 2233–2243. [Google Scholar] [CrossRef]
  155. Lee, J.; Bagheri, B.; Kao, H.A. A cyber–physical systems architecture for industry 4.0-based manufacturing systems. Manuf. Lett. 2015, 3, 18–23. [Google Scholar] [CrossRef]
  156. Knappmeyer, M.; Kiani, S.L.; Reetz, E.S.; Baker, N.; Tonjes, R. Survey of context provisioning middleware. IEEE Commun. Surv. Tutor. 2013, 15, 1492–1519. [Google Scholar] [CrossRef] [Green Version]
  157. Guinard, D.; Trifa, V.; Mattern, F.; Wilde, E. From the Internet of things to the web of things: Resource-oriented architecture and best practices. In Architecting the Internet of Things; Springer: Berlin/Heidelberg, Germany, 2011; pp. 97–129. [Google Scholar]
  158. Huang, Y.; Li, G. Descriptive models for Internet of Things. In Proceedings of the 2010 International Conference on Intelligent Control and Information Processing, Dalian, China, 13–15 August 2010; IEEE: Piscataway, NJ, USA, 2010; pp. 483–486. [Google Scholar]
  159. Spiess, P.; Karnouskos, S.; Guinard, D.; Savio, D.; Baecker, O.; De Souza, L.M.S.; Trifa, V. SOA-based integration of the Internet of things in enterprise services. In Proceedings of the 2009 IEEE International Conference on Web Services, Los Angeles, CA, USA, 6–10 July 2009; IEEE: Piscataway, NJ, USA, 2009; pp. 968–975. [Google Scholar]
  160. Guinard, D.; Trifa, V.M.; Wilde, E. Architecting a Mashable Open World Wide Web of Things; Technical Report; Department of Computer Science, Swiss Federal Institute of Technology Zurich: Zürich, Switzerland, 2010. [Google Scholar]
  161. Guinard, D.; Trifa, V.; Karnouskos, S.; Spiess, P.; Savio, D. Interacting with the soa-based Internet of things: Discovery, query, selection, and on-demand provisioning of web services. IEEE Trans. Serv. Comput. 2010, 3, 223–235. [Google Scholar] [CrossRef] [Green Version]
  162. Guinard, D.; Trifa, V.; Wilde, E. A resource oriented architecture for the web of things. In Proceedings of the 2010 Internet of Things (IOT), Tokyo, Japan, 29 November–1 December 2010; IEEE: Piscataway, NJ, USA, 2010; pp. 1–8. [Google Scholar]
  163. Kindberg, T.; Barton, J.; Morgan, J.; Becker, G.; Caswell, D.; Debaty, P.; Gopal, G.; Frid, M.; Krishnan, V.; Morris, H.; et al. People, places, things: Web presence for the real world. Mob. Netw. Appl. 2002, 7, 365–376. [Google Scholar] [CrossRef]
  164. Blackstock, M.; Lea, R. IoT mashups with the WoTKit. In Proceedings of the 2012 3rd IEEE International Conference on the Internet of Things, Wuxi, China, 24–26 October 2012; IEEE: Piscataway, NJ, USA, 2012; pp. 159–166. [Google Scholar]
  165. Borgia, E. The Internet of Things vision: Key features, applications and open issues. Comput. Commun. 2014, 54, 1–31. [Google Scholar] [CrossRef]
  166. Islam, S.M.R.; Kwak, D.; Kabir, M.D.H.; Hossain, M.; Kwak, K.S. The Internet of things for health care: A comprehensive survey. IEEE Access 2015, 3, 678–708. [Google Scholar] [CrossRef]
  167. Wang, S.; Wan, J.; Zhang, D.; Li, D.; Zhang, C. Towards smart factory for industry 4.0: A self-organized multi-agent system with Big Data based feedback and coordination. Comput. Netw. 2016, 101, 158–168. [Google Scholar] [CrossRef] [Green Version]
  168. Akyildiz, I.F.; Su, W.; Sankarasubramaniam, Y.; Cayirci, E. Wireless sensor networks: A survey. Comput. Netw. 2002, 38, 393–422. [Google Scholar] [CrossRef] [Green Version]
  169. Schleich, B.; Anwer, N.; Mathieu, L.; Wartzack, S. Shaping the Digital Twin for design and production engineering. CIRP Ann. 2017, 66, 141–144. [Google Scholar] [CrossRef] [Green Version]
  170. Moreno, A.; Velez, G.; Ardanza, A.; Barandiaran, I.; de Infante, Á.R.; Chopitea, R. Virtualisation process of a sheet metal punching machine within the Industry 4.0 vision. Int. J. Interact. Des. Manuf. (IJIDeM) 2017, 11, 365–373. [Google Scholar] [CrossRef]
  171. Sepasgozar, S.M. Differentiating Digital Twin from Digital Shadow: Elucidating a paradigm shift to expedite a smart, sustainable built environment. Buildings 2021, 11, 151. [Google Scholar] [CrossRef]
  172. Madria, S.; Kumar, V.; Dalvi, R. Sensor cloud: A cloud of virtual sensors. IEEE Softw. 2014, 31, 70–77. [Google Scholar] [CrossRef]
  173. Ocean, M.J.; Bestavros, A.; Kfoury, A.J. snBench: Programming and virtualization framework for distributed multitasking sensor networks. In Proceedings of the 2nd International Conference on Virtual Execution Environments, Ottawa, ON, Canada, June 14–16 2006; pp. 89–99. [Google Scholar]
  174. Khan, I.; Belqasmi, F.; Glitho, R.; Crespi, N.; Morrow, M.; Polakos, P. Wireless sensor network virtualization: A survey. IEEE Commun. Surv. Tutor. 2016, 18, 553–576. [Google Scholar] [CrossRef] [Green Version]
  175. Bröring, A.; Schmid, S.; Schindhelm, C.K.; Khelil, A.; Kabisch, S.; Kramer, D.; Le Phuoc, D.; Mitic, J.; Anicic, D.; Teniente López, E. Enabling IoT ecosystems through platform interoperability. IEEE Softw. 2017, 34, 54–61. [Google Scholar] [CrossRef] [Green Version]
  176. Desai, P.; Sheth, A.; Anantharam, P. Semantic gateway as a service architecture for iot interoperability. In Proceedings of the 2015 IEEE International Conference on Mobile Services, New York, NY, USA, 27 June–2 July 2015; IEEE: Piscataway, NJ, USA, 2015; pp. 313–319. [Google Scholar]
  177. Al-Fuqaha, A.; Khreishah, A.; Guizani, M.; Rayes, A.; Mohammadi, M. Toward better horizontal integration among IoT services. IEEE Commun. Mag. 2015, 53, 72–79. [Google Scholar] [CrossRef]
  178. Raza, M.; Aslam, N.; Le-Minh, H.; Hussain, S.; Cao, Y.; Khan, N.M. A Critical Analysis of Research Potential, Challenges, and Future Directives in Industrial Wireless Sensor Networks. IEEE Commun. Surv. Tutor. 2018, 20, 39–95. [Google Scholar] [CrossRef]
  179. Das, A.K.; Zeadally, S.; He, D. Taxonomy and analysis of security protocols for Internet of Things. Future Gener. Comput. Syst. 2018, 89, 110–125. [Google Scholar] [CrossRef]
  180. Keoh, S.L.; Kumar, S.S.; Tschofenig, H. Securing the Internet of things: A standardization perspective. IEEE Internet Things J. 2014, 1, 265–275. [Google Scholar] [CrossRef]
  181. Granjal, J.; Monteiro, E.; Silva, J.S. Security for the Internet of things: A survey of existing protocols and open research issues. IEEE Commun. Surv. Tutor. 2015, 17, 1294–1312. [Google Scholar] [CrossRef]
  182. Kouicem, D.E.; Bouabdallah, A.; Lakhlef, H. Internet of Things Security: A top-down survey. Comput. Netw. 2018, 141, 199–221. [Google Scholar] [CrossRef] [Green Version]
  183. Zhou, Y.; Fang, Y.; Zhang, Y. Securing wireless sensor networks: A survey. IEEE Commun. Surv. Tutor. 2008, 10, 6–28. [Google Scholar] [CrossRef]
  184. Chen, X.; Makki, K.; Yen, K.; Pissinou, N. Sensor network security: A survey. IEEE Commun. Surv. Tutor. 2009, 11, 52–73. [Google Scholar] [CrossRef] [Green Version]
  185. Zhang, Z.K.; Cho, M.C.Y.; Wang, C.W.; Hsu, C.W.; Chen, C.K.; Shieh, S. IoT security: Ongoing challenges and research opportunities. In Proceedings of the 2014 IEEE 7th International Conference on Service-Oriented Computing and Applications, Matsue, Japan, 17–19 November 2014; IEEE: Piscataway, NJ, USA, 2014; pp. 230–234. [Google Scholar]
  186. Singh, J.; Pasquier, T.; Bacon, J.; Ko, H.; Eyers, D. Twenty Security Considerations for Cloud-Supported Internet of Things. IEEE Internet Things J. 2016, 3, 269–284. [Google Scholar] [CrossRef] [Green Version]
  187. Heer, T.; Garcia-Morchon, O.; Hummen, R.; Keoh, S.L.; Kumar, S.S.; Wehrle, K. Security Challenges in the IP-based Internet of Things. Wirel. Pers. Commun. 2011, 61, 527–542. [Google Scholar] [CrossRef] [Green Version]
  188. Yaqoob, I.; Ahmed, E.; ur Rehman, M.H.; Ahmed, A.I.A.; Al-garadi, M.A.; Imran, M.; Guizani, M. The rise of ransomware and emerging security challenges in the Internet of Things. Comput. Netw. 2017, 129, 444–458. [Google Scholar] [CrossRef]
  189. Zhao, K.; Ge, L. A survey on the Internet of things security. In Proceedings of the 2013 Ninth International Conference on Computational Intelligence and Security, Emeishan, China, 14–15 December 2013; IEEE: Piscataway, NJ, USA, 2013; pp. 663–667. [Google Scholar]
  190. Sarma, A.C.; Girão, J. Identities in the future Internet of things. Wirel. Pers. Commun. 2009, 49, 353–363. [Google Scholar] [CrossRef]
  191. Roman, R.; Najera, P.; Lopez, J. Securing the Internet of things. Computer 2011, 44, 51–58. [Google Scholar] [CrossRef] [Green Version]
  192. Sadeghi, A.R.; Wachsmann, C.; Waidner, M. Security and privacy challenges in industrial Internet of things. In Proceedings of the 2015 52nd ACM/EDAC/IEEE Design Automation Conference (DAC), San Francisco, CA, USA, 8–12 June 2015; IEEE: Piscataway, NJ, USA, 2015; pp. 1–6. [Google Scholar]
  193. Akyildiz, I.F.; Su, W.; Sankarasubramaniam, Y.; Cayirci, E. A survey on sensor networks. IEEE Commun. Mag. 2002, 40, 102–114. [Google Scholar] [CrossRef] [Green Version]
  194. Bi, Z.; Da Xu, L.; Wang, C. Internet of things for enterprise systems of modern manufacturing. IEEE Trans. Ind. Inform. 2014, 10, 1537–1546. [Google Scholar]
  195. Doukas, C.; Maglogiannis, I. Bringing IoT and cloud computing towards pervasive healthcare. In Proceedings of the 2012 Sixth International Conference on Innovative Mobile and Internet Services in Ubiquitous Computing, Palermo, Italy, 4–6 July 2012; IEEE: Piscataway, NJ, USA, 2012; pp. 922–926. [Google Scholar]
  196. Amendola, S.; Lodato, R.; Manzari, S.; Occhiuzzi, C.; Marrocco, G. RFID technology for IoT-based personal healthcare in smart spaces. IEEE Internet Things J. 2014, 1, 144–152. [Google Scholar] [CrossRef]
  197. Catarinucci, L.; De Donno, D.; Mainetti, L.; Palano, L.; Patrono, L.; Stefanizzi, M.L.; Tarricone, L. An IoT-aware architecture for smart healthcare systems. IEEE Internet Things J. 2015, 2, 515–526. [Google Scholar] [CrossRef]
  198. Kumari, A.; Tanwar, S.; Tyagi, S.; Kumar, N. Fog computing for Healthcare 4.0 environment: Opportunities and challenges. Comput. Electr. Eng. 2018, 72, 1–13. [Google Scholar] [CrossRef]
  199. Vlacheas, P.; Giaffreda, R.; Stavroulaki, V.; Kelaidonis, D.; Foteinos, V.; Poulios, G.; Demestichas, P.; Somov, A.; Biswas, A.R.; Moessner, K. Enabling Smart Cities through a cognitive management framework for the Internet of things. IEEE Commun. Mag. 2013, 51, 102–111. [Google Scholar] [CrossRef]
  200. Krylovskiy, A.; Jahn, M.; Patti, E. Designing a Smart City Internet of things platform with microservice architecture. In Proceedings of the 2015 3rd International Conference on Future Internet of Things and Cloud, Rome, Italy, 24–26 August 2015; IEEE: Piscataway, NJ, USA, 2015; pp. 25–30. [Google Scholar]
  201. Al-Turjman, F.; Lemayian, J.P. Intelligence, security, and vehicular sensor networks in Internet of things (IoT)-enabled smart-cities: An overview. Comput. Electr. Eng. 2020, 87, 106776. [Google Scholar] [CrossRef]
  202. Kamilaris, A.; Pitsillides, A. Mobile phone computing and the Internet of things: A survey. IEEE Internet Things J. 2016, 3, 885–898. [Google Scholar] [CrossRef]
  203. Bertalanffy, L.V. General System Theory: Foundations, Development, Applications; G. Braziller: New York, NY, USA, 1969. [Google Scholar]
  204. Moussaïd, M.; Schinazi, V.R.; Kapadia, M.; Thrash, T. Virtual Sensing and Virtual Reality: How New Technologies Can Boost Research on Crowd Dynamics. Front. Robot. AI 2018, 5, 82. [Google Scholar] [CrossRef] [Green Version]
  205. Kiefer, P.; Giannopoulos, I.; Raubal, M. Where am I? Investigating map matching during self-localization with mobile eye tracking in an urban environment. Trans. GIS 2014, 18, 660–686. [Google Scholar] [CrossRef]
  206. Petäjäjärvi, J.; Mikhaylov, K.; Yasmin, R.; Hämäläinen, M.; Iinatti, J. Evaluation of LoRa LPWAN technology for indoor remote health and wellbeing monitoring. Int. J. Wirel. Inf. Netw. 2017, 24, 153–165. [Google Scholar] [CrossRef] [Green Version]
  207. Singh, S.; Prasad, D. Wireless Body Area Network (WBAN): A review of schemes and protocols. Mater. Today Proc. 2021. [Google Scholar] [CrossRef]
  208. Kushalnagar, N.; Montenegro, G.; Schumacher, C. IPv6 over Low-Power Wireless Personal Area Networks (6LoWPANs): Overview, Assumptions, Problem Statement, and Goals; Technical Report; IETF: Fremont, CA, USA, 2007. [Google Scholar]
  209. White, A.L.; Stoughton, M.; Feng, L. Servicizing: The Quiet Transition to Extended Product Responsibility; Tellus Institute: Boston, CA, USA, 1999; Volume 97. [Google Scholar]
  210. Toffel, M.W. Contracting for Servicizing. Harvard Business School Technology & Operations Mgt. Unit Research Paper No. 08-063. 4 February 2008. Available online: https://ssrn.com/abstract=1090237 (accessed on 1 June 2022).
  211. Plepys, A.; Heiskanen, E.; Mont, O. European policy approaches to promote servicizing. J. Clean. Prod. 2015, 97, 117–123. [Google Scholar] [CrossRef]
  212. Agrawal, V.V.; Bellos, I. The potential of servicizing as a green business model. Manag. Sci. 2016, 63, 1545–1562. [Google Scholar] [CrossRef]
  213. Alam, S.; Chowdhury, M.M.R.; Noll, J. Senaas: An event-driven sensor virtualization approach for Internet of things cloud. In Proceedings of the 2010 IEEE International Conference on Networked Embedded Systems for Enterprise Applications, Suzhou, China, 25–26 November 2010; IEEE: Piscataway, NJ, USA, 2010; pp. 1–6. [Google Scholar]
  214. Distefano, S.; Merlino, G.; Puliafito, A. A utility paradigm for IoT: The sensing Cloud. Pervasive Mob. Comput. 2015, 20, 127–144. [Google Scholar] [CrossRef]
  215. Chen, Y.; Hu, H. Internet of intelligent things and robot as a service. Simul. Model. Pract. Theory 2013, 34, 159–171. [Google Scholar] [CrossRef]
  216. Toulios, L. Remote sensing as a service tool for the detection of water resources and crop conditions monitoring. In Hellenic Agricultural Organization DEMETER; Institute of Soil Classification & Mapping: Larissa, Greece, 2015. [Google Scholar]
  217. Ouaddah, A.; Mousannif, H.; Elkalam, A.A.; Ouahman, A.A. Access control in the Internet of Things: Big challenges and new opportunities. Comput. Netw. 2017, 112, 237–262. [Google Scholar] [CrossRef]
  218. Association, I.S. IEEE Standard for Local and Metropolitan Area Networks—Part 15.4: Low-Rate Wireless Personal Area Networks (LR-WPANs); IEEE Std 802.15.4-2011; IEEE Computer Society: New York, NY, USA, 2011. [Google Scholar]
  219. Association, I.S. IEEE Standard for Local and Metropolitan Area Networks—Part 15.4: Low-Rate Wireless Personal Area Networks (LR-WPANs) Amendment 1: MAC sublayer; IEEE Std 802.15. 4e-2012 (Amendment to IEEE Std 802.15. 4-2011); IEEE Computer Society: New York, NY, USA, 2012. [Google Scholar]
  220. Bormann, C.; Castellani, A.P.; Shelby, Z. Coap: An application protocol for billions of tiny Internet nodes. IEEE Internet Comput. 2012, 16, 62–67. [Google Scholar] [CrossRef]
  221. Raza, S.; Shafagh, H.; Hewage, K.; Hummen, R.; Voigt, T. Lithe: Lightweight secure CoAP for the Internet of things. IEEE Sens. J. 2013, 13, 3711–3720. [Google Scholar] [CrossRef]
  222. Glissa, G.; Meddeb, A. 6LowPSec: An end-to-end security protocol for 6LoWPAN. Ad Hoc Netw. 2019, 82, 100–112. [Google Scholar] [CrossRef]
  223. Mayzaud, A.; Badonnel, R.; Chrisment, I. A Taxonomy of Attacks in RPL-based Internet of Things. Int. J. Netw. Secur. 2016, 18, 459–473. [Google Scholar]
  224. Villegas, D.; Bobroff, N.; Rodero, I.; Delgado, J.; Liu, Y.; Devarakonda, A.; Fong, L.; Sadjadi, S.M.; Parashar, M. Cloud federation in a layered service model. J. Comput. Syst. Sci. 2012, 78, 1330–1344. [Google Scholar] [CrossRef] [Green Version]
  225. Hachem, S.; Teixeira, T.; Issarny, V. Ontologies for the Internet of things. In Proceedings of the 8th Middleware Doctoral Symposium, Lisbon, Portugal, 12–16 December 2011; p. 3. [Google Scholar]
  226. Wang, W.; De, S.; Toenjes, R.; Reetz, E.; Moessner, K. A comprehensive ontology for knowledge representation in the Internet of things. In Proceedings of the 2012 IEEE 11th International Conference on Trust, Security and Privacy in Computing and Communications, Liverpool, UK, 25–27 June 2012; IEEE: Piscataway, NJ, USA, 2012; pp. 1793–1798. [Google Scholar]
  227. Song, Z.; Cárdenas, A.A.; Masuoka, R. Semantic middleware for the Internet of things. In Proceedings of the 2010 Internet of Things (IOT), Tokyo, Japan, 29 November–1 December 2010; IEEE: Piscataway, NJ, USA, 2010; pp. 1–8. [Google Scholar]
  228. Ngu, A.H.; Gutierrez, M.; Metsis, V.; Nepal, S.; Sheng, Q.Z. IoT middleware: A survey on issues and enabling technologies. IEEE Internet Things J. 2017, 4, 1–20. [Google Scholar]
  229. Razzaque, M.A.; Milojevic-Jevric, M.; Palade, A.; Clarke, S. Middleware for Internet of Things: A survey. IEEE Internet Things J. 2016, 3, 70–95. [Google Scholar] [CrossRef] [Green Version]
  230. Mineraud, J.; Mazhelis, O.; Su, X.; Tarkoma, S. A gap analysis of Internet-of-Things platforms. Comput. Commun. 2016, 89, 5–16. [Google Scholar] [CrossRef] [Green Version]
  231. Linthicum, D.S. PaaS death watch? IEEE Cloud Comput. 2017, 4, 6–9. [Google Scholar] [CrossRef]
  232. Atzori, L.; Iera, A.; Morabito, G. Understanding the Internet of Things: Definition, potentials, and societal role of a fast evolving paradigm. Ad Hoc Netw. 2017, 56, 122–140. [Google Scholar] [CrossRef]
  233. Baheti, R.; Gill, H. Cyber–physical systems. Impact Control. Technol. 2011, 12, 161–166. [Google Scholar]
  234. Lee, E.A. The past, present and future of cyber–physical systems: A focus on models. Sensors 2015, 15, 4837–4869. [Google Scholar] [CrossRef]
  235. Rho, S.; Vasilakos, A.V.; Chen, W. Cyber physical systems technologies and applications. Future Gener. Comput. Syst. 2016, 56, 436–437. [Google Scholar] [CrossRef]
  236. Alam, K.M.; El Saddik, A. C2ps: A Digital Twin architecture reference model for the cloud-based cyber–physical systems. IEEE Access 2017, 5, 2050–2062. [Google Scholar] [CrossRef]
  237. Um, J.S. Drones as Cyber–Physical Systems; Springer: Berlin/Heidelberg, Germany, 2019. [Google Scholar]
  238. Sheth, A. Internet of things to smart iot through semantic, cognitive, and perceptual computing. IEEE Intell. Syst. 2016, 31, 108–112. [Google Scholar] [CrossRef]
  239. Rawat, P.; Singh, K.D.; Bonnin, J.M. Cognitive Radio for M2M and Internet of Things: A survey. Comput. Commun. 2016, 94, 1–29. [Google Scholar] [CrossRef]
  240. Mainetti, L.; Patrono, L.; Vilei, A. Evolution of wireless sensor networks towards the Internet of things: A survey. In Proceedings of the SoftCOM 2011, 19th International Conference on Software, Telecommunications and Computer Networks, Split, Croatia, 15–17 September 2011; IEEE: Piscataway, NJ, USA, 2011; pp. 1–6. [Google Scholar]
  241. Welbourne, E.; Battle, L.; Cole, G.; Gould, K.; Rector, K.; Raymer, S.; Balazinska, M.; Borriello, G. Building the Internet of things using RFID: The RFID ecosystem experience. IEEE Internet Comput. 2009, 13, 48–55. [Google Scholar] [CrossRef]
  242. Li, C.; Mo, L.; Zhang, D. Review on UHF RFID localization methods. IEEE J. Radio Freq. Identif. 2019, 3, 205–215. [Google Scholar] [CrossRef]
  243. Shao, C.; Roh, H.; Lee, W. Next-generation RF-powered networks for Internet of Things: Architecture and research perspectives. J. Netw. Comput. Appl. 2018, 123, 23–31. [Google Scholar] [CrossRef]
  244. Rastogi, V.; Suciu, D.; Welbourne, E. Access control over uncertain data. Proc. Vldb Endow. 2008, 1, 821–832. [Google Scholar] [CrossRef] [Green Version]
  245. Bormann, C.; Ersue, M.; Keranen, A. Terminology for Constrained-Node Networks; Technical Report; Internet Engineering Task Force (IETF): Fremont, CA, USA, 2014. [Google Scholar] [CrossRef]
  246. Hahm, O.; Baccelli, E.; Petersen, H.; Tsiftes, N. Operating systems for low-end devices in the Internet of things: A survey. IEEE Internet Things J. 2016, 3, 720–734. [Google Scholar] [CrossRef] [Green Version]
  247. Javed, F.; Afzal, M.K.; Sharif, M.; Kim, B.S. Internet of Things (IoTs) Operating Systems Support, Networking Technologies, Applications, and Challenges: A Comparative Review. IEEE Commun. Surv. Tutor. 2018, 20, 2062–2100. [Google Scholar] [CrossRef]
  248. Li, X.; Lu, R.; Liang, X.; Shen, X.; Chen, J.; Lin, X. Smart community: An Internet of things application. IEEE Commun. Mag. 2011, 49, 68–75. [Google Scholar] [CrossRef]
  249. Sanchez, L.; Muñoz, L.; Galache, J.A.; Sotres, P.; Santana, J.R.; Gutierrez, V.; Ramdhany, R.; Gluhak, A.; Krco, S.; Theodoridis, E.; et al. SmartSantander: IoT experimentation over a Smart City testbed. Comput. Netw. 2014, 61, 217–238. [Google Scholar] [CrossRef] [Green Version]
  250. Perera, C.; Liu, C.H.; Jayawardena, S.; Chen, M. A survey on Internet of things from industrial market perspective. IEEE Access 2014, 2, 1660–1679. [Google Scholar] [CrossRef]
  251. Kothari, M.; Mistry, Z.; Kamat, A.; Ragavendran, U. Multi-antenna-Enabled Technologies for IoT-Driven Smart Cities. In IoT and IoE Driven Smart Cities; Springer: Berlin/Heidelberg, Germany, 2022; pp. 45–72. [Google Scholar]
  252. Gazis, D.C.; Herman, R. The moving and “phantom” bottlenecks. Transp. Sci. 1992, 26, 223–229. [Google Scholar] [CrossRef]
  253. Nitti, M.; Pilloni, V.; Colistra, G.; Atzori, L. The virtual object as a major element of the Internet of things: A survey. IEEE Commun. Surv. Tutor. 2016, 18, 1228–1240. [Google Scholar] [CrossRef] [Green Version]
  254. Rathore, M.M.; Ahmad, A.; Paul, A.; Rho, S. Urban planning and building Smart Cities based on the Internet of things using Big Data analytics. Comput. Netw. 2016, 101, 63–80. [Google Scholar] [CrossRef]
  255. Anda, C.; Medina, S.A.O.; Axhausen, K.W. Synthesising Digital Twin travellers: Individual travel demand from aggregated mobile phone data. Transp. Res. Part C Emerg. Technol. 2021, 128, 103118. [Google Scholar] [CrossRef]
  256. Satyanarayanan, M. Pervasive Computing: Vision and challenges. IEEE Pers. Commun. 2001, 8, 10–17. [Google Scholar] [CrossRef]
  257. Ma, J.; Yang, L.T.; Apduhan, B.O.; Huang, R.; Barolli, L.; Takizawa, M. Towards a smart world and ubiquitous intelligence: A walkthrough from smart things to smart hyperspaces and UbicKids. Int. J. Pervasive Comput. Commun. 2005, 1, 53–68. [Google Scholar] [CrossRef] [Green Version]
  258. Kortuem, G.; Kawsar, F.; Sundramoorthy, V.; Fitton, D. Smart objects as building blocks for the Internet of things. IEEE Internet Comput. 2010, 14, 44–51. [Google Scholar] [CrossRef] [Green Version]
  259. Fortino, G.; Guerrieri, A.; Russo, W. Agent-oriented smart objects development. In Proceedings of the 2012 IEEE 16th International Conference on Computer Supported Cooperative Work in Design (CSCWD), Wuhan, China, 23–25 May 2012; IEEE: Piscataway, NJ, USA, 2012; pp. 907–912. [Google Scholar]
  260. Atzori, L.; Iera, A.; Morabito, G. From “smart objects” to “social objects”: The next evolutionary step of the Internet of things. IEEE Commun. Mag. 2014, 52, 97–105. [Google Scholar] [CrossRef]
  261. Tan, L.; Wang, N. Future Internet: The Internet of things. In Proceedings of the 2010 3rd International Conference on Advanced Computer Theory and Engineering(ICACTE), Chengdu, China, 20–22 August 2010; IEEE: Piscataway, NJ, USA, 2010; Volume 5, p. V5-376. [Google Scholar]
  262. Khan, R.; Khan, S.U.; Zaheer, R.; Khan, S. Future Internet: The Internet of things architecture, possible applications and key challenges. In Proceedings of the 2012 10th International Conference on Frontiers of Information Technology, Islamabad, Pakistan, 17–19 December 2012; IEEE: Piscataway, NJ, USA, 2012; pp. 257–260. [Google Scholar]
  263. Stankovic, J.A. Research directions for the Internet of things. IEEE Internet Things J. 2014, 1, 3–9. [Google Scholar] [CrossRef]
  264. Hurlburt, G.F.; Voas, J.; Miller, K.W. The Internet of things: A reality check. IT Prof. 2012, 14, 56–59. [Google Scholar] [CrossRef]
  265. Ashton, K. That ‘Internet of things’ thing. RFID J. 2009, 22, 97–114. [Google Scholar]
  266. Gluhak, A.; Krco, S.; Nati, M.; Pfisterer, D.; Mitton, N.; Razafindralambo, T. A survey on facilities for experimental Internet of things research. IEEE Commun. Mag. 2011, 49, 58–67. [Google Scholar] [CrossRef] [Green Version]
  267. Chen, Y.K. Challenges and opportunities of Internet of things. In Proceedings of the 17th Asia and South Pacific design automation conference, Sydney, NSW, Australia, 30 January–2 February 2012; pp. 383–388. [Google Scholar]
  268. Gubbi, J.; Buyya, R.; Marusic, S.; Palaniswami, M. Internet of Things (IoT): A vision, architectural elements, and future directions. Future Gener. Comput. Syst. 2013, 29, 1645–1660. [Google Scholar] [CrossRef] [Green Version]
  269. Singh, D.; Tripathi, G.; Jara, A.J. A survey of Internet-of-Things: Future vision, architecture, challenges and services. In Proceedings of the 2014 IEEE World Forum on Internet of Things (WF-IoT), Seoul, Korea, 6–8 March 2014; IEEE: Piscataway, NJ, USA, 2014; pp. 287–292. [Google Scholar]
  270. Want, R.; Schilit, B.N.; Jenson, S. Enabling the Internet of things. Computer 2015, 48, 28–35. [Google Scholar] [CrossRef]
  271. Al-Fuqaha, A.; Guizani, M.; Mohammadi, M.; Aledhari, M.; Ayyash, M. Internet of things: A survey on enabling technologies, protocols, and applications. IEEE Commun. Surv. Tutor. 2015, 17, 2347–2376. [Google Scholar] [CrossRef]
  272. Whitmore, A.; Agarwal, A.; Da Xu, L. The Internet of Things—A survey of topics and trends. Inf. Syst. Front. 2015, 17, 261–274. [Google Scholar] [CrossRef]
  273. Madakam, S.; Ramaswamy, R.; Tripathi, S. Internet of Things (IoT): A literature review. J. Comput. Commun. 2015, 3, 164. [Google Scholar] [CrossRef] [Green Version]
  274. Palattella, M.R.; Dohler, M.; Grieco, A.; Rizzo, G.; Torsner, J.; Engel, T.; Ladid, L. Internet of things in the 5G era: Enablers, architecture, and business models. IEEE J. Sel. Areas Commun. 2016, 34, 510–527. [Google Scholar] [CrossRef] [Green Version]
  275. Lin, J.; Yu, W.; Zhang, N.; Yang, X.; Zhang, H.; Zhao, W. A survey on Internet of things: Architecture, enabling technologies, security and privacy, and applications. IEEE Internet Things J. 2017, 4, 1125–1142. [Google Scholar] [CrossRef]
  276. Qiu, T.; Chen, N.; Li, K.; Atiquzzaman, M.; Zhao, W. How can heterogeneous Internet of things build our future: A survey. IEEE Commun. Surv. Tutor. 2018, 20, 2011–2027. [Google Scholar] [CrossRef]
  277. Shafique, K.; Khawaja, B.A.; Sabir, F.; Qazi, S.; Mustaqim, M. Internet of things (IoT) for next-generation smart systems: A review of current challenges, future trends and prospects for emerging 5G-IoT scenarios. IEEE Access 2020, 8, 23022–23040. [Google Scholar] [CrossRef]
  278. Zorzi, M.; Gluhak, A.; Lange, S.; Bassi, A. From today’s intranet of things to a future Internet of things: A wireless-and mobility-related view. IEEE Wirel. Commun. 2010, 17, 44–51. [Google Scholar] [CrossRef]
  279. Wu, J.; Zhao, W. Design and realization of wInternet: From net of things to Internet of things. ACM Trans. Cyber Phys. Syst. 2017, 1, 2. [Google Scholar] [CrossRef]
  280. Alamri, A.; Ansari, W.S.; Hassan, M.M.; Hossain, M.S.; Alelaiwi, A.; Hossain, M.A. A survey on sensor-cloud: Architecture, applications, and approaches. Int. J. Distrib. Sens. Netw. 2013, 9, 917923. [Google Scholar] [CrossRef]
  281. Misra, S.; Chatterjee, S.; Obaidat, M.S. On theoretical modeling of sensor cloud: A paradigm shift from wireless sensor network. IEEE Syst. J. 2017, 11, 1084–1093. [Google Scholar] [CrossRef]
  282. Pfisterer, D.; Romer, K.; Bimschas, D.; Kleine, O.; Mietz, R.; Truong, C.; Hasemann, H.; Kröller, A.; Pagel, M.; Hauswirth, M.; et al. SPITFIRE: Toward a semantic web of things. IEEE Commun. Mag. 2011, 49, 40–48. [Google Scholar] [CrossRef]
  283. Bonomi, F.; Milito, R.; Zhu, J.; Addepalli, S. Fog computing and its role in the Internet of things. In Proceedings of the First Edition of the MCC Workshop on Mobile Cloud Computing, Helsinki, Finland, 13–17 August 2012; pp. 13–16. [Google Scholar]
  284. Guo, B.; Zhang, D.; Wang, Z.; Yu, Z.; Zhou, X. Opportunistic IoT: Exploring the harmonious interaction between human and the Internet of things. J. Netw. Comput. Appl. 2013, 36, 1531–1539. [Google Scholar] [CrossRef]
  285. Paraskevakos, T. Apparatus for Generating and Transmitting Digital Information. US Patent 3,812,296, 21 May 1974. [Google Scholar]
  286. Datta, S.K.; Bonnet, C.; Nikaein, N. An IoT gateway centric architecture to provide novel M2M services. In Proceedings of the 2014 IEEE World Forum on Internet of Things (WF-IoT), Seoul, Korea, 6–8 March 2014; IEEE: Piscataway, NJ, USA, 2014; pp. 514–519. [Google Scholar]
  287. Andreev, S.; Galinina, O.; Pyattaev, A.; Gerasimenko, M.; Tirronen, T.; Torsner, J.; Sachs, J.; Dohler, M.; Koucheryavy, Y. Understanding the IoT connectivity landscape: A contemporary M2M radio technology roadmap. IEEE Commun. Mag. 2015, 53, 32–40. [Google Scholar] [CrossRef] [Green Version]
  288. Gazis, V. A Survey of Standards for Machine-to-Machine and the Internet of Things. IEEE Commun. Surv. Tutor. 2017, 19, 482–511. [Google Scholar] [CrossRef]
  289. Patil, S.; Gokhale, P. Systematic Review of Resource Allocation Methods Using Scheduling for M2M (Machine to Machine Communication) in IoT Network. In Security Issues and Privacy Threats in Smart Ubiquitous Computing; Springer: Cham, Switzerland, 2021; pp. 213–224. [Google Scholar]
  290. Baronti, P.; Pillai, P.; Chook, V.W.C.; Chessa, S.; Gotta, A.; Hu, Y.F. Wireless sensor networks: A survey on the state of the art and the 802.15. 4 and ZigBee standards. Comput. Commun. 2007, 30, 1655–1695. [Google Scholar] [CrossRef]
  291. Yick, J.; Mukherjee, B.; Ghosal, D. Wireless sensor network survey. Comput. Netw. 2008, 52, 2292–2330. [Google Scholar] [CrossRef]
  292. Khorov, E.; Lyakhov, A.; Krotov, A.; Guschin, A. A survey on IEEE 802.11 ah: An enabling networking technology for Smart Cities. Comput. Commun. 2015, 58, 53–69. [Google Scholar] [CrossRef]
  293. Wang, H.; Fapojuwo, A.O. A survey of enabling technologies of low power and long range machine-to-machine communications. IEEE Commun. Surv. Tutor. 2017, 19, 2621–2639. [Google Scholar] [CrossRef]
  294. Ishaq, I.; Carels, D.; Teklemariam, G.K.; Hoebeke, J.; den Abeele, F.V.; Poorter, E.D.; Moerman, I.; Demeester, P. IETF standardization in the field of the Internet of things (IoT): A survey. J. Sens. Actuator Netw. 2013, 2, 235–287. [Google Scholar] [CrossRef] [Green Version]
  295. Sheng, Z.; Yang, S.; Yu, Y.; Vasilakos, A.; Mccann, J.; Leung, K. A survey on the ietf protocol suite for the Internet of things: Standards, challenges, and opportunities. IEEE Wirel. Commun. 2013, 20, 91–98. [Google Scholar] [CrossRef]
  296. Bachir, A.; Dohler, M.; Watteyne, T.; Leung, K.K. MAC essentials for wireless sensor networks. IEEE Commun. Surv. Tutor. 2010, 12, 222–248. [Google Scholar] [CrossRef] [Green Version]
  297. Rajandekar, A.; Sikdar, B. A survey of MAC layer issues and protocols for machine-to-machine communications. IEEE Internet Things J. 2015, 2, 175–186. [Google Scholar] [CrossRef]
  298. Sotenga, P.Z.; Djouani, K.; Kurien, A.M. Media access control in large-scale Internet of Things: A review. IEEE Access 2020, 8, 55834–55859. [Google Scholar] [CrossRef]
  299. International Organization for Standardization. The Basic Reference Model for Open Systems Interconnection (ISO/IEC Standard No. 7498-1). 2016. Available online: https://www.iso.org/standard/20269.html (accessed on 1 June 2022).
  300. Chen, K.C.; Lien, S.Y. Machine-to-machine communications: Technologies and challenges. Ad Hoc Netw. 2014, 18, 3–23. [Google Scholar] [CrossRef]
  301. Aijaz, A.; Aghvami, A.H. Cognitive machine-to-machine communications for Internet-of-Things: A protocol stack perspective. IEEE Internet Things J. 2015, 2, 103–112. [Google Scholar] [CrossRef]
  302. Akyildiz, I.F.; Jornet, J.M. The Internet of nano-things. IEEE Wirel. Commun. 2010, 17, 58–63. [Google Scholar] [CrossRef]
  303. Akyildiz, I.F.; Pierobon, M.; Balasubramaniam, S.; Koucheryavy, Y. The Internet of bio-nano things. IEEE Commun. Mag. 2015, 53, 32–40. [Google Scholar] [CrossRef]
  304. Gravina, R.; Alinia, P.; Ghasemzadeh, H.; Fortino, G. Multi-sensor fusion in body sensor networks: State-of-the-art and research challenges. Inf. Fusion 2017, 35, 68–80. [Google Scholar] [CrossRef]
  305. Miraz, M.; Ali, M.; Excell, P.; Picking, R. Internet of nano-things, things and everything: Future growth trends. Future Internet 2018, 10, 68. [Google Scholar] [CrossRef] [Green Version]
  306. Arslan, H.; Chen, Z.N.; Di Benedetto, M.G. Ultra Wideband Wireless Communication; John Wiley & Sons: Hoboken, NJ, USA, 2006. [Google Scholar]
  307. Tozlu, S.; Senel, M.; Mao, W.; Keshavarzian, A. Wi-Fi enabled sensors for Internet of things: A practical approach. IEEE Commun. Mag. 2012, 50, 134–143. [Google Scholar] [CrossRef]
  308. Ratasuk, R.; Mangalvedhe, N.; Zhang, Y.; Robert, M.; Koskinen, J.P. Overview of narrowband IoT in LTE Rel-13. In Proceedings of the 2016 IEEE Conference on Standards for Communications and Networking (CSCN), Berlin, Germany, 31 October–2 November 2016; IEEE: Piscataway, NJ, USA, 2016; pp. 1–7. [Google Scholar]
  309. Raza, U.; Kulkarni, P.; Sooriyabandara, M. Low power wide area networks: An overview. IEEE Commun. Surv. Tutor. 2017, 19, 855–873. [Google Scholar] [CrossRef] [Green Version]
  310. Nolan, K.E.; Guibene, W.; Kelly, M.Y. An evaluation of low power wide area network technologies for the Internet of Things. In Proceedings of the 2016 International Wireless Communications and Mobile Computing Conference (IWCMC), Paphos, Cyprus, 5–9 September 2016; IEEE: Piscataway, NJ, USA, 2016; pp. 439–444. [Google Scholar]
  311. Oliveira, L.M.L.; De Sousa, A.F.; Rodrigues, J.J.P.C. Routing and mobility approaches in IPv6 over LoWPAN mesh networks. Int. J. Commun. Syst. 2011, 24, 1445–1466. [Google Scholar] [CrossRef]
  312. Bor, M.C.; Roedig, U.; Voigt, T.; Alonso, J.M. Do LoRa low-power wide-area networks scale? In Proceedings of the 19th ACM International Conference on Modeling, Analysis and Simulation of Wireless and Mobile Systems, Malta, Malta, 13–17 November 2016; pp. 59–67. [Google Scholar]
  313. Petäjäjärvi, J.; Mikhaylov, K.; Pettissalo, M.; Janhunen, J.; Iinatti, J. Performance of a low-power wide-area network based on LoRa technology: Doppler robustness, scalability, and coverage. Int. J. Distrib. Sens. Netw. 2017, 13, 1550147717699412. [Google Scholar] [CrossRef] [Green Version]
  314. Georgiou, O.; Raza, U. Low power wide area network analysis: Can LoRa scale? IEEE Wirel. Commun. Lett. 2017, 6, 162–165. [Google Scholar] [CrossRef] [Green Version]
  315. Palattella, M.R.; Accettura, N.; Vilajosana, X.; Watteyne, T.; Grieco, L.A.; Boggia, G.; Dohler, M. Standardized protocol stack for the Internet of (important) things. IEEE Commun. Surv. Tutor. 2013, 15, 1389–1406. [Google Scholar] [CrossRef] [Green Version]
  316. Kurunathan, H.; Severino, R.; Koubaa, A.; Tovar, E. IEEE 802.15. 4e in a Nutshell: Survey and Performance Evaluation. IEEE Commun. Surv. Tutor. 2018, 20, 1989–2010. [Google Scholar] [CrossRef]
  317. Hui, J.W.; Culler, D.E. IP is dead, long live IP for wireless sensor networks. In Proceedings of the 6th ACM Conference on Embedded Network Sensor Systems, Raleigh, NC, USA, 5–7 November 2008; pp. 15–28. [Google Scholar]
  318. Hong, S.; Kim, D.; Ha, M.; Bae, S.; Park, S.J.; Jung, W.; Kim, J.E. SNAIL: An IP-based wireless sensor network approach to the Internet of things. IEEE Wirel. Commun. 2010, 17, 34–42. [Google Scholar] [CrossRef]
  319. Rodrigues, J.J.P.C.; Neves, P.A.C.S. A survey on IP-based wireless sensor network solutions. Int. J. Commun. Syst. 2010, 23, 963–981. [Google Scholar] [CrossRef]
  320. Imani, M.; Moghadam, A.Q.; Zarif, N.; Noshiri, O.; Faramarzi, K.; Arabnia, H.; Joudaki, M. A Comprehensive Survey on Addressing Methods in the Internet of Things. arXiv 2018, arXiv:1807.02173. [Google Scholar]
  321. Bonomi, F.; Milito, R.; Natarajan, P.; Zhu, J. Fog computing: A platform for Internet of things and analytics. In Big Data and Internet of Things: A Roadmap for Smart Environments; Springer: Berlin/Heidelberg, Germany, 2014; pp. 169–186. [Google Scholar]
  322. Stojmenovic, I.; Wen, S. The fog computing paradigm: Scenarios and security issues. In Proceedings of the 2014 Federated Conference on Computer Science and Information Systems, Warsaw, Poland, 7–10 September 2014; IEEE: Piscataway, NJ, USA, 2014; pp. 1–8. [Google Scholar]
  323. Jalali, F.; Hinton, K.; Ayre, R.; Alpcan, T.; Tucker, R.S. Fog computing may help to save energy in cloud computing. IEEE J. Sel. Areas Commun. 2016, 34, 1728–1739. [Google Scholar] [CrossRef]
  324. Soliman, M.; Abiodun, T.; Hamouda, T.; Zhou, J.; Lung, C.H. Smart home: Integrating Internet of things with web services and cloud computing. In Proceedings of the 2013 IEEE 5th International Conference on Cloud Computing Technology and Science, Bristol, UK, 2–5 December 2013; IEEE: Piscataway, NJ, USA, 2013; pp. 317–320. [Google Scholar]
  325. Botta, A.; De Donato, W.; Persico, V.; Pescapé, A. On the integration of cloud computing and Internet of things. Future Internet Things Cloud (FiCloud) 2014, 27, 23–30. [Google Scholar]
  326. Fortino, G.; Guerrieri, A.; Russo, W.; Savaglio, C. Integration of agent-based and cloud computing for the smart objects-oriented IoT. In Proceedings of the 2014 IEEE 18th International Conference on Computer Supported Cooperative Work in Design (CSCWD), Hsinchu, Taiwan, 21–23 May 2014; IEEE: Piscataway, NJ, USA, 2014; pp. 493–498. [Google Scholar]
  327. Cavalcante, E.; Pereira, J.; Alves, M.P.; Maia, P.; Moura, R.; Batista, T.; Delicato, F.C.; Pires, P.F. On the interplay of Internet of Things and Cloud Computing: A systematic mapping study. Comput. Commun. 2016, 89, 17–33. [Google Scholar] [CrossRef]
  328. Bittencourt, L.; Immich, R.; Sakellariou, R.; Fonseca, N.; Madeira, E.; Curado, M.; Villas, L.; da Silva, L.; Lee, C.; Rana, O. The Internet of Things, Fog and Cloud Continuum: Integration and Challenges. Internet Things 2018, 3, 134–155. [Google Scholar] [CrossRef] [Green Version]
  329. Aazam, M.; Huh, E.N. Fog computing and smart gateway based communication for cloud of things. In Proceedings of the 2014 International Conference on Future Internet of Things and Cloud, Barcelona, Spain, 27–29 August 2014; IEEE: Piscataway, NJ, USA, 2014; pp. 464–470. [Google Scholar]
  330. Satyanarayanan, M.; Simoens, P.; Xiao, Y.; Pillai, P.; Chen, Z.; Ha, K.; Hu, W.; Amos, B. Edge analytics in the Internet of things. IEEE Pervasive Comput. 2015, 14, 24–31. [Google Scholar] [CrossRef] [Green Version]
  331. Chiang, M.; Zhang, T. Fog and IoT: An overview of research opportunities. IEEE Internet Things J. 2016, 3, 854–864. [Google Scholar] [CrossRef]
  332. Mao, Y.; You, C.; Zhang, J.; Huang, K.; Letaief, K.B. A survey on mobile edge computing: The communication perspective. IEEE Commun. Surv. Tutor. 2017, 19, 2322–2358. [Google Scholar] [CrossRef] [Green Version]
  333. Naha, R.K.; Garg, S.; Georgakopoulos, D.; Jayaraman, P.P.; Gao, L.; Xiang, Y.; Ranjan, R. Fog Computing: Survey of Trends, Architectures, Requirements, and Research Directions. IEEE Access 2018, 6, 47980–48009. [Google Scholar] [CrossRef]
  334. Vaquero, L.M.; Rodero-Merino, L. Finding your way in the fog: Towards a comprehensive definition of fog computing. ACM SIGCOMM Comput. Commun. Rev. 2014, 44, 27–32. [Google Scholar] [CrossRef]
  335. Giang, N.K.; Blackstock, M.; Lea, R.; Leung, V.C.M. Developing IoT applications in the fog: A distributed dataflow approach. In Proceedings of the 2015 5th International Conference on the Internet of Things (IOT), Seoul, Korea, 26–28 October 2015; IEEE: Piscataway, NJ, USA, 2015; pp. 155–162. [Google Scholar]
  336. Dastjerdi, A.V.; Buyya, R. Fog computing: Helping the Internet of Things realize its potential. Computer 2016, 49, 112–116. [Google Scholar] [CrossRef]
  337. Aazam, M.; Zeadally, S.; Harras, K.A. Offloading in fog computing for IoT: Review, enabling technologies, and research opportunities. Future Gener. Comput. Syst. 2018, 87, 278–289. [Google Scholar] [CrossRef]
  338. Sarkar, S.; Chatterjee, S.; Misra, S. Assessment of the Suitability of Fog Computing in the Context of Internet of Things. IEEE Trans. Cloud Comput. 2018, 6, 46–59. [Google Scholar] [CrossRef]
  339. Negash, B.; Westerlund, T.; Tenhunen, H. Towards an interoperable Internet of Things through a web of virtual things at the Fog layer. Future Gener. Comput. Syst. 2019, 91, 96–107. [Google Scholar] [CrossRef]
  340. Alam, M.G.R.; Hassan, M.M.; Uddin, M.Z.; Almogren, A.; Fortino, G. Autonomic computation offloading in mobile edge for IoT applications. Future Gener. Comput. Syst. 2019, 90, 149–157. [Google Scholar] [CrossRef]
  341. Liyanage, M.; Chang, C.; Srirama, S.N. mePaaS: Mobile-embedded platform as a service for distributing fog computing to edge nodes. In Proceedings of the 2016 17th International Conference on Parallel and Distributed Computing, Applications and Technologies (PDCAT), Guangzhou, China, 16–18 December 2016; IEEE: Piscataway, NJ, USA, 2016; pp. 73–80. [Google Scholar]
  342. Dogo, E.M.; Salami, A.F.; Aigbavboa, C.O.; Nkonyana, T. Taking cloud computing to the extreme edge: A review of mist computing for Smart Cities and industry 4.0 in Africa. Edge Comput. 2019, 107–132. [Google Scholar] [CrossRef]
  343. Hong, K.; Lillethun, D.; Ramachandran, U.; Ottenwälder, B.; Koldehofe, B. Mobile fog: A programming model for large-scale applications on the Internet of things. In Proceedings of the Second ACM SIGCOMM Workshop on Mobile Cloud Computing, Hong Kong, China, 12–16 August 2013; pp. 15–20. [Google Scholar]
  344. Rindos, A.; Wang, Y. Dew computing: The complementary piece of cloud computing. In Proceedings of the 2016 IEEE International Conferences on Big Data and Cloud Computing (BDCloud), Social Computing and Networking (SocialCom), Sustainable Computing and Communications (SustainCom)(BDCloud-SocialCom-SustainCom), Atlanta, GA, USA, 8–10 October 2016; IEEE: Piscataway, NJ, USA, 2016; pp. 15–20. [Google Scholar]
  345. Ray, P.P. An introduction to dew computing: Definition, concept and implications. IEEE Access 2017, 6, 723–737. [Google Scholar] [CrossRef]
  346. Gushev, M. Dew computing architecture for cyber–physical systems and IoT. Internet Things 2020, 11, 100186. [Google Scholar] [CrossRef]
  347. Bhattacharyya, S.; Katramatos, D.; Yoo, S. Why wait? Let us start computing while the data is still on the wire. Future Gener. Comput. Syst. 2018, 89, 563–574. [Google Scholar] [CrossRef]
  348. Oxford Dictionary. Subsidiarity. 2018. Available online: https://www.lexico.com/definition/subsidiarity (accessed on 1 June 2022).
  349. Campbell, J.B.; Wynne, R.H. Introduction to Remote Sensing; Guilford Press: New York, NY, USA, 2011. [Google Scholar]
  350. Viani, F.; Rocca, P.; Oliveri, G.; Massa, A. Pervasive remote sensing through WSNs. In Proceedings of the 2012 6th European Conference on Antennas and Propagation (EUCAP), Prague, Czech Republic, 26–30 March 2012; IEEE: Piscataway, NJ, USA, 2012; pp. 49–50. [Google Scholar]
  351. Sutaria, R.; Govindachari, R. Making sense of interoperability: Protocols and Standardization initiatives in IOT. In Proceedings of the 2nd International Workshop on Computing and Networking for Internet of Things, Mumbai, India, 3–6 January 2013; p. 7. [Google Scholar]
  352. Triantafyllou, A.; Sarigiannidis, P.; Bibi, S. Precision agriculture: A remote sensing monitoring system architecture. Information 2019, 10, 348. [Google Scholar] [CrossRef] [Green Version]
  353. Ruiz-Fernández, D.; Marcos-Jorquera, D.; Gilart-Iglesias, V.; Vives-Boix, V.; Ramírez-Navarro, J. Empowerment of patients with hypertension through BPM, IoT and remote sensing. Sensors 2017, 17, 2273. [Google Scholar] [CrossRef] [Green Version]
  354. Demetri, S.; Zúñiga, M.; Picco, G.P.; Kuipers, F.; Bruzzone, L.; Telkamp, T. Automated estimation of link quality for LoRa: A remote sensing approach. In Proceedings of the 2019 18th ACM/IEEE International Conference on Information Processing in Sensor Networks (IPSN), Montreal, QC, Canada, 15–18 April 2019; IEEE: Piscataway, NJ, USA, 2019; pp. 145–156. [Google Scholar]
  355. Abdelwahab, S.; Hamdaoui, B.; Guizani, M.; Rayes, A. Enabling smart cloud services through remote sensing: An Internet of everything enabler. IEEE Internet Things J. 2014, 1, 276–288. [Google Scholar] [CrossRef]
  356. Ismail, A.; Bagula, B.A.; Tuyishimire, E. Internet-of-things in motion: A uav coalition model for remote sensing in Smart Cities. Sensors 2018, 18, 2184. [Google Scholar] [CrossRef] [Green Version]
  357. Fraga-Lamas, P.; Ramos, L.; Mondéjar-Guerra, V.; Fernández-Caramés, T.M. A review on IoT deep learning UAV systems for autonomous obstacle detection and collision avoidance. Remote Sens. 2019, 11, 2144. [Google Scholar] [CrossRef] [Green Version]
  358. Srivastava, A.N.; Oza, N.C.; Stroeve, J. Virtual sensors: Using data mining techniques to efficiently estimate remote sensing spectra. IEEE Trans. Geosci. Remote Sens. 2005, 43, 590–600. [Google Scholar] [CrossRef]
  359. Wang, W.; He, Z.; Huang, D.; Zhang, X. Research on service platform of Internet of things for Smart City. Int. Arch. Photogramm. Remote Sens. Spat. Inf. Sci. 2014, 40, 301. [Google Scholar] [CrossRef] [Green Version]
  360. Milgram, P.; Kishino, F. A taxonomy of mixed reality visual displays. IEICE Trans. Inf. Syst. 1994, 77, 1321–1329. [Google Scholar]
  361. Batty, M. A perspective on city Dashboards. Reg. Stud. Reg. Sci. 2015, 2, 29–32. [Google Scholar] [CrossRef] [Green Version]
  362. Bednarz, T.; Branchaud, D.; Wang, F.; Baker, J.; Marquarding, M. Digital Twin of the Australian Square Kilometre Array (ASKAP). In SIGGRAPH Asia 2020 Posters; Association for Computing Machinery: New York, NY, USA, 2020; SA ’20. [Google Scholar] [CrossRef]
  363. Spielhofer, R.; Thrash, T.; Wissen Hayek, U.; Grêt-Regamey, A.; Salak, B.; Grübel, J.; Schinazi, V.R. Physiological and behavioral reactions to renewable energy systems in various landscape types. Renew. Sustain. Energy Rev. 2021, 135, 110410. [Google Scholar] [CrossRef]
  364. Hua, J.; Lee, S.; Roman, G.C.; Julien, C. Arciot: Enabling intuitive device control in the Internet of things through Augmented Reality. In Proceedings of the 2021 IEEE International Conference on Pervasive Computing and Communications Workshops and other Affiliated Events (PerCom Workshops), Kassel, Germany, 22–26 March 2021; IEEE: Piscataway, NJ, USA, 2021; pp. 558–564. [Google Scholar]
  365. Weyrich, M.; Ebert, C. Reference architectures for the Internet of things. IEEE Softw. 2016, 33, 112–116. [Google Scholar] [CrossRef]
  366. Di Martino, B.; Rak, M.; Ficco, M.; Esposito, A.; Maisto, S.A.; Nacchia, S. Internet of things reference architectures, security and interoperability: A survey. Internet Things 2018, 1, 99–112. [Google Scholar] [CrossRef]
  367. Lynn, T.; Endo, P.T.; Ribeiro, A.M.N.; Barbosa, G.B.; Rosati, P. The Internet of things: Definitions, key concepts, and reference architectures. In The Cloud-to-Thing Continuum; Palgrave Macmillan: Cham, Switzerland, 2020; pp. 1–22. [Google Scholar]
  368. Tsiatsis, V.; Gluhak, A.; Bauge, T.; Montagut, F.; Bernat, J.; Bauer, M.; Villalonga, C.; Barnaghi, P.; Krco, S. The SENSEI Real World Internet Architecture. In Towards the Future Internet: Emerging Trends from European Research; IOS Press: Amsterdam, The Netherlands, 2010; pp. 247–256. [Google Scholar]
  369. De, S.; Barnaghi, P.; Bauer, M.; Meissner, S. Service modelling for the Internet of Things. In Proceedings of the 2011 Federated Conference on Computer Science and Information Systems (FedCSIS), Szczecin, Poland, 18–21 September 2011; IEEE: Piscataway, NJ, USA, 2011; pp. 949–955. [Google Scholar]
  370. Asghari, P.; Rahmani, A.M.; Javadi, H.H.S. Service composition approaches in IoT: A systematic review. J. Netw. Comput. Appl. 2018, 120, 61–77. [Google Scholar] [CrossRef]
  371. Hui, T.K.L.; Sherratt, R.S.; Sánchez, D.D. Major requirements for building Smart Homes in Smart Cities based on Internet of Things technologies. Future Gener. Comput. Syst. 2017, 76, 358–369. [Google Scholar] [CrossRef] [Green Version]
  372. Laput, G.; Harrison, C. Exploring the efficacy of sparse, general-purpose sensor constellations for wide-area activity sensing. Proc. Acm Interact. Mob. Wearable Ubiquitous Technol. 2019, 3, 1–19. [Google Scholar] [CrossRef]
  373. Chianese, A.; Piccialli, F. Designing a smart museum: When cultural heritage joins IoT. In Proceedings of the 2014 Eighth International Conference on Next Generation Mobile Apps, Services and Technologies, Oxford, UK, 10–12 September 2014; IEEE: Piscataway, NJ, USA, 2014; pp. 300–306. [Google Scholar]
  374. Suciu, G.; Vulpe, A.; Halunga, S.; Fratu, O.; Todoran, G.; Suciu, V. Smart Cities built on resilient cloud computing and secure Internet of things. In Proceedings of the 2013 19th International Conference on Control Systems and Computer Science, Bucharest, Romania, 29–31 May 2013; IEEE: Piscataway, NJ, USA, 2013; pp. 513–518. [Google Scholar]
  375. Zanella, A.; Bui, N.; Castellani, A.; Vangelista, L.; Zorzi, M. Internet of things for Smart Cities. IEEE Internet Things J. 2014, 1, 22–32. [Google Scholar] [CrossRef]
  376. Alavi, A.H.; Jiao, P.; Buttlar, W.G.; Lajnef, N. Internet of Things-enabled Smart Cities: State-of-the-art and future trends. Measurement 2018, 129, 589–606. [Google Scholar] [CrossRef]
  377. Matern, A.; Binder, J.; Noack, A. Smart regions: Insights from hybridization and peripheralization research. Eur. Plan. Stud. 2020, 28, 2060–2077. [Google Scholar] [CrossRef] [Green Version]
  378. Sun, Y.; Song, H.; Jara, A.J.; Bie, R. Internet of things and Big Data analytics for smart and connected communities. IEEE Access 2016, 4, 766–773. [Google Scholar] [CrossRef]
  379. Glaeser, E.L.; Berry, C.R. Why are smart places getting smarter. In Rappaport Institute/Taubman Center Policy Brief No. PB-2006-2; Rappaport Institute/Taubman Center: Cambridge, MA, USA, 2006; Volume 2. [Google Scholar]
  380. Meijer, A.; Bolívar, M.P.R. Governing the Smart City: A review of the literature on smart urban governance. Int. Rev. Adm. Sci. 2016, 82, 392–408. [Google Scholar] [CrossRef]
  381. Shrouf, F.; Ordieres, J.; Miragliotta, G. Smart factories in Industry 4.0: A review of the concept and of energy management approached in production based on the Internet of Things paradigm. In Proceedings of the 2014 IEEE International Conference on Industrial Engineering and Engineering Management, Selangor, Malaysia, 9–12 December 2014; IEEE: Piscataway, NJ, USA, 2014; pp. 697–701. [Google Scholar]
  382. Lu, Y.; Xu, X.; Wang, L. Smart manufacturing process and system automation–a critical review of the standards and envisioned scenarios. J. Manuf. Syst. 2020, 56, 312–325. [Google Scholar] [CrossRef]
  383. Kochovski, P.; Stankovski, V. Supporting smart construction with dependable edge computing infrastructures and applications. Autom. Constr. 2018, 85, 182–192. [Google Scholar] [CrossRef]
  384. Rohokale, V.M.; Prasad, N.R.; Prasad, R. A cooperative Internet of Things (IoT) for rural healthcare monitoring and control. In Proceedings of the 2011 2nd International Conference on Wireless Communication, Vehicular Technology, Information Theory and Aerospace & Electronic Systems Technology (Wireless VITAE), Chennai, India, 28 February–3 March 2011; IEEE: Piscataway, NJ, USA, 2011; pp. 1–6. [Google Scholar]
  385. Faria, R.; Brito, L.; Baras, K.; Silva, J. Smart mobility: A survey. In Proceedings of the 2017 International Conference on Internet of Things for the Global Community (IoTGC), Funchal, Portugal, 10–13 July 2017; IEEE: Piscataway, NJ, USA, 2017; pp. 1–8. [Google Scholar]
  386. Kontogianni, A.; Alepis, E. Smart tourism: State of the art and literature review for the last six years. Array 2020, 6, 100020. [Google Scholar] [CrossRef]
  387. Szilagyi, I.; Wira, P. Ontologies and Semantic Web for the Internet of Things-a survey. In Proceedings of the IECON 2016—42nd Annual Conference of the IEEE Industrial Electronics Society, Florence, Italy, 23–26 October 2016; IEEE: Piscataway, NJ, USA, 2016; pp. 6949–6954. [Google Scholar]
  388. De Matos, E.; Tiburski, R.T.; Moratelli, C.R.; Johann Filho, S.; Amaral, L.A.; Ramachandran, G.; Krishnamachari, B.; Hessel, F. Context information sharing for the Internet of Things: A survey. Comput. Netw. 2020, 166, 106988. [Google Scholar] [CrossRef]
  389. Perera, C.; Zaslavsky, A.; Christen, P.; Georgakopoulos, D. Context aware computing for the Internet of things: A survey. IEEE Commun. Surv. Tutor. 2014, 16, 414–454. [Google Scholar] [CrossRef] [Green Version]
  390. Fremantle, P.; Scott, P. A survey of secure middleware for the Internet of Things. PeerJ Comput. Sci. 2017, 3, e114. [Google Scholar] [CrossRef] [Green Version]
  391. Almusaylim, Z.A.; Zaman, N. A review on smart home present state and challenges: Linked to context-awareness Internet of things (IoT). Wirel. Netw. 2019, 25, 3193–3204. [Google Scholar] [CrossRef]
  392. Rajagopalan, R.; Varshney, P.K. Data aggregation techniques in sensor networks: A survey. IEEE Commun. Surv. Tutor. 2006, 8, 48–63. [Google Scholar] [CrossRef] [Green Version]
  393. Wang, F.; Liu, J. Networked wireless sensor data collection: Issues, challenges, and approaches. IEEE Commun. Surv. Tutor. 2011, 13, 673–687. [Google Scholar] [CrossRef] [Green Version]
  394. Li, S.; Xu, L.D.; Wang, X. Compressed sensing signal and data acquisition in wireless sensor networks and Internet of things. IEEE Trans. Ind. Inform. 2013, 9, 2177–2186. [Google Scholar] [CrossRef] [Green Version]
  395. Zhang, Y.; Meratnia, N.; Havinga, P.J.M. Outlier detection techniques for wireless sensor networks: A survey. IEEE Commun. Surv. Tutor. 2010, 12, 159–170. [Google Scholar] [CrossRef] [Green Version]
  396. Alsheikh, M.A.; Lin, S.; Niyato, D.; Tan, H.P. Machine learning in wireless sensor networks: Algorithms, strategies, and applications. IEEE Commun. Surv. Tutor. 2014, 16, 1996–2018. [Google Scholar] [CrossRef] [Green Version]
  397. Bagaa, M.; Challal, Y.; Ksentini, A.; Derhab, A.; Badache, N. Data aggregation scheduling algorithms in wireless sensor networks: Solutions and challenges. IEEE Commun. Surv. Tutor. 2014, 16, 1339–1368. [Google Scholar] [CrossRef]
  398. Al-Turjman, F.M. Information-centric sensor networks for cognitive IoT: An overview. Ann. Telecommun. 2017, 72, 3–18. [Google Scholar] [CrossRef]
  399. Bin, S.; Yuan, L.; Xiaoyi, W. Research on data mining models for the Internet of things. In Proceedings of the 2010 International Conference on Image Analysis and Signal Processing, Zhejiang, China, 9–11 April 2010; IEEE: Piscataway, NJ, USA, 2010; pp. 127–132. [Google Scholar]
  400. Tsai, C.W.; Lai, C.F.; Chiang, M.C.; Yang, L.T. Data mining for Internet of Things: A survey. IEEE Commun. Surv. Tutor. 2014, 16, 77–97. [Google Scholar] [CrossRef]
  401. Ganz, F.; Puschmann, D.; Barnaghi, P.; Carrez, F. A practical evaluation of information processing and abstraction techniques for the Internet of things. IEEE Internet Things J. 2015, 2, 340–354. [Google Scholar] [CrossRef] [Green Version]
  402. Chen, F.; Deng, P.; Wan, J.; Zhang, D.; Vasilakos, A.V.; Rong, X. Data mining for the Internet of things: Literature review and challenges. Int. J. Distrib. Sens. Netw. 2015, 11, 431047. [Google Scholar] [CrossRef] [Green Version]
  403. Ahmed, E.; Yaqoob, I.; Hashem, I.A.T.; Khan, I.; Ahmed, A.I.A.; Imran, M.; Vasilakos, A.V. The role of Big Data analytics in Internet of Things. Comput. Netw. 2017, 129, 459–471. [Google Scholar] [CrossRef]
  404. Shadroo, S.; Rahmani, A.M. Systematic survey of Big Data and data mining in Internet of things. Comput. Netw. 2018, 139, 19–47. [Google Scholar] [CrossRef]
  405. Ge, M.; Bangui, H.; Buhnova, B. Big Data for Internet of Things: A Survey. Future Gener. Comput. Syst. 2018, 87, 601–614. [Google Scholar] [CrossRef]
  406. Hadi, M.S.; Lawey, A.Q.; El-Gorashi, T.E.H.; Elmirghani, J.M.H. Big Data analytics for wireless and wired network design: A survey. Comput. Netw. 2018, 132, 180–199. [Google Scholar] [CrossRef]
  407. Kulkarni, R.V.; Forster, A.; Venayagamoorthy, G.K. Computational intelligence in wireless sensor networks: A survey. IEEE Commun. Surv. Tutor. 2011, 13, 68–96. [Google Scholar] [CrossRef]
  408. Al-Garadi, M.A.; Mohamed, A.; Al-Ali, A.K.; Du, X.; Ali, I.; Guizani, M. A survey of machine and deep learning methods for Internet of things (IoT) security. IEEE Commun. Surv. Tutor. 2020, 22, 1646–1685. [Google Scholar] [CrossRef] [Green Version]
  409. Atitallah, S.B.; Driss, M.; Boulila, W.; Ghézala, H.B. Leveraging Deep Learning and IoT Big Data analytics to support the Smart Cities development: Review and future directions. Comput. Sci. Rev. 2020, 38, 100303. [Google Scholar] [CrossRef]
  410. Christidis, K.; Devetsikiotis, M. Blockchains and smart contracts for the Internet of things. IEEE Access 2016, 4, 2292–2303. [Google Scholar] [CrossRef]
  411. Zheng, Z.; Xie, S.; Dai, H.; Chen, X.; Wang, H. An overview of blockchain technology: Architecture, consensus, and future trends. In Proceedings of the 2017 IEEE International Congress on Big Data (BigData Congress), Honolulu, HI, USA, 25–30 June 2017; IEEE: Piscataway, NJ, USA, 2017; pp. 557–564. [Google Scholar]
  412. Dorri, A.; Kanhere, S.S.; Jurdak, R.; Gauravaram, P. Blockchain for IoT security and privacy: The case study of a smart home. In Proceedings of the 2017 IEEE International Conference on Pervasive Computing and Communications Workshops (PerCom Workshops), Kona, HI, USA, 13–17 March 2017; IEEE: Piscataway, NJ, USA, 2017; pp. 618–623. [Google Scholar]
  413. Khan, M.A.; Salah, K. IoT security: Review, blockchain solutions, and open challenges. Future Gener. Comput. Syst. 2018, 82, 395–411. [Google Scholar] [CrossRef]
  414. Mistry, I.; Tanwar, S.; Tyagi, S.; Kumar, N. Blockchain for 5G-enabled IoT for industrial automation: A systematic review, solutions, and challenges. Mech. Syst. Signal Process. 2020, 135, 106382. [Google Scholar] [CrossRef]
  415. Kranz, M.; Holleis, P.; Schmidt, A. Embedded interaction: Interacting with the Internet of things. IEEE Internet Comput. 2010, 14, 46–53. [Google Scholar] [CrossRef]
  416. Koreshoff, T.L.; Leong, T.W.; Robertson, T. Approaching a human-centred Internet of things. In Proceedings of the OzCHI: The 25th Australian Computer-Human Interaction Conference, Adelaide, Australia, 25–29 November 2013; pp. 363–366. [Google Scholar]
  417. Atzori, L.; Iera, A.; Morabito, G. Siot: Giving a social structure to the Internet of things. IEEE Commun. Lett. 2011, 15, 1193–1195. [Google Scholar] [CrossRef]
  418. Vinyals, M.; Rodriguez-Aguilar, J.A.; Cerquides, J. A survey on sensor networks from a multiagent perspective. Comput. J. 2011, 54, 455–470. [Google Scholar] [CrossRef] [Green Version]
  419. Wu, Q.; Ding, G.; Xu, Y.; Feng, S.; Du, Z.; Wang, J.; Long, K. Cognitive Internet of things: A new paradigm beyond connection. IEEE Internet Things J. 2014, 1, 129–143. [Google Scholar] [CrossRef] [Green Version]
  420. Kataria, P. Towards a Semantic Classification of Possible Human-to-Environment Interactions in IoT. In International Conference on Human-Computer Interaction; Springer: Berlin/Heidelberg, Germany, 2021; pp. 135–152. [Google Scholar]
  421. Gath-Morad, M.; Aguilar, L.; Dalton, R.C.; Hölscher, C. cogarch: Simulating wayfinding by architecture in multilevel buildings. In Proceedings of the 11th Annual Symposium on Simulation for Architecture and Urban Design, Virtual Event, Austria, 25–27 May 2020; pp. 1–8. [Google Scholar]
  422. Jacko, J.A. Human Computer Interaction Handbook: Fundamentals, Evolving Technologies, and Emerging Applications; CRC Press: Boca Raton, FL, USA, 2012. [Google Scholar]
  423. Helander, M.G. Handbook of Human-Computer Interaction; Elsevier: Amsterdam, The Netherlands, 2014. [Google Scholar]
  424. Chettri, L.; Bera, R. A comprehensive survey on Internet of Things (IoT) toward 5G wireless systems. IEEE Internet Things J. 2019, 7, 16–32. [Google Scholar] [CrossRef]
  425. Jo, D.; Kim, G.J. ARIoT: Scalable Augmented Reality framework for interacting with Internet of Things appliances everywhere. IEEE Trans. Consum. Electron. 2016, 62, 334–340. [Google Scholar] [CrossRef]
  426. Park, Y.; Yun, S.; Kim, K.H. When IoT met Augmented Reality: Visualizing the source of the wireless signal in AR view. In Proceedings of the 17th Annual International Conference on Mobile Systems, Applications, and Services, Seoul, Korea, 17–21 June 2019; pp. 117–129. [Google Scholar]
  427. Michalakis, K.; Aliprantis, J.; Caridakis, G. Visualizing the Internet of things: Naturalizing human–computer interaction by incorporating AR features. IEEE Consum. Electron. Mag. 2018, 7, 64–72. [Google Scholar] [CrossRef]
  428. Carneiro, J.; Rossetti, R.J.; Silva, D.C.; Oliveira, E.C. BIM, GIS, IoT, and AR/VR integration for smart maintenance and management of road networks: A review. In Proceedings of the 2018 IEEE International Smart Cities Conference (ISC2), Kansas City, MO, USA, 16–19 September 2018; IEEE: Piscataway, NJ, USA, 2018; pp. 1–7. [Google Scholar]
  429. Jo, D.; Kim, G.J. AR enabled IoT for a smart and interactive environment: A survey and future directions. Sensors 2019, 19, 4330. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  430. Chen, H.; Hou, L.; Zhang, G.K.; Moon, S. Development of BIM, IoT and AR/VR technologies for fire safety and upskilling. Autom. Constr. 2021, 125, 103631. [Google Scholar] [CrossRef]
  431. Madsen, J.B.; Madsen, C.B. An interactive visualization of the past using a situated simulation approach. In Proceedings of the 2013 Digital Heritage International Congress (DigitalHeritage), Marseille, France, 28 October–1 November 2013; IEEE: Piscataway, NJ, USA, 2013; Volume 1, pp. 307–314. [Google Scholar]
  432. Schall, G.; Mendez, E.; Kruijff, E.; Veas, E.; Junghanns, S.; Reitinger, B.; Schmalstieg, D. Handheld Augmented Reality for underground infrastructure visualization. Pers. Ubiquitous Comput. 2009, 13, 281–291. [Google Scholar] [CrossRef]
  433. Xu, X.; Ding, L.; Luo, H.; Ma, L. From building information modeling to city information modeling. J. Inf. Technol. Constr. 2014, 19, 292–307. [Google Scholar]
  434. ElSayed, N.; Thomas, B.; Marriott, K.; Piantadosi, J.; Smith, R. Situated Analytics. In Proceedings of the 2015 Big Data Visual Analytics (BDVA), Hobart, TAS, Australia, 22–25 September 2015; IEEE: Piscataway, NJ, USA, 2015; pp. 1–8. [Google Scholar]
  435. Zollmann, S.; Poglitsch, C.; Ventura, J. VISGIS: Dynamic situated visualization for geographic information systems. In Proceedings of the 2016 international conference on image and vision computing New Zealand (IVCNZ), Palmerston North, New Zealand, 21–22 November 2016; IEEE: Piscataway, NJ, USA, 2016; pp. 1–6. [Google Scholar]
  436. Abao, R.P.; Malabanan, C.V.; Galido, A.P. Design and development of foodgo: A mobile application using situated analytics to augment product information. Procedia Comput. Sci. 2018, 135, 186–193. [Google Scholar] [CrossRef]
  437. Dave, B.; Buda, A.; Nurminen, A.; Främling, K. A framework for integrating BIM and IoT through open standards. Autom. Constr. 2018, 95, 35–45. [Google Scholar] [CrossRef]
  438. Lochhead, I.; Hedley, N. Communicating Multilevel Evacuation Context Using Situated Augmented Reality. In Proceedings of the ISPRS Annals of Photogrammetry, Remote Sensing & Spatial Information Sciences, Delft, The Netherlands, 1–2 October 2018; Volume 4. [Google Scholar]
  439. Pratt, P.; Ives, M.; Lawton, G.; Simmons, J.; Radev, N.; Spyropoulou, L.; Amiras, D. Through the HoloLens™ looking glass: Augmented Reality for extremity reconstruction surgery using 3D vascular models with perforating vessels. Eur. Radiol. Exp. 2018, 2, 1–7. [Google Scholar] [CrossRef]
  440. Stojanovic, V.; Trapp, M.; Richter, R.; Hagedorn, B.; Döllner, J. Towards the generation of Digital Twins for facility management based on 3D point clouds. Management 2018, 270, 279. [Google Scholar]
  441. Lee, E.Y.; Tran, V.T.; Kim, D. A novel head mounted display based methodology for balance evaluation and rehabilitation. Sustainability 2019, 11, 6453. [Google Scholar] [CrossRef] [Green Version]
  442. Guarese, R.; Becker, J.; Fensterseifer, H.; Walter, M.; Freitas, C.; Nedel, L.; Maciel, A. Augmented situated visualization for spatial and context-aware decision-making. In Proceedings of the International Conference on Advanced Visual Interfaces, Salerno, Italy, 28 September–2 October 2020; pp. 1–5. [Google Scholar]
  443. Mourtzis, D.; Siatras, V.; Angelopoulos, J. Real-time remote maintenance support based on Augmented Reality (AR). Appl. Sci. 2020, 10, 1855. [Google Scholar] [CrossRef] [Green Version]
  444. Vidal-Balea, A.; Blanco-Novoa, O.; Picallo-Guembe, I.; Celaya-Echarri, M.; Fraga-Lamas, P.; Lopez-Iturri, P.; Azpilicueta, L.; Falcone, F.; Fernández-Caramés, T.M. Analysis, design and practical validation of an Augmented Reality teaching system based on microsoft HoloLens 2 and edge computing. Eng. Proc. 2020, 2, 52. [Google Scholar]
  445. Vidal-Balea, A.; Blanco-Novoa, O.; Fraga-Lamas, P.; Vilar-Montesinos, M.; Fernández-Caramés, T.M. Creating collaborative Augmented Reality experiences for industry 4.0 training and assistance applications: Performance evaluation in the shipyard of the future. Appl. Sci. 2020, 10, 9073. [Google Scholar] [CrossRef]
  446. Tang, S.; Shelden, D.R.; Eastman, C.M.; Pishdad-Bozorgi, P.; Gao, X. A review of building information modeling (BIM) and the Internet of things (IoT) devices integration: Present status and future trends. Autom. Constr. 2019, 101, 127–139. [Google Scholar] [CrossRef]
  447. Ang, L.M.; Seng, K.P. Big sensor data applications in urban environments. Big Data Res. 2016, 4, 1–12. [Google Scholar] [CrossRef]
  448. Boubiche, S.; Boubiche, D.E.; Bilami, A.; Toral-Cruz, H. Big Data challenges and data aggregation strategies in wireless sensor networks. IEEE Access 2018, 6, 20558–20571. [Google Scholar] [CrossRef]
  449. Jones, D.; Snider, C.; Nassehi, A.; Yon, J.; Hicks, B. Characterising the Digital Twin: A systematic literature review. CIRP J. Manuf. Sci. Technol. 2020, 29, 36–52. [Google Scholar] [CrossRef]
  450. Zhou, F.; Duh, H.B.L.; Billinghurst, M. Trends in Augmented Reality tracking, interaction and display: A review of ten years of ISMAR. In Proceedings of the 2008 7th IEEE/ACM International Symposium on Mixed and Augmented Reality, Cambridge, UK, 15–18 September 2008; IEEE: Piscataway, NJ, USA, 2008; pp. 193–202. [Google Scholar]
  451. Nizam, S.M.; Abidin, R.Z.; Hashim, N.C.; Lam, M.C.; Arshad, H.; Majid, N. A review of multimodal interaction technique in Augmented Reality environment. Int. J. Adv. Sci. Eng. Inf. Technol 2018, 8, 4–8. [Google Scholar]
  452. Kim, K.; Billinghurst, M.; Bruder, G.; Duh, H.B.L.; Welch, G.F. Revisiting trends in Augmented Reality research: A review of the 2nd decade of ISMAR (2008–2017). IEEE Trans. Vis. Comput. Graph. 2018, 24, 2947–2962. [Google Scholar] [CrossRef]
  453. Liu, X.; Wang, X.; Wright, G.; Cheng, J.C.; Li, X.; Liu, R. A state-of-the-art review on the integration of Building Information Modeling (BIM) and Geographic Information System (GIS). ISPRS Int. J. Geo-Inf. 2017, 6, 53. [Google Scholar] [CrossRef] [Green Version]
  454. Helbing, D.; Fanitabasi, F.; Giannotti, F.; Hänggli, R.; Hausladen, C.I.; van den Hoven, J.; Mahajan, S.; Pedreschi, D.; Pournaras, E. Ethics of Smart Cities: Towards value-sensitive design and co-evolving city life. Sustainability 2021, 13, 11162. [Google Scholar] [CrossRef]
  455. Fonnet, A.; Prie, Y. Survey of immersive analytics. IEEE Trans. Vis. Comput. Graph. 2019, 27, 2101–2122. [Google Scholar] [CrossRef]
Figure 1. Comparison of a Digital Twin and Fused Twins. In both rows, the same lobby of a building is shown. (a) the DT of a sensor system with 390 nodes is visualised in Virtual Reality (VR) [20,21]. The white areas visualise the areas of observation and the coloured spheres encode the type of sensor that is located at the respective position in the building. (b) The FTs [17] are visualised in Mixed Reality (MR) [22] by embedding the DT in situ in its PT, thus fusing the two. The fused representation enables Situated Analytics [19] which facilitates the user’s understanding of analysis through spatial cues of the environment.
Figure 1. Comparison of a Digital Twin and Fused Twins. In both rows, the same lobby of a building is shown. (a) the DT of a sensor system with 390 nodes is visualised in Virtual Reality (VR) [20,21]. The white areas visualise the areas of observation and the coloured spheres encode the type of sensor that is located at the respective position in the building. (b) The FTs [17] are visualised in Mixed Reality (MR) [22] by embedding the DT in situ in its PT, thus fusing the two. The fused representation enables Situated Analytics [19] which facilitates the user’s understanding of analysis through spatial cues of the environment.
Remotesensing 14 03095 g001
Figure 2. Components of a Digital Twin. A DT is often thought of as the interaction between the PT and the DT [2], but any DT implements five core components, even if only partially. These components are the Physical Environment, the Data Environment, the Analytical Environment, the Virtual Environment, and the Connection Environment [23]. The PT is measured through the Physical Environment of the DT, and the raw data are stored in the Data Environment. The Analytical Environment provides simulations, services and automatisation. Users often only perceive the DT represented in the Virtual Environment as a model of the PT without being aware of the involvement of all the other components. The Connection Environment is the invisible glue that holds the different components together and enables composition with other DTs if properly implemented. Generated with http://draw.io.
Figure 2. Components of a Digital Twin. A DT is often thought of as the interaction between the PT and the DT [2], but any DT implements five core components, even if only partially. These components are the Physical Environment, the Data Environment, the Analytical Environment, the Virtual Environment, and the Connection Environment [23]. The PT is measured through the Physical Environment of the DT, and the raw data are stored in the Data Environment. The Analytical Environment provides simulations, services and automatisation. Users often only perceive the DT represented in the Virtual Environment as a model of the PT without being aware of the involvement of all the other components. The Connection Environment is the invisible glue that holds the different components together and enables composition with other DTs if properly implemented. Generated with http://draw.io.
Remotesensing 14 03095 g002
Figure 3. The Smart City with a Digital Twin Foundation. Smart Cities rely on effective data collection, analysis, and communication to work. A DT can provide the required processing of the data collection systems to formalise the PT. In a Smart City, a DT provides analysis for governance, management, and entertainment. While governance refers to directing the city’s future, management refers to the present state of the city, and entertainment refers to the use of the city. Generated with http://draw.io.
Figure 3. The Smart City with a Digital Twin Foundation. Smart Cities rely on effective data collection, analysis, and communication to work. A DT can provide the required processing of the data collection systems to formalise the PT. In a Smart City, a DT provides analysis for governance, management, and entertainment. While governance refers to directing the city’s future, management refers to the present state of the city, and entertainment refers to the use of the city. Generated with http://draw.io.
Remotesensing 14 03095 g003
Figure 4. Composition of Digital Twins to describe higher order complex systems. Multiple DTs are composed into higher order systems such as cities. Here, an open stationary city DT is composed of different open or closed and mobile or stationary DTs such as building DTs, transportation DTs, and infrastructure DTs. Each of the DTs contains complex nested DTs of different types such as for healthcare or industrial settings. The complexity level can be arbitrarily increased through nesting. Generated with http://draw.io.
Figure 4. Composition of Digital Twins to describe higher order complex systems. Multiple DTs are composed into higher order systems such as cities. Here, an open stationary city DT is composed of different open or closed and mobile or stationary DTs such as building DTs, transportation DTs, and infrastructure DTs. Each of the DTs contains complex nested DTs of different types such as for healthcare or industrial settings. The complexity level can be arbitrarily increased through nesting. Generated with http://draw.io.
Remotesensing 14 03095 g004
Figure 5. Immersive Analytics compared to Situated Analytics. The real world is schematically represented in grey. AR content is shown in blue. On the left side, a user is displaying a Smart City environment (blue model) in non-situated AR. Interactables are highlighted in white, and several pop-ups showing information (white squares) are opened. Immersive Analytics make use of 3D space to visualise content but can be independent of the spatial context surrounding them. On the right side, a user (in yellow) is walking through the Physical Environment of the Smart City. In Situated Analytics, interactables are AR content that is displayed in situ for the Physical Environment that they describe. Similarly to (a), pop-ups (blue squares) are showing content to the user in AR space. Generated with https://miro.com and https://rhino3d.com with royalty-free models from https://cgtrader.com.
Figure 5. Immersive Analytics compared to Situated Analytics. The real world is schematically represented in grey. AR content is shown in blue. On the left side, a user is displaying a Smart City environment (blue model) in non-situated AR. Interactables are highlighted in white, and several pop-ups showing information (white squares) are opened. Immersive Analytics make use of 3D space to visualise content but can be independent of the spatial context surrounding them. On the right side, a user (in yellow) is walking through the Physical Environment of the Smart City. In Situated Analytics, interactables are AR content that is displayed in situ for the Physical Environment that they describe. Similarly to (a), pop-ups (blue squares) are showing content to the user in AR space. Generated with https://miro.com and https://rhino3d.com with royalty-free models from https://cgtrader.com.
Remotesensing 14 03095 g005
Figure 6. Examples of Embodied Digital Twins and Fused Twins. We observe the difference between non-situated AR (top row) and situated AR (bottom row) for anchoring content at the body of the user, an object, or the immediate environment. The distinction between body-, object-, and environment-centred AR is inspired by the classification of different physicalities of embodiment offered by [90,91]. With a body anchor, the DT is anchored on the users’ bodies, similar to the directly embodied approach. With an object anchor, the DT is anchored on an object with which the user can interact, similar to the manipulated approach. With an environment anchor, the DT is anchored in a room, similar to the augmented approach. In the situated environment (implementing the FTs paradigm), the relationship between the DT and the representation is homologous such that the body, object, or environment is used to anchor its own DT in a one-to-one correspondence. However, in the non-situated environment, this relationship is non-homologous because the mapping from the DT to the representation is transformation instead of a one-to-one correspondence.
Figure 6. Examples of Embodied Digital Twins and Fused Twins. We observe the difference between non-situated AR (top row) and situated AR (bottom row) for anchoring content at the body of the user, an object, or the immediate environment. The distinction between body-, object-, and environment-centred AR is inspired by the classification of different physicalities of embodiment offered by [90,91]. With a body anchor, the DT is anchored on the users’ bodies, similar to the directly embodied approach. With an object anchor, the DT is anchored on an object with which the user can interact, similar to the manipulated approach. With an environment anchor, the DT is anchored in a room, similar to the augmented approach. In the situated environment (implementing the FTs paradigm), the relationship between the DT and the representation is homologous such that the body, object, or environment is used to anchor its own DT in a one-to-one correspondence. However, in the non-situated environment, this relationship is non-homologous because the mapping from the DT to the representation is transformation instead of a one-to-one correspondence.
Remotesensing 14 03095 g006
Figure 7. Situated Analytics [19,126] within a DT context. Situated Analytics use the Digital/Physical Divide to define the generative steps in situ in immersive analytics in the physical world (bottom row) or digital world (top row). There is a remarkable similarity between the components of DTs (blue boxes) and the generation of Situated Analytics (white boxes). The mapping is visually nested to be highlight. An observer in the physical world sees the physical referent or PT (in the DT context). The observer also sees the physical presentation of information (Virtual Environment of a DT). In the digital world, the raw data are received from the referent, which covers the processes of gathering data by the Physical Environment and storing data by the Data Environment. The Visualisation Pipeline produces a visual output (Analytical Environment of a DT). Lastly, the data are placed into the physical presentation (Virtual Environment). Generated with http://draw.io.
Figure 7. Situated Analytics [19,126] within a DT context. Situated Analytics use the Digital/Physical Divide to define the generative steps in situ in immersive analytics in the physical world (bottom row) or digital world (top row). There is a remarkable similarity between the components of DTs (blue boxes) and the generation of Situated Analytics (white boxes). The mapping is visually nested to be highlight. An observer in the physical world sees the physical referent or PT (in the DT context). The observer also sees the physical presentation of information (Virtual Environment of a DT). In the digital world, the raw data are received from the referent, which covers the processes of gathering data by the Physical Environment and storing data by the Data Environment. The Visualisation Pipeline produces a visual output (Analytical Environment of a DT). Lastly, the data are placed into the physical presentation (Virtual Environment). Generated with http://draw.io.
Remotesensing 14 03095 g007
Figure 8. Digital Shadows and Digital Models in the context of Digital Twins. The left side demonstrates the relations between a Digital Shadow, a digital model and a DT [152]. The right side demonstrates the mapping to our definition. The Digital Shadow corresponds to the Data Environment and relies on the Physical Environment to collect data. The Engineering, Digital Model and their DT simulation correspond to the Analytical Environment. The experimentable DT is akin to the Virtual Environment. In our definition, all blocks are components of the DT, whereas in the other definition, the DT is a piece in the diagram. The reduction of the DT to a simulation removes the HCI from the system and requires terms such as Digital Shadow and Digital Models to differentiate DTs. Under our definition, these terms reflect different aspects of a DT and map to different components. Generated with http://draw.io.
Figure 8. Digital Shadows and Digital Models in the context of Digital Twins. The left side demonstrates the relations between a Digital Shadow, a digital model and a DT [152]. The right side demonstrates the mapping to our definition. The Digital Shadow corresponds to the Data Environment and relies on the Physical Environment to collect data. The Engineering, Digital Model and their DT simulation correspond to the Analytical Environment. The experimentable DT is akin to the Virtual Environment. In our definition, all blocks are components of the DT, whereas in the other definition, the DT is a piece in the diagram. The reduction of the DT to a simulation removes the HCI from the system and requires terms such as Digital Shadow and Digital Models to differentiate DTs. Under our definition, these terms reflect different aspects of a DT and map to different components. Generated with http://draw.io.
Remotesensing 14 03095 g008
Figure 9. Example Composition of DTs. Multiple Smart Objects (SOs) and their DTs are composed into a DT of a Household. Remote sensing provides the external context to the building. Locally, FTs (SOs) provide physical user interfaces, enabling end users to directly interact with the DT in a natural way. Generated with http://draw.io.
Figure 9. Example Composition of DTs. Multiple Smart Objects (SOs) and their DTs are composed into a DT of a Household. Remote sensing provides the external context to the building. Locally, FTs (SOs) provide physical user interfaces, enabling end users to directly interact with the DT in a natural way. Generated with http://draw.io.
Remotesensing 14 03095 g009
Figure 10. Composition of Components to form Digital Twins. Each DT is represented as a column. The Connection Environment (represented by arrows) connects instances in all environments across DTs. The other environments are aligned in rows to show the implementation across DTs. Instances are shown in each environment as rectangles, cylinders, triangles, or circles. Data from the Data Environment can processed in the Analytical Environment or displayed directly in the Virtual Environment. DTs can also be siloed (e.g., DT C) but more often will use the Connection Environment with complex dependencies across all environments. For example, DT A uses Data 3 and Analysis 3 from PT B. PT B has two DTs where DT B2 is composed into DT B1 in the Virtual Environment by combining interface 2 and 3. DT B1 and B2 use data that is collected from all PTs. Despite DT C being siloed, PT C is evaluated in DT A through Data 3 and Analysis 2. Note that this abstraction makes no claims about the mode of data collection in the Physical Environment that may be with either local sensors or remote sensing. Generated with http://draw.io.
Figure 10. Composition of Components to form Digital Twins. Each DT is represented as a column. The Connection Environment (represented by arrows) connects instances in all environments across DTs. The other environments are aligned in rows to show the implementation across DTs. Instances are shown in each environment as rectangles, cylinders, triangles, or circles. Data from the Data Environment can processed in the Analytical Environment or displayed directly in the Virtual Environment. DTs can also be siloed (e.g., DT C) but more often will use the Connection Environment with complex dependencies across all environments. For example, DT A uses Data 3 and Analysis 3 from PT B. PT B has two DTs where DT B2 is composed into DT B1 in the Virtual Environment by combining interface 2 and 3. DT B1 and B2 use data that is collected from all PTs. Despite DT C being siloed, PT C is evaluated in DT A through Data 3 and Analysis 2. Note that this abstraction makes no claims about the mode of data collection in the Physical Environment that may be with either local sensors or remote sensing. Generated with http://draw.io.
Remotesensing 14 03095 g010
Figure 11. Mapping of the “as a Service” paradigm to the components of a Digital Twin and the weather services example. This example application is an alert for a nearby weather event. Presented from left to right, the PT forms the Physical Environment of the DT and is captured by local sensors and networks and remote sensing as IaaS. The Data Environment of the DT serves as a PaaS. The Analytical Environment of the DT models the weather and calculates the proximity to the predicted event as a SaaS. The Virtual Environment is implemented as a smartphone application that provides an alert on an abstract Dashboard. The Connection Environment spans across all services and environments providing an API that allows for one environment to be replaced by another. This makes the DT both modular and composable. Generated with http://draw.io.
Figure 11. Mapping of the “as a Service” paradigm to the components of a Digital Twin and the weather services example. This example application is an alert for a nearby weather event. Presented from left to right, the PT forms the Physical Environment of the DT and is captured by local sensors and networks and remote sensing as IaaS. The Data Environment of the DT serves as a PaaS. The Analytical Environment of the DT models the weather and calculates the proximity to the predicted event as a SaaS. The Virtual Environment is implemented as a smartphone application that provides an alert on an abstract Dashboard. The Connection Environment spans across all services and environments providing an API that allows for one environment to be replaced by another. This makes the DT both modular and composable. Generated with http://draw.io.
Remotesensing 14 03095 g011
Figure 12. Google search trends by search term in web searches. Interests are averaged with LOESS across months from 2003 to early 2022. In general, the Internet of Things has been falling since 2017, and Digital Twin has been steadily rising. This year, Digital Twin has become the most searched term among the cognates other than Internet of Things. Terms for underlying technologies and previous system names are stagnating or declining in use. Data collected from https://trends.google.com on 8 February 2022.
Figure 12. Google search trends by search term in web searches. Interests are averaged with LOESS across months from 2003 to early 2022. In general, the Internet of Things has been falling since 2017, and Digital Twin has been steadily rising. This year, Digital Twin has become the most searched term among the cognates other than Internet of Things. Terms for underlying technologies and previous system names are stagnating or declining in use. Data collected from https://trends.google.com on 8 February 2022.
Remotesensing 14 03095 g012
Figure 13. Computing Hydrology of computation-on-the-wire. Computing power can be located locally at the sensor/actuator or more remotely at the gateway, a local cluster, or remote server. The computations form the basis for a DT. Computing in IoT and Smart Cities is often conceptualised with hydrological terminology (i.e., cloud, fog, mist, and dew) that maps to the computing location and encapsulates different trade-offs in computational intensity, traffic, latency, accuracy, energy consumption, and security [42,191]. These trade-offs are best evaluated under the subsidiarity principle to decide where to place the computation physically. Edge or grid computing [150,151] is another terminology to cover all non-cloud computing. DTs rely heavily on these different data and computing sources while at the same time abstracting away their existence in the end users’ view. Generated with http://draw.io.
Figure 13. Computing Hydrology of computation-on-the-wire. Computing power can be located locally at the sensor/actuator or more remotely at the gateway, a local cluster, or remote server. The computations form the basis for a DT. Computing in IoT and Smart Cities is often conceptualised with hydrological terminology (i.e., cloud, fog, mist, and dew) that maps to the computing location and encapsulates different trade-offs in computational intensity, traffic, latency, accuracy, energy consumption, and security [42,191]. These trade-offs are best evaluated under the subsidiarity principle to decide where to place the computation physically. Edge or grid computing [150,151] is another terminology to cover all non-cloud computing. DTs rely heavily on these different data and computing sources while at the same time abstracting away their existence in the end users’ view. Generated with http://draw.io.
Remotesensing 14 03095 g013
Figure 14. Devices to access Digital Twins on the Reality–Virtuality Continuum [360]. On the Continuum, FTs span instances of Extended Reality (XR) input modalities and media from Augmented Reality (AR) via Mixed Reality (MR) to Augmented Virtuality (AV). Typical media include tablets and phones but also holographic displays and smart objects. Reality-embedded access to DT focuses on abstract Dashboards, whereas virtuality-embedded access focuses on remote analysis of the DT. The FTs spectrum covers direct interaction with the PT that is fused with augmentations from the DT either displayed via AR on tablet, MR on HoloLens, or haptically through AV smart objects. Generated with http://draw.io.
Figure 14. Devices to access Digital Twins on the Reality–Virtuality Continuum [360]. On the Continuum, FTs span instances of Extended Reality (XR) input modalities and media from Augmented Reality (AR) via Mixed Reality (MR) to Augmented Virtuality (AV). Typical media include tablets and phones but also holographic displays and smart objects. Reality-embedded access to DT focuses on abstract Dashboards, whereas virtuality-embedded access focuses on remote analysis of the DT. The FTs spectrum covers direct interaction with the PT that is fused with augmentations from the DT either displayed via AR on tablet, MR on HoloLens, or haptically through AV smart objects. Generated with http://draw.io.
Remotesensing 14 03095 g014
Figure 15. Comparison of Digital Twin Architectures. Data flows from the bottom (Thing) to the top (User). The functionality of each layer is displayed in comparison to other architectures where possible by spawning each layer vertically such that it contains all functionality that similar layers in other architectures provide. For example, Presentation and Session in the OSI Reference Model are subsumed in Application in the IETF LLN Stack. The labels at the top provide the name of the models, and the labels at the bottom references the origin of the respective model. From left to right: OSI Reference Model, Variation of IETF Internet Model [316], Digital Twin [2], CPS [234], ETSI M2M [165] and SENSEI [368], SoA [154], 5C [155], and Context Aware [156]. Generally, management and security columns spawning all layers are not shown. Our own definition is mapped to these architectures in the last column. Generated with http://draw.io.
Figure 15. Comparison of Digital Twin Architectures. Data flows from the bottom (Thing) to the top (User). The functionality of each layer is displayed in comparison to other architectures where possible by spawning each layer vertically such that it contains all functionality that similar layers in other architectures provide. For example, Presentation and Session in the OSI Reference Model are subsumed in Application in the IETF LLN Stack. The labels at the top provide the name of the models, and the labels at the bottom references the origin of the respective model. From left to right: OSI Reference Model, Variation of IETF Internet Model [316], Digital Twin [2], CPS [234], ETSI M2M [165] and SENSEI [368], SoA [154], 5C [155], and Context Aware [156]. Generally, management and security columns spawning all layers are not shown. Our own definition is mapped to these architectures in the last column. Generated with http://draw.io.
Remotesensing 14 03095 g015
Figure 16. Fused Twins Torus Stack. Generalised technology stack for the FTs paradigm based on [17]. The stack implements a full DT and expands the Virtual Environment to accommodate the FTs. Data flow is shown through the arrows. The stack becomes a torus because the chain of technologies and environments requires circular links of the data. Changes to a DT in the Virtual Environment flow back to the PT through the Data Environment. Generated with http://draw.io.
Figure 16. Fused Twins Torus Stack. Generalised technology stack for the FTs paradigm based on [17]. The stack implements a full DT and expands the Virtual Environment to accommodate the FTs. Data flow is shown through the arrows. The stack becomes a torus because the chain of technologies and environments requires circular links of the data. Changes to a DT in the Virtual Environment flow back to the PT through the Data Environment. Generated with http://draw.io.
Remotesensing 14 03095 g016
Table 1. Taxonomy of Digital Twin Context Types.
Table 1. Taxonomy of Digital Twin Context Types.
System Type
OpenClosed
Location TypeStationaryCities, Economies & CountriesIndustrial processes
MobileEnd users, goods & vehiclesBody (health) & machines (internal)
Table 2. Task Classification for Active and Passive Devices.
Table 2. Task Classification for Active and Passive Devices.
TaskDescriptionActivePassive
AccessingAccessing information on an object linked to the device.
ActuatingTriggering an action of the object linked to the device.
LocatingIdentifying the location of an object linked to the device.
SensingMeasure characteristic of the object linked to the device.
Table 3. Waves of IoT generations.
Table 3. Waves of IoT generations.
Generation IGeneration IIGeneration IIIGeneration IV
Earliest Mention1999201020112012
TechnologiesRFIDInter-networkingCloud computingFog Computing
M2MWeb of ThingsICNDTs
Object integrationsConstrained devicesSemantic IoTOpportunistic IoT
FocusDigitalisationNetworkingCentralisationDecentralisation
Source[232,265][232,282][232,268]own definition, [2,283,284]
Table 4. Communication technologies for PHY layer.
Table 4. Communication technologies for PHY layer.
ScopeAbbreviationTechnologiesReviews
Body AreaBANBLE, Zigbee[207,302,303,304,305]
Personal AreaPANUWB, Bluetooth, WiFi[306,307]
Local AreaLANWiFi, LTE[307,308]
Wide AreaWANLoRaWAN, Sigfox, NB-IoT[20,309,310,311,312,313,314]
Table 5. Research Areas for Digital Twin Analytics.
Table 5. Research Areas for Digital Twin Analytics.
DomainDescriptionReview
OntologiesMake semantic sense of the data and produce knowledge and context [156,225,226,227,387,388,389]
MiddlewareInfrastructure to identify, process, store, secure and present information[228,229,230,390,391]
Data preprocessingData aggregation and data fusion [304,392,393,394,395,396,397,398]
Data miningCloud-based Big Data analysis on unstructured data[399,400,401,402,403,404,405,406]
AI&MLMachine learning and deep learning [396,407,408,409]
BlockchainSecurity, trust, and privacy through smart contracts[410,411,412,413,414]
HCIHuman-Computer Interaction for improved access [70,85,86,260,415,416,417]
ABMAgent-based modelling Objects[418,419,420,421]
Table 6. Recent and Ongoing Research Applying the Fused Twins Paradigm.
Table 6. Recent and Ongoing Research Applying the Fused Twins Paradigm.
Environments
Application TypeYearPhysicalDataAnalyticalVirtualConnection
Infrastructure Management [432]2009**-
City Information [18]2009*-
History [431]2013*--
City Information (concept) [433]2014*****
Shopping [434]2015**-
IoT device interaction [425]2016-*
City Information [435]2016**-
Shopping [436]2018**-
Application TypeYearPhysicalDataAnalyticalVirtualConnection
Smart Campus [437]2018
Evacuation [438]2018*---
Medicine [439]2018--
Facility Management [440]2018-
Rehabilitation [441]2019*-
IoT device visualisation [426]2019---
Seat Selection [442]2020**-
Equipment Maintenance [443]2020*--
In situ visualisation [101]2020***
Education [444]2020*
Training & Assistance [445]2020**-*
Firefighting [430]2021---
General [17]2021*
IoT device interaction [364]2021-
General [100]2022*
Note: * Partially implemented or insufficient information.
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Grübel, J.; Thrash, T.; Aguilar, L.; Gath-Morad, M.; Chatain, J.; Sumner, R.W.; Hölscher, C.; Schinazi, V.R. The Hitchhiker’s Guide to Fused Twins: A Review of Access to Digital Twins In Situ in Smart Cities. Remote Sens. 2022, 14, 3095. https://doi.org/10.3390/rs14133095

AMA Style

Grübel J, Thrash T, Aguilar L, Gath-Morad M, Chatain J, Sumner RW, Hölscher C, Schinazi VR. The Hitchhiker’s Guide to Fused Twins: A Review of Access to Digital Twins In Situ in Smart Cities. Remote Sensing. 2022; 14(13):3095. https://doi.org/10.3390/rs14133095

Chicago/Turabian Style

Grübel, Jascha, Tyler Thrash, Leonel Aguilar, Michal Gath-Morad, Julia Chatain, Robert W. Sumner, Christoph Hölscher, and Victor R. Schinazi. 2022. "The Hitchhiker’s Guide to Fused Twins: A Review of Access to Digital Twins In Situ in Smart Cities" Remote Sensing 14, no. 13: 3095. https://doi.org/10.3390/rs14133095

APA Style

Grübel, J., Thrash, T., Aguilar, L., Gath-Morad, M., Chatain, J., Sumner, R. W., Hölscher, C., & Schinazi, V. R. (2022). The Hitchhiker’s Guide to Fused Twins: A Review of Access to Digital Twins In Situ in Smart Cities. Remote Sensing, 14(13), 3095. https://doi.org/10.3390/rs14133095

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop