Next Article in Journal
LARO: Opposition-Based Learning Boosted Artificial Rabbits-Inspired Optimization Algorithm with Lévy Flight
Next Article in Special Issue
A Recognition Method of Ancient Architectures Based on the Improved Inception V3 Model
Previous Article in Journal
Controlling Problem within a Class of Two-Level Positive Maps
Previous Article in Special Issue
Hand Gesture Recognition with Symmetric Pattern under Diverse Illuminated Conditions Using Artificial Neural Network
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Distributed Network of Adaptive and Self-Reconfigurable Active Vision Systems

Department of Electronics and Communication Engineering, Delhi Technological University, Delhi 110042, India
*
Author to whom correspondence should be addressed.
Symmetry 2022, 14(11), 2281; https://doi.org/10.3390/sym14112281
Submission received: 29 September 2022 / Revised: 18 October 2022 / Accepted: 25 October 2022 / Published: 31 October 2022
(This article belongs to the Special Issue Computer Vision, Pattern Recognition, Machine Learning, and Symmetry)

Abstract

:
The performance of a computer vision system depends on the accuracy of visual information extracted by the sensors and the system’s visual-processing capabilities. To derive optimum information from the sensed data, the system must be capable of identifying objects of interest (OOIs) and activities in the scene. Active vision systems intend to capture OOIs with the highest possible resolution to extract the optimum visual information by calibrating the configuration spaces of the cameras. As the data processing and reconfiguration of cameras are interdependent, it becomes very challenging for advanced active vision systems to perform in real time. Due to limited computational resources, model-based asymmetric active vision systems only work in known conditions and fail miserably in unforeseen conditions. Symmetric/asymmetric systems employing artificial intelligence, while they manage to tackle unforeseen environments, require iterative training and thus are not reliable for real-time applications. Thus, the contemporary symmetric/asymmetric reconfiguration systems proposed to obtain optimum configuration spaces of sensors for accurate activity tracking and scene understanding may not be adequate to tackle unforeseen conditions in real time. To address this problem, this article presents an adaptive self-reconfiguration (ASR) framework for active vision systems operating co-operatively in a distributed blockchain network. The ASR framework enables active vision systems to share their derived learning about an activity or an unforeseen environment, which learning can be utilized by other active vision systems in the network, thus lowering the time needed for learning and adaptation to new conditions. Further, as the learning duration is reduced, the duration of the reconfiguration of the cameras is also reduced, yielding better performance in terms of understanding of a scene. The ASR framework enables resource and data sharing in a distributed network of active vision systems and outperforms state-of-the-art active vision systems in terms of accuracy and latency, making it ideal for real-time applications.

1. Introduction

A smart camera network (SCN), defined by Reisslein et al. [1], is a real-time distributed embedded system that is configured to perform computer-vision tasks by processing sensor data obtained from a plurality of cameras via cooperative sensing. The smart camera network is generally deployed to perform complex computer-vision tasks that require more than one camera to extract visual information. Visual surveillance of large areas [2], complex sports analytics [3], situation cognizance [4], tele-immersion, automated driverless vehicles [5], ambient assisted living [6], computer-vision-based disaster management, robotic vision, and other applications rely on SCNs to achieve desired functionalities. The smart camera network aims to co-operatively fuse sensor data to develop scene understanding and further associates scene understanding with a control system configured to achieve the overall functionality of a computer vision system.
The advancement in computer-vision applications has taken a tremendous leap in the last decade. According to a survey reported in [2], the growth of computer-vision applications is estimated at a compound annual growth rate (CAGR) of 7.6% from 2020 to 2027. Computer vision systems [7] aim to leverage human efforts by developing an understanding of events through the processing of data obtained from a number of sensors (hereinafter interchangeably referred to as “cameras” or “camera sensors”). Further, based on understanding, most of the smart computer vision systems aim to enable the operation of control systems to deliver a desired functionality automatically without any human interaction. An ideal computer vison system should extract data from the activities detected by the sensors, generate an understanding based on processing of the data, and provide optimal functionality through an appropriate control system.
Computer vision systems [7] rely on the processing of digital images or videos obtained from one or more cameras to obtain an understanding of an activity or an event. Understanding of the activity or event highly depends on the quality of data in terms of information about objects of interest. To obtain optimum information about an OOI from captured data, the cameras must be configured to capture OOIs with the highest possible resolution. In some respects, it can be concluded that the objects of interest must be captured as close to the center of the camera’s field of view as possible.
A computer vision system [7] needs to process the data to identify an OOI and critical activities in the scene to derive scene understanding, based on which the cameras can be reconfigured to capture the OOI in the center of the FOV. Thus, data processing and camera calibration are interdependent. A system capable of reconfiguring the parameters of the cameras to manipulate the viewpoints of the cameras in order to investigate the environment and obtain better information from it is known as an active vision system. As reconfiguration of the cameras and the processing performance of the active vision system are interdependent, designing advanced active vision systems employing a network of cameras co-operatively working towards a desired functionality in real time is very challenging.
Traditional model-based active vision systems have limited computational resources, due to which they can only identify activities in known environments for which they are designed and fail badly to identify and understand new activities in unforeseen conditions. Active vision systems employing artificial intelligence (AI) manage to tackle unforeseen conditions; however, due to iterative training processes, they are not reliable for real-time applications. The impacts of unforeseen conditions and uncertainties in computer vision systems are presented in [2].
To address the abovementioned problem, this article showcases an adaptive self-reconfiguration (ASR) framework for active vision systems operating co-operatively in a distributed blockchain network. The ASR framework facilitates active vision systems to share information about their learning towards new activities and unforeseen environments with other systems in the distributed network. The information shared can be utilized by any system in the distributed blockchain network to tackle an identical condition and thus saves a lot of time which would otherwise have been taken up with iterative model training. Further, as the learning duration is reduced, the duration of the reconfiguration of the cameras is also reduced, thus yielding better performance in terms of understanding a scene.
To develop a good understanding of the abovementioned reconfiguration problem and its impact on the performance of an active vision system, this article primarily provides a detailed discussion of the challenges in developing an active vision system at different operational levels. We further provide an extended survey of various systems and methods proposed for addressing the challenges. Additionally, this article highlights a trend in the state-of-the-art systems and methods proposed to address the active vision challenges and showcases a common threat for most of the contemporary solutions. This article further provides definitions of the concepts of self-reconfiguration and self-adaptation and presents a role for self-reconfiguration and self-adaptation in the enhancement of the performance of active vision systems. Finally, this article provides an adaptive self-reconfiguration (ASR) framework to enhance the performance of active vision systems and their applications.
A computer vision system generally includes one or more camera sensors, software or circuitry for data processing, and a control system. The camera sensors are configured for co-operative sensing in the operating environment and gathering raw sensor data. The raw sensor data are fused to obtain pre-processed visual information that is utilized for visual processing. The visual processing includes several steps to process and furnish the sensor data and obtain a visual understanding. A control signal is generated based on the visual understanding and is transmitted to the control system. Based on the control signal received, the control system actuates one or more components of the system such that the desired functionality of the system is achieved. The abovementioned process is depicted in Figure 1.
A multi-camera active vision system [8] (hereinafter interchangeably referred to as an “active vision system”) is basically a computer vision system that employs an SCN with the capability of altering the viewpoints of the sensor nodes, thereby yielding better data for processing. The operation of active vision systems can be differentiated into two levels: a deployment level (i.e., for data extraction) and a processing level (i.e., where the visual processing takes place to obtain an understanding). The challenges associated with the deployment of an active vision system employing one or more SCNs can also be classified into two categories. Challenges of the first type [9] are associated with reconfiguration of the sensor nodes to obtain sensor data bearing optimal information, relative to each camera’s resource limitations. Challenges of the second type [2] are associated with the data-processing level of operation, occurring at the time of processing the data to develop an understanding of a scene while keeping the computational complexity as low as possible. Thus, an efficient active vision system demands efforts at both hardware and software levels of deployment. Detailed reviews of the challenges at the data-processing level and the sensor-calibration level of active vision systems are presented in [2] and [9], respectively. A multi-tier taxonomy of challenges for active vision systems employing camera networks on the basis of operational classification is shown in Figure 2.

2. Challenges

A smart camera network (SCN) enhances the availability of raw data and thus the chances of deriving better information; however, it also increases the computational complexity of the active vision system of which it is a part. The configuration space of each sensor depends on the intrinsic parameters (such as optical center, focal length, etc.) and the extrinsic parameters of the sensor (such as rotation and translation, etc.), which require calibration [8].
Apparently, an active vision system utilizing a single fixed camera has much lower computational complexity and thus faces fewer challenges at the deployment level than an active vision system employing a network of mobile pan–tilt–zoom (PTZ) cameras. However, an active vision system employing multiple camera sensors can have an occlusion-avoidance capability that may be lacking in an active vision system employing a single camera for sensing.

2.1. Deployment-Level Challenges

Due to limited resources with respect to most of the sensor nodes and the high computational complexity of active vision systems, it becomes challenging to allocate tasks to the nodes. Additionally, optimization of the resources by each sensor node is equally important. Further, designing a sensor architecture and accurate sensor placement according to the overall objective of the system is another challenge at the deployment level. Furthermore, calibration of the sensors of the SCN and occlusion handling by the system also add to the challenges at the deployment level.

2.1.1. Sensor Placement

Generally, cameras in SCNs are placed with overlapping FOVs to reproduce the entire operating environment. However, in some cases, with limited resources and larger operational areas, it becomes very difficult to place cameras with overlapping FOVs. The placement of cameras has a direct impact on the quality and quantity of data available for processing. For example, if an object is captured in the center of a camera’s FOV, the quality of the data, and thus the visual information available from the data, is much higher than if the object is captured at the edges of the camera’s FOV. In addition, camera placement must ensure maximum coverage of events and thus camera placement becomes a critical part of the deployment of SCNs. For the above reasons, sensor placement is critical to the deployment of an SCN.

2.1.2. Calibration

The configuration space of an SCN includes the internal and external camera parameters of the sensor nodes or cameras of the SCN. It is critical to change the configuration space of each camera, considering the available and used resources. When the target changes its position in the field of view (FOV) of a sensor or moves from the FOV of one sensor to another, the configuration of the sensor must be changed by modifying its parameters so that data can be acquired to yield optimal information (i.e., by keeping the object as close to the center of the FOV as possible).
Some SCN designs also emphasize minimizing resource utilization by inactivating sensors if no activity is detected over a specified period of time, which alters the dynamic topology of the network. Thus, dealing with such dynamically changing situations is very challenging. At the operational level, calibration challenges are further classified as follows.
  • Sensor modelling
The parameters of the configuration space of each sensor depend on the type of sensors used in the SCN. The sensor model provides information about the configuration space as well as the available resources, such as power, bandwidth, and the overall quality of service (QoS), possessed by the sensors. Therefore, sensor modelling plays a crucial role in active vision systems.
  • Localization
When an object moves away from the FOV of a sensor, it is important to hand over the object to another sensor in the SCN. Since the topology of the SCN is dynamic, the network must be configured to receive information about the active nodes and their relative positions at all times. This information is obtained through the active localization of sensor nodes in the SCN. Localization plays an important role in the system’s deployment, as it aids in determining the dynamic relative positions of active sensor nodes in the SCN.
  • Parameter estimation and correction
Real-time calibration of the active sensors participating in the SCN is required to capture the objects of interest (OOIs) with maximum information. Therefore, parameter estimation and correction play important roles in active vision systems employing SCNs.

2.1.3. Resource Optimization and Task-Load Balancing

Most of the present-day active vision systems require sensor mobility and thus rely on batteries and wireless communication. Due to limited resources, optimum resource utilization becomes critical. Resource optimization can be achieved in two ways, either by using low-power-consuming components at the nodes or by using the sensor nodes smartly. Low-power-consuming components are generally expensive and thus add to the overall expense of deploying an SCN. In a pre-established network, employing new low-power-consuming components may lead to having to discard the entire system and build a new one from the scratch. On the other hand, using the sensors smartly (only when required) does not necessarily require changing the existing hardware or components in the SCN. However, getting a machine to learn when to activate while responding in real time can be challenging for a system. An outcome of a dynamically changing state of the nodes in the SCN is dynamic task-load allocation to the active nodes in the SCN.
The network topology changes dynamically as the sensors switch from active to inactive states. Additionally, in order to avoid deviating from the SCN’s main goal (i.e., obtaining sensor data with high-quality visual information), the system’s overall functionality is dynamically distributed amongst the nodes in the form of task loads. As a result, the SCN must compute the dynamic topology in real time, determine the nodes’ localizations, and simultaneously distribute the task load among the active nodes in the SCN.

2.1.4. Handling Occlusions

The OOI can occasionally become obscured while being observed by a sensor, resulting in low-quality data due to the occlusion blocking the OOI. One way to deal with this situation is to hand over the OOI to the subsequent nearest sensor node. However, finding a next-best node for the OOI in real-time also creates challenges. Moreover, it becomes difficult if any nearby nodes are unavailable or if the OOI is outside the field of view (FOV) of any nearby nodes. There are a few prediction-based approaches that rely on computation; however, these approaches lack accuracy and may result in missing critical data, making them unreliable.

2.2. Challenges in Data Processing

Processing sensor data to obtain understanding is the most critical part of any active vision system. Active vision systems utilize data from multiple sources (i.e., multiple sensor nodes) and therefore require highly complex computational capabilities. The visual processing becomes more challenging when an application has to deliver results in real time. The data-processing challenges of an active vision system are discussed below.

2.2.1. Selection of a Processing Platform

The choice of processing platform is as important as the algorithms used to process the sensor data. The processing requirements of active vision are selected based on the system’s overall functionality, which is typically determined by two factors: processing time and functional complexity. For instance, the architecture of a system that requires complex computations in real time might be more complex and therefore more expensive than one with a relaxed processing window for applications with simpler computations.

2.2.2. Scene Reconstruction

The data obtained by a number of active sensors need to be synchronized to obtain useful information. An individual sensor’s data need to be stitched together in such a way that they result in the determination of an action performed by the OOI. The dynamically changing network topology makes this even more challenging.

2.2.3. Data-Processing Challenges

The active vision system relies on visual-processing algorithms to achieve the overall functionality of the system. Data-processing challenges include detecting an event or activity by the OOI and further deriving an understanding about the detected event or activity. Various visual-processing challenges in an active vision system are discussed below.
  • Object detection
While detecting an OOI, an active vision system divides the acquired sensor data into a foreground and a background. It should be noted that in an active vision system, object detection and data extraction by sensors are interdependent, making it difficult for any active vision system to accurately detect the OOI. In addition, one or more objects in the observed scene may be performing multiple tasks, increasing the likelihood of problems. Variations in viewpoints, shifts in lighting conditions, occlusions, and similar factors often lead to issues with object detection.
  • Object classification and tracking
Detected objects undergo classification to distinguish one object from another. Object classification requires the selection of appropriate methods for pattern recognition, clustering, and data segmentation. There are several methods for pattern recognition, clustering, and segmentation and there are associated advantages and challenges. Therefore, the selection of an appropriate method for object classification is very critical. Further, information about the activity performed by the object(s) participating in an event under observation is extracted by tracking the OOI(s) in consecutive frames. Deciding tracking features based on the activity of an OOI can be challenging because the OOI perceived by the sensor in successive frames is likely to have different viewpoints (i.e., the shape of the OOI perceived by the sensors may appear variable).
  • Object Re-identification
Many active-vision applications rely on relating activities of an OOI observed in non-consecutive frames. For instance, an understanding of an event can be obtained by comparing the actions of an OOI seen in various camera FOVs and at various times. For the system to reach a conclusion, it must re-identify an already identified OOI and correlate different activities of the OOI performed at different times.
  • Pose and behavior estimation
Consecutive frames are subjected to spatio-temporal evaluations of consecutive frames to obtain pattern and pose information from the OOI tracked in the scene. An accurate relationship between the activities and the changes in the pose of the OOI is required to relate the visual information to the corresponding understanding of the activity. Further, the active vision system must be configured to estimate variation in the pose of the OOI and thus derive the behavior of the OOI based on the pattern of the poses.

2.2.4. Activity Recognition and Understanding

Activity detection and recognition follows two paradigms: static and dynamic. An activity can be detected by analyzing a single frame, known as static recognition, and requires only spatial evaluation of the frame. Dynamic approaches require the evaluation of multiple consecutive frames, utilizing scene reconstruction (also known as dynamic recognition), and thus require spatio-temporal computational capabilities. Static recognition processes do not require pose estimation. Dynamic recognition, on the other hand, is generally used to solve complex problems where the information is obtained by pose estimation, such that the pose information is subsequently processed.
It is worth noting that the process flow of an active vision system utilizing an SCN includes two basic components: one derived from the functionality of the SCN (i.e., for the dynamic calibration of parameters) and the other derived from the computer vision system (i.e., for visual processing). In addition, it must also be noted that both the abovementioned components are inter-related co-operatively.

2.3. System-Level Challenges

In an active vision system employing an SCN to extract sensor data, the calibration of parameters of the sensor nodes participating in the SCN is highly dependent on the detection and interpretation of the sensor data by the active vision system. Further, the effectiveness and performance of the active vision system (i.e., the data processing) depend on the quality of data extracted by the SCN nodes. Thus, it can be inferred that the functionality of the SCN and the data processing system are interdependent. Moreover, real-time decision making is even more challenging for such an interdependent system.
In some respects, an active vision system can be considered a type of computer vision system wherein the system is configured to make alterations in the configuration space of the sensor(s) employed by the system in real time in a manner pre-defined by way of design protocols. However, an active vision system is not capable of handling most unforeseen challenges because it is protocol-driven and therefore cannot make independent decisions. Thus, the system-level challenges mentioned above are very critical and hard to overcome. If the system is configured with self-reconfigurable properties, it can overcome the abovementioned challenges. If such a system is configured with adaptive capabilities, this may further facilitate an enhancement in performance through active learning. The self-reconfiguration and adaptive properties of an active vision system are discussed later in Section 5.

3. Existing Solutions

To understand the development and progress at each operational level of active vision systems, we have performed an extended search survey for the last two decades. The results were generated with the keywords “active vision”, “computer vision”, “smart camera network”, AND “the name of challenge”, as discussed in Section 2, using the Google Scholar search tool.

3.1. Sensor Placement

The two major challenges to be addressed while deploying an SCN in reference to sensor placement are the maximization of the surveillance area and the handling of non-overlapping FOVs (i.e., managing handover). Indu et al. [10] and Zhang et al. [11] proposed methods for sensor placement in networks aiming to maximize the surveillance areas covered by the sensors. Silva et al. [12] proposed a system for coordination among the sensor nodes of an unmanned aerial vehicle (UAV) network for efficient surveillance. The systems and methods proposed in [10,11,12] present camera-placement solutions for optimized functionality; however, they lack flexibility in architecture and prioritization of surveillance space. An activity-based prioritization of the surveillance area is proposed in [13] by Jamshed et al., addressing the abovementioned concern. Vejdanparast [14] addressed the camera-placement problem for maximizing the area of surveillance by enhancing the fidelity of each camera in the network. Wang et al. [15] proposed Latin-Hypercube-based Resampling Particle Swarm Optimization (LH-RPSO) based on a camera placement algorithm for IoT devices and networks.
Redding et al. [16] proposed object handover based on multiple features, such as Zernike moments, scale-invariant feature transformation, gray-level co-occurrence matrices, color models, etc., using cross-matching for non-overlapping FOVs. In [17], Esterele et al. presented a method for the generation of an online real-time vision graph for the handover of information in a decentralized network with non-overlapping FOVs. The method proposed in [17] demonstrated that no prior knowledge of nodes is required and that it is easy to add or remove nodes from the network. Lin et al. [18] proposed an active handover control for real-time handover of single objects using multiple PTZ cameras, using the shortest distance rule and spatial relations. The method proposed in [18] proposed the readiness of a receiving camera before handover. A year-wise representation of the abovementioned sensor placement techniques along with their advantages is shown in Table 1.

3.2. Calibration

The major challenges in the calibration of nodes participating in an SCN concern sensor (camera) modelling, localization, and parameter estimation and correction. Some of the proposed solutions addressing various calibration challenges are discussed hereinbelow.

3.2.1. Camera Modelling

Some basic models for camera calibration are the thin-lens camera model, the pinhole camera model (linear-perspective projection model), the orthographic projection model, the scaled orthographic projection model, and the para-perspective projection model.
The thin-lens model is a linear calibration model that accounts for the effects of translation and rotation relative to a view plane. The pinhole model later introduced the effect of linear perspective projection; however, it has high computational complexity.
To overcome the high computational complexity of the pinhole model, Hall et al. [19] proposed a much simpler and computationally efficient linear model based on 3D affine transformation with linear perspective projection. The abovementioned linear models did not perform well, as they were unable to account for non-linear distortion, which was addressed by improving 3D affine transformation with non-linear perspective projection models by Tsai et al. in [20], Toscani’s non-linear calibration model in [21], and by Wang et al. in [22].

3.2.2. Localization

Camera localization helps estimate the relative positions, orientations, and poses of active nodes in a network. Identifiers or markers in the form of lines, points, features, cones, circles, spheres, etc., are commonly used for the localization of nodes in a camera network.
Such identifiers are commonly used in unknown network environments but can also be used in known environment for improved accuracy and better scene mapping. Utilizing perspective points as markers, the Perspective-n-Point (PnP) algorithm can be used in a known network environment. Simultaneous localization and mapping (SLAM), as presented in [23,24], and structure from motion (SFM), as presented in [25], can be used for dynamically changing network environments. The SFM technique in [25] is based on human vision perseverance to estimate a 3D scene using 2D image data by combining image motion information with frame data. The Monte Carlo method in [26] uses a particle filter for localization and recursive Bayesian estimation for sorting and sampling.
Montzel et al. [27] proposed a distributed energy-efficient camera network localization method using sparse overlapping in 2004. In [28], Brachmann and Rother proposed 6D pose estimation using an end-to-end localization pipeline. The geometric localization obtained via the head-to-foot location (poles) of pedestrians using an estimated distribution algorithm (EDA) in [29] can be utilized for self-calibration of the nodes in a network. A year-wise representation of the abovementioned localization techniques along with their advantages is shown in Table 2.

3.2.3. Parameter Estimation and Correction

Zheng et al. [30] proposed a focal-length estimation method using parallel particle swarm optimization (PSO) with low time complexity and efficient performance. Führ and Jung [31] proposed a self-calibration method for the surveillance of pedestrians in a static camera network using a projection matrix obtained from non-linear optimization of an initial projection matrix obtained after pole extraction. The information in the projection matrix was used for the localization of cameras in the network. In [32], Yao et al. proposed a self-calibration model for dynamic multi-view cameras using golf and soccer datasets based on a field model.
Li et al. [33] proposed a greedy-descent-optimization-based parameter-estimation and scene-reconstruction framework for camera–projector pairs for self-calibration. A network of such systems can be used for efficient tele-immersion applications. Janne and Heikkilä [34] proposed a self-reconfiguration solution for a camera network with focal-length estimation using homography from unknown planar scenes. In [35], Tang et al. proposed a simultaneous distortion-correction self-configuration method using an evolutionary optimization scheme on an estimated distribution algorithm (EDA) for tracking and segmentation. A year-wise representation of the abovementioned parameter estimation techniques along with their advantages is shown in Table 3.

3.3. Resource Optimization: Topology Estimation and Task-Load Balancing

As the network topology changes, the task load needs to be altered to ensure that the overall functionality of the system is achieved. Marinakis and Dudek [36] proposed a system to estimate the topology of a visual network in the form of a weighted directed graph using statistical Monte Carlo expectation and sampling models. Hangel et al. in [37] addressed the problem of topology estimation for a large camera network and proposed a window-occupancy-based method as a solution. The method in [37] required a lot of assumptions and could not handle large numbers of data. Detmold et al. [38] proposed a topology-estimation method capable of handling data from a large number of nodes in the network by scalable collective stream processing using an exclusion algorithm in distributed clusters of nodes. The method proposed by Detmold et al. [38] is similar to the decentralized processing scheme. Clarot et al. [39] proposed an activity-matching-based network topology for distributed networks. Zhou et al. [40] proposed topology estimation by means of a statistical approach in a distributed network environment, utilizing identity and appearance similarity.
In [41], Farrel and Davis proposed network topology estimation for decentralized data processing. Zhu et al. [42] proposed a centralized processing approach for topology discovery using pipeline processing of lightning variations. In [43], Goutam and Misra proposed a trust-based topology management system for a distributed camera network. Tan et al. [44] proposed a method for topology estimation using blind distance as a parameter. In [45], Li et al. proposed topology estimation using Gaussian and mean cross-correlation functions for a distributed camera network. A year-wise representation of the abovementioned topology-estimation techniques along with their advantages is shown in Table 4.
An efficient computer vision system differentiates the overall functionality into a number of small tasks to optimize the system’s functionality. The task load for each active node depends on its local state, orientation, and available resources. In some respects, resource utilization is related to task-load balancing at each active node in the SCN. Kansal et al. [46] presented a distributed approach for adaptive task-load assignment on the basis of available energy from the network environment, which significantly improved the lifetime of the system. Rinner et al. [47] proposed a heterogeneous multiple mobile-agent-based task-allocation framework utilizing a distributed multi-view camera network. Later, Rinner et al. [48] presented an updated approach to allocate tasks for traffic surveillance, proposing clustered surveillance areas. In [49], Karuppiah et al. proposed a hierarchy-based automatic resource allotment and task-load balancing algorithm using fault tolerance based on activity density for a distributed network. Dieber et al. [50] proposed expectation-maximization-based task-load assignment to optimize monitoring performance, with efficient resource utilization. Dieber et al. [51] extended their work on task-load balancing with market-based handover for real-time tracking with optimized resource utilization. In [52], Christos et al. proposed a market-based bidding framework for multi-task allocation for a distributed camera configuration. A year-wise representation of the abovementioned task-load balancing techniques along with their advantages is shown in Table 5.

3.4. Occlusion Handling

Occlusion-handling approaches either aim at handing over the OOI to the next-best sensor node or predicting the occluded part of the OOI and reproducing it virtually to obtain the missing visual information. Occluded objects in the camera field can result in loss of activity information and thus compromise the functionality of the control system utilizing the understanding provided by the computer vision. Wang et al. [53] proposed occlusion estimation at each point of a scene flow field with patch-match optimization utilizing feature consistency and smoothness regularization as performance parameters in space with an improved red–green–blue dense model. In [54], Quyang et al. proposed a framework based on a part-based deep model for pedestrian detection. The proposed model is capable of estimating information loss due to occlusion in the form of errors in detector scores, using visibility of parts as a parameter.
Shahzad et al. [55] proposed multi-object tracking with effective occlusion handling by modelling the foreground using a K-means algorithm, where the object information is associated after occlusion using a statistical approach. Rehman et al. [56] proposed clustering based on a variational Bayesian method and multi-object tracking based on concepts of attractive and repulsive forces depending upon Euclidean distances between objects, utilizing a social force model to avoid the effects of occlusion. In [57], Chang et al. proposed a convolutional-neural-network (CNN)-based tracking system with sparse coding for the pre-training of a network with the capability of handling occlusion effectively for the surveillance and classification of vehicles. Zhao et al. [58] proposed an adaptive background formulation based on a Gaussian model for occlusion handling and object tracking in a coarse-to-fine-manner without-affecting-appearance model present in the system. In [59], Liu et al. proposed a distraction-aware tracking system based on a 3D mean-shift algorithm, capable of altering its appearance model and occlusion handling by utilizing depth information of the OOI. A year-wise representation of the abovementioned occlusion handling techniques along with their advantages is shown in Table 6.

3.5. Selection of a Processing Platform

Selection of a platform to develop a computer vision system is as critical as designing or selecting algorithms specific to the functionality. Most commonly, the processing platforms may either be software-based, as with a central processing unit (CPU) or a graphical processing unit (GPU), or hardware-based, as with a field-programmable gate array (FPGA) and application-specific integrated circuits (ASICs). The selection of the platform depends on the requirements of processing capabilities, result accuracy, flexibility, timeliness, and resource utilization. A comparative evaluation of the selection of processing platforms for computer vision systems is presented by Feng et al. [60].
Systems that require flexible functionality usually prefer CPU- or GPU-based processing platforms; however, the efficiency of such systems is low. On the other hand, ASICs and FPGA are used for systems that require high efficiency and better and faster computations; however, these systems lack operational flexibility.
Hørup et al. [61] presented a comparative analysis of general-purpose computations performed by CPUs and GPUs in computer vision systems. Guo et al. [62] proposed a fast and flexible CPU-based computation system for human pose estimation. Tan et al. [63] proposed a fast yet flexible deep-learning-based computer vision system utilizing a GPU. Irmak et al. [64], Costa et al. [65], and Carbajal et al. [66] proposed FPGA-based computer vision systems, whereas Xiong et al. [67] presented an ASIC-based computer vision system with enhanced operational flexibility. To obtain the advantages of the two kinds of platform, a hybrid model, i.e., a platform with a hardware–software combination such as the one presented in [68], can also be utilized for computation.
The state-of-the-art research on the selection of processing platforms aims to serve the ends of efficiency and flexibility of computation. CPU- and GPU-based systems aim to improve efficiency and computational speed, whereas systems based on FPGA and ASICs are intended to enhance operational flexibility.

3.6. Scene Reconstruction

The sensors in the SCN obtain raw sensor data from their respective FOVs. The raw data are then fused together, utilizing spatio-temporal information (obtained through the frame count and relative localization of sensors) associated with the data. The integration of sensor data into a virtual environment or a scene is called scene reconstruction. R. Szeliski [69] proposed a novel volumetric scene-reconstruction method using a layered structure and multiple depth maps. Martinec et al. [70] proposed 3D reconstruction using an uncalibrated image dataset and a pipelining approach to detect regions of interest (ROIs) and match them using random sample consensus (RANSAC). Peng et al. [71] addressed the network geometry-estimation problem utilized for scene reconstruction and proposed two-view geometry estimation using a local-structure-constraint-based L2-estimation–local-structure-constraint (L2E-LSC) algorithm.
For efficient scene reconstruction, effective point matching is imperative. Brito et al. [72] compared different state-of-the-art point correspondence methods, such as Scale Invariant Feature Transform (SIFT), Fast Retina Keypoints (FREAK), Oriented Fast and Rotated Brief (ORB), Binary Robust Invariant Scalable Keypoints (BRISK), and Speeded-Up Robust Features (SURF). Milani [73] introduced localization-based reconstruction for a heterogeneous network. Aliakbarpour et al. [74] reviewed different scene-reconstruction methodologies and presented a reconstruction method using parametric homography. Wang and Guo [75] presented a reconstruction method using plane primitives of an RGB-D frame. Ma et al. [76] proposed a mesh-reconstruction system using an adaptive octree division algorithm for point-cloud segmentation and mesh relabeling and reconstruction for scene reconstruction. Ichimru et al. [77] presented 3D scene reconstruction using a CNN under water, utilizing transfer learning with a bubble dataset to avoid distortions.

3.7. Data Processing

The overall objective of a computer vision system is to detect an activity or event of interest, perform processing on the data containing information about the event, develop an understanding of the event, and generate an action by way of a control unit based on the understanding. The visual processing includes a number of stages. Recent advancements in the most common stages of visual processing in computer vision systems are discussed hereinbelow.

3.7.1. Object Detection

The first step towards the visual processing of data in a computer vision system is detection of the OOI (also referred to as the foreground). Some traditional object-detection methods are the Viola and Jones technique [78], scale-invariant feature transformation (SIFT) [79], HOG-based detection [80], optical flow [81,82], and background subtraction [83]. Most of the recent computer vision systems involve machine-learning-based object detection, including neural-network-based object detection [84], “you only look once” (YOLO) [85], region proposals (R-CNN) [86], single-shot refinement neural networks (RefineDet) [87], Retina-Net [88], and single-shot multi-box detectors (SSDs) [89]. The recent neural-network-based object-detection methods provide much better accuracy as compared to the traditional detection methods, but they are highly dependent on the training data. A survey of the evolution of detection techniques in computer vision from probabilistic-prediction approaches to advanced machine-learning approaches is presented in [2].
Some challenges in object detection arise due to dynamic illumination changes, the movement of objects, and occlusions. Roy and Ghosh [90] proposed an adaptive background model (a histogram min–max bucket) using a single sliding window to add adaptability to the background detection. The adaptive background model used a median-finding algorithm incorporated to handle dynamic illumination changes. Bharti et al. [5] proposed an adaptive real-time occlusion-handling kernelized correlation framework for UAVs capable of updating location and boundary information based on the confidence values of the tracker. Min et al. [91] proposed a multiple-object-detection approach using pixel lifespan to blend ghost shadows to the background and a classifier based on a state vector machine (SVM) and a convolutional neural network (CNN) to avoid occlusions.
Detecting an object in a moving FOV is a tedious task and requires substantial approximations. Wu et al. [92] proposed an effective computational model to solve this problem. They evaluated a coarse foreground using singular-value decomposition and reconstructed the background using the foreground information obtained through a fast in-painting technique. Mean-shift segmentation was used for further refinement of the foreground. Hu et al. [93] presented a tensor-based approach to detect mobile objects without changing the scene dynamics. For initial foreground detection, saliently fused sparse regularization was used and tensor nuclear norms were utilized to handle background redundancy. The foreground was further improved using a 3D locally adaptive regression kernel, which was used to compute spatio-temporal variations. A year-wise representation of the abovementioned object detection techniques and their advantages is shown in Table 7.

3.7.2. Object Classification and Tracking

The OOI can be classified after detection on the basis of one or more appearance parameters [94] called features that classify the object based on color, texture, shape, pixel motion, etc. Some basic features for object representation presented in [95] are points, shapes, and silhouettes or contours. Conventionally, object-classification methods [96] can be categorized as decision-based, statistical-probability-based, and soft-computing-based techniques. Some common decision-based classification methods are decision trees [97,98] and random forests [99]. Bayesian classification [100,101,102], discriminant analysis [103], logical regression [104], and nearest-neighbor [105] approaches use statistical probability for classification. State vector machines [106], multi-layered perceptrons [107], and neural networks [108,109] use soft computing for classification.
One of the major challenges in object tracking is distortion of the OOI. Villiers et al. [110] proposed real-time inverse distortion for distortion correction. A lot of methods for distortion correction and calibration use properties of vanishing points, as was first proposed in [111]. A distributed algorithm proposed by Caprile et al. [35] illustrated the utilization of tracking waking humans as poles to derive vanishing points for radial distortion correction and self-calibration. Radial distortion correction was addressed in [112,113] by estimating the center of distortion. Huang et al. [114] proposed linear-transformation-based radial distortion correction, whereas Zhao et al. [115] used a pipelined process for radial distortion correction. Methods for the correction of radial as well as tangential distortion were proposed in [116,117,118]. Yang et al. [119] proposed estimation and correction of perspective distortion utilizing depth information. Color-calibration theory, discussed by Finlayson et al. in [120], has been used to address the challenge of optical distortion. In [121], Wong et al. presented a color-calibration approach using a multi-spectral camera model.
Another challenge in object tracking is to obtain the best possible information for each object of interest in the scene while dealing with multiple OOIs in the scene. One of the factors affecting the performance of this task is the motion of the camera nodes to capture the objects precisely. The motion blur sometimes increases to such an extent that the effect of moving the camera in the environment is nullified. Han et al. [122] proposed a motion-aware tracker to address the abovementioned problem by filling in the tracking fragments caused by occlusion or blur. Meinhardit et al. [123] proposed a former tracking system to address the challenges of multi-object tracking.

3.7.3. Object Re-Identification

Object re-identification encounters multiple challenges through occlusions, false object detection due to ghost shadows [124], illumination changes, and change in viewpoints. Zhang et al. [125] proposed an adaptive re-identification framework for spatio-temporal alignment utilizing Fisher vector learning to address illumination changes in the re-identification of OOIs. Yang et al. [126] used logical determinant metric learning to tackle re-identification through different camera views to overcome occlusions in the re-identification of objects. Multiple features, when fused together, improve re-identification capabilities; however, the role of each feature and the weights are critical. Geng et al. [127] proposed a feature-fusion method based on weighted-center graph theory to obtain the role of each feature in re-identification. Yang et al. [128] presented re-identification using partial information which can be utilized with occluded data.

3.7.4. Pose and Behavior Estimation

Pose-estimation methods utilize models to associate the patterns of postures, poses, or shapes of OOIs detected and track them to obtain meaningful information, thus yielding understanding. The major challenges in pose and behavior estimation are: the selection of a pose-estimation model and the association of information with a sequence of poses obtained through tracking an OOI. Pose-estimation methods can either be model-based and may use kinematic modelling, planar modelling, or volumetric modelling, or they can be model-free. Kinematic models rely on tracking the movements of points on the OOI, planar models rely on contours, whereas volumetric models rely on changes in the volume distributions of OOIs tracked over time. Kinematic models are easy to process but are not reliable. Planar and volumetric models are more reliable but have more computational complexity. Addressing the challenge of low computational complexity and high accuracy in pose and behavior estimation, Chen et al. [129] proposed an anatomically aware 3D pose-estimation model for human behavior analysis. Staraka et al. [130] proposed a kinematic skeletal-model-based pose-estimation method with real-time and accurate behavior estimation.

3.8. Visual Understanding

Systems achieve visual understanding by relating one or more pose behaviors and performing spatio-temporal analyses of the patterns of poses corresponding to events in a scene. Campbell et al. [131] utilized phase-space constraints to depict human motion. Oren et al. [132] used single-frame wavelet templates for pedestrian detection. Image captioning [133], manuscript reviewing in the medical field [134] and academia [135] are some applications of static activity recognition. Nguyen et al. [136] used multi-objective optimization for real-time activity monitoring. In [3,137], dynamic recognition was utilized for sports analysis. Xiang et al. [138] proposed a system capable of making decisions for multiple-object tracking, whereas Wu et al. [4] proposed a dynamic activity recognition system for smart homes. In [139], Laptev et al. proposed a state-vector-machine-based abnormal human activity recognition system.

4. Contemporary Solutions

From the above, it can be observed that there has been a shift from model-based approaches to artificial-intelligence (AI)-based methods. We have also considered some of the earlier publications to understand the concepts of some of the recently proposed models. Further, it has also been observed that the emerging research fields in active vision systems are multi-object detection and tracking, occlusion handling, and sensor reconfiguration.
Artificial intelligence (AI) drives down the time taken to perform a task. It enables multi-tasking and eases workloads for existing resources. AI further facilitates decision making by making the process faster and smarter. For these reasons, most of the state-of-the-art systems addressing the abovementioned challenges rely on one or more artificial-intelligence (AI)-based approaches [140,141,142,143,144,145,146,147,148,149,150,151,152,153,154,155,156]. According to market research reported in [157], AI in computer vision has a 45% compound annual growth rate, which is the growth rate of research in the area. The main reason to switch from traditional model-driven approaches to artificial-intelligence-based systems is the high level of accuracy AI systems can provide as compared to the former. It has also been observed that most of the AI-based approaches are dependent on machine-learning (ML)- or deep-learning (DL)-based models.
Some of the recently proposed systems utilize the concepts of traditional model-based methods along with modern AI-based approaches to generate highly accurate hybrid systems [147,151], while some systems [153] capable of addressing the challenges through AI-based models utilize the basic principles of traditional model-based methods. Some of the state-of-the-art AI-based systems addressing the challenges of self-reconfiguration faced by active vision systems are presented in Table 8.
ML-based systems highly rely on training datasets to develop operational models. To make an ML-based system adaptive, AI models need to learn the unforeseen by deriving information from experience. Such systems fail drastically in centralized networks, as the nodes are trained on different datasets based on their specific experiences of distinct events in their surroundings.
Thus, the biggest challenge for adaptive systems is the sharing of information about events between the nodes in networks such that each event in the network of networks can be handled with equal accuracy by each node.
ML systems further suffer from visual attacks [158], the most common of which are adversarial attacks [159,160] which misguide the classifier and thus impact the accuracy of the system. One study [161] showcases the effects of adversarial attacks on the performance of machine-learning-based approaches. The systems are attacked by visual attacks which make small alterations in the weights of the classifier, thus misleading the classifier and reducing the accuracy of classification. Over a period of time, such attacks result in drastic degradation in the performance of the system. Refs. [162,163] present detailed surveys on the types of visual attacks and the methods proposed to detect and mitigate such attacks. Visual attacks [158] can be targeted if the effect of the attacks is predicted correctly by the model; however, in most scenarios, such attacks introduce random noise and thus are very difficult to reverse, causing permanent degradation in the classifier’s performance over time. To overcome the problems of data loss and degradation of system performance due to such attacks, ML systems must be incorporated into a distributed network, such that the data are distributed throughout the entire network. Such a network is capable of detecting an attack in the initial stage itself and providing back-up for any kind of data loss to each participating node in the network.

5. Self-Adaptation and Self-Reconfiguration

The challenges discussed above have received solutions designed to tackle particular problems in well-known settings. However, none of the aforementioned approaches addresses the challenge with unforeseen conditions optimally due to unexpected changes in the environment. To overcome this limitation, an active vision system must be capable of understanding the changes in its environment and reconfiguring the parameters of the active sensors participating in the system on its own. Further, active vision systems must also enable the sharing of information with each other to handle unforeseen changes in a better way.
An active vision system capable of self-adjusting its configuration space is called a “self-reconfigurable” system. The ability of an SCN to adapt to such changes and reconfigure its parameters for optimized performance in dynamic and unforeseen conditions is called self-reconfiguration of the SCN [164]. Such a self-reconfigurable SCN for UAVs used for surveillance was proposed by Leong et al. in [165]. A comparative survey presented by Natarajan et al. [166] illustrates a number of self-reconfiguration models for computer vision systems with multiple active nodes participating in data extraction. Martinel et al. [167] proposed the distributed self-reconfiguration of an SCN to address the vehicle re-identification problem utilizing deep-learning models.
The capability of a system to share information to learn and adapt can be achieved through “self-adaptation”. Self-adaptation is the ability of a network (in this case, specifically an SCN) to enhance its performance by enabling the network to update the configuration space of nodes, active participants, protocols, and functional algorithms. Some systems configured for self-adaptation require learning from past experience or the sharing of performance statistics from different parts of the network and opt for the settings which worked best for similar events.
Self-adaptation can be achieved through self-expression and self-awareness capabilities in a system. A system having self-awareness (SA) possesses knowledge of its state; self-awareness gives the system the ability to share its state in the form of parameters and overall quality of service (QoS). Additionally, each node in the self-adaptive system possesses the ability to question its current state, investigate alternative configurations for better QoS, and change its state using active learning.
Rinner et al. [168] used a market-based approach for self-awareness and proposed six major steps for self-adaptation, namely, resource monitoring, object tracking, topology learning, object handover, strategy selection and objective formation. Lewis et al. [169] classified events as explicit and implicit and discussed the privacy, extent, and quality of self-adaptation. Lewis and Chandra [170] discussed formal models for self-adaptation and the application of self-adaptation in systems with artificial intelligence, conceptual systems, engineering, automotive systems, computing, etc. Wang et al. [171] discussed methods of self-adaptation with capabilities of online learning. In [172], Ali et al. proposed an auto-adaptive multi-stream architecture using multiple heterogeneous sensors with pipelined switches between processing states and ideal states to reduce power, using an FPGA implementation that demonstrated inter-frame adaptation capability with a relatively low overhead. Guettalfi et al. [173] proposed an architecture utilizing quality of service (QoS), resource estimation, a feedback mechanism, and state estimation for public and private self-awareness using actuators. Zhu et al. [174] and Lin et al. [175] proposed a self-adaptation-based person-re-identification system based on unsupervised learning. Wu et al. [176] and Rudolph et al. [177] proposed an adaptive self-reconfiguration framework for computer vision systems. Both frameworks proposed the sharing of information between the sensors in a network for its utilization. However, both adaptive self-reconfiguration frameworks were designed for centralized network configurations and thus possess limited scopes of learning.
An adaptive self-reconfigurable framework presented hereinbelow, when utilized for an SCN-enabled active vision system, can provide adaptive calibration of SCN sensor parameters in near real time. Further, in the data-processing part of the system, the framework can be utilized to adapt to the best specifications by learning from the experiences obtained from other sources and thus result in the minimization of re-iterative training for models that are specific to the development of scene understanding.

6. Adaptive Self-Reconfiguration Framework

Piciarelli et al. [8] proposed a dynamic reconfiguration framework for SCNs, as shown in Figure 3. The framework used a local state (f), resources (r), and QoS information (q) from a number of nodes to generate the overall state (F), overall resources (R), and overall quality (Q) of the system, utilizing an SCN. The dynamic reconfiguration framework showcased the functionality of a reconfigurator, which was configured to determine changes in parameters based on a resource model and the objectives of the system. The dynamic reconfiguration framework in [8] was designed for an independent SCN system working in a centralized environment and thus lacks adaptiveness to deal with unforeseen events. The framework needs to reconfigure itself from scratch if a new type of activity is discovered and thus is not suitable for real-time active-vision applications. Further, almost all the symmetric/asymmetric systems and frameworks discussed in Section 3 and Section 4 lack adaptiveness and thus are likely to fail miserably in unforeseen conditions. Recent reconfiguration systems [177,178,179] based on the reconfigurator model of [8] have tried to improve the accuracy of detection; however, due to centralized networks of operation, they have limited scopes for learning and are not suitable to tackle unforeseen conditions in real time.
As discussed earlier, the reconfiguration of sensor nodes is highly reliant on the scene understanding gained by the processing of sensor data and vice versa. Due to this interdependency, it is very challenging for a system to reconfigure its configuration space while dealing with unforeseen situations. With prior knowledge of an event, scene understanding can be greatly improved and thus the reconfiguration of the sensor network.
This article proposes an adaptive self-reconfiguration (ASR) framework to extend the scope of the framework proposed by Piciarelli et al. [8] to a number of networks, such that each network is configured to learn and reconfigure adaptively by utilizing the experiences and models of others to obtain the best visual understanding possible and thus perform optimal reconfiguration of the SCN. In spite of using a centralized reconfigurator, we propose a distributed network of systems comprising a number of datacenters or cloud servers to perform data computation and reconfiguration. For illustration, the architecture is inspired by the NEAR blockchain network, comprising a number of datacenters, such that each datacenter is capable of sharing datasets, performance parameters, and even trained models. Due to the symmetrical architecture of the distributed blockchain network, the proposed ASR system provides unbiased functionality for each smart camera network utilizing the network. This allows the model, datasets, and parameters to be used by any SCN to deal with an unforeseen situation if a similar condition has been encountered and dealt with by another SCN in the distributed network. Further, the ASR framework proposes the distribution of critical data throughout the blockchain network, providing data security, such that if an SCN is attacked by adversarial attack, the critical data can be retrieved. In some respects, the network of systems developed or deployed using the ASR framework utilizing a distributed network (blockchain) can be considered a self-adaptive system of active vision systems.

Model

An exemplary embodiment of the proposed ASR-framework-based architecture for deployment of the adaptive self-reconfiguration of active vision systems utilizing SCNs is shown in Figure 4. The framework consists of “m” number of active vision systems (AVSs) connected together in a distributed network. The architecture comprises “m” number smart camera networks coupled to each other by way of a distributed blockchain network “B”. Sensed data and a local configuration space are associated with each sensor node of each SCN, which results in input data from each sensor being processed for reconfiguration by one or more datacenters in the blockchain network. Input data, local states, resources, and quality of service information from each sensor “si” are cumulatively passed to a fusion block of the SCN (represented by the set {Ci}) to obtain an overall system configuration space for the SCN. Each SCN has an overall functionality, defined by a number of objectives. The sensed data, the system’s configuration space, and the SCN objectives are communicated as self-expression data of the SCN “Ei” to the blockchain network “B” for processing.
The blockchain network comprises “M” number of datacenters (cloud servers) that are configured to perform computations on the self-expression data (Ei) for each SCN. A datacenter is selected as per the objectives of the SCN, based on the resources available at the datacenter. The datacenter generates a model based on the objectives of the SCN and derives an understanding by way of the sensor’s data using unsupervised artificial intelligence. Each activity and/or event determined by a model (by the datacenter) is assigned a pattern vector for identification, such that the pattern vector is generated based on the objectives and the understanding of the event by the model. The model further derives a best-possible configuration space corresponding to the event iteratively. Each pattern vector along with the model specifications is distributed to every datacenter in the blockchain network “B”. Each event detected by any datacenter generates a pattern vector that is distributed throughout the entire blockchain “B”. Based on the understanding of the detected activity, the selected datacenter generates self-awareness data (Ai) comprising a revised configuration space set {Ci} for the SCN. Based on the revised configuration space set, each camera sensor (sij) is calibrated.
If an unforeseen condition is faced by another SCN in the network, the associated datacenter generates a pattern vector and compares the pattern vector with a pre-existing pattern vector in the blockchain network. Based on the comparison of the pattern vector with the pre-existing pattern vectors, a suitable datacenter is allotted and thus the latency of iterative processing is minimized. The improved configuration space is reverted back to the SCN through self-awareness data that are utilized by the reconfigurator to calibrate the parameters of each sensor of the SCN.
Notation used:
Xi: ith smart camera network;
sij: jth camera sensor of the ith active vision system;
{Ci}: Set of input data for the ith SCN;
{Ci}: Set of output data for the ith SCN;
Ei: Self-expression data for the ith SCN;
Ai: Self-awareness data for the ith SCN;
ni: Number of camera sensors in the ith active vision system;
m: Number of active vision systems utilizing the ASR framework;
B: Distributed blockchain network comprising “M” number of datacenters;
Dk: kth datacenter in the distributed blockchain network.
To generate the pattern vector corresponding to each event, we propose an auto-encoder-based unsupervised learning model, such that, for each event, the sensor data received by a selected datacenter is converted to lower-dimensional data. The auto-encoder model determines a reconstruction error to restore the sensor data back to their original form. The reconstruction error is propagated to the auto-encoder to iteratively alter the weights of the model for training in order to minimize the reconstruction error. The data of weights of the trained model along with the objectives of the SCN are utilized to generate the pattern vector, such that the pattern vector is distributed throughout the blockchain. The datacenter further computes the reconfiguration parameters based on the understanding of the event and sends the reconfiguration parameters to the SCN in the form of self-expression data, which are utilized by the reconfigurator of the SCN to calibrate the parameters. If an identical event occurs in any other SCN in the blockchain, the selected datacenter generates a pattern vector for the event. The pattern vector is matched with the pre-existing pattern vectors in the blockchain, and the reconfiguration model of the closest pattern vector is shared with the SCN.
For an unbiased and seamless flow of operations, the distributed blockchain network protocols were designed based on consensus mechanisms. The datacenters are categorized as processor datacenter nodes and validator datacenter nodes. The categorization of the datacenters is based on the proof-of-active-participation (POAP) consensus mechanism [180]. The processor datacenter nodes generate a model and a pattern vector based on the sensor data, whereas the validator nodes validate the model and the pattern vectors. The processor nodes further share details of their resource consumption (bandwidth and computational resources, etc.) to obtain rewards. The validator nodes validate or authenticate the details about resource consumption, models, and pattern vectors generated by the processor nodes. The validator nodes further compare the pattern vector generated by the processor node with the pre-existing pattern vectors to determine the best suitable reconfiguration configurations. To ensure the unbiased functionality of the validator and the processor nodes, each participating processor node and validator node is required to raise a stake based on a proof-of-stake consensus mechanism [181]. Upon validation of the model and the reconfiguration, the processor nodes and the validator nodes receive their stake as well as a reward from the corresponding SCN network. A flow diagram of the process is shown in Figure 5.

7. Results

For illustration of the functionality of the adaptive self-reconfiguration (ASR) framework, we utilized multiple surveillance video datasets and simulated results, comparing the centralized reconfiguration of [8] to our proposed distributed adaptive self-reconfiguration. Each frame of each video dataset was standardized to a resolution of 640 × 360 pixels. Activity maps were generated using the centralized approach presented in [8] and the proposed ASR, respectively, such that the activity maps updated with each frame added to an event. The performances of the systems were compared in terms of multi-object tracking accuracy (MOTA) results obtained for both activity maps. Further, due to the unavailability of resources necessary to develop a private blockchain network with datacenters, we utilized lower processing capabilities and thus, in spite of accurate processing latency, the results are represented with respect to training cycles (Ti) (each of 15 minutes duration); however, the processing capabilities of a datacenter are much higher and can be utilized to achieve results in near real time. We utilized standard regions with a convolutional neural network (R-CNN) model for multi-object detection by the centralized system of [8] as well as the distributed system based on the proposed ASR framework. All the results presented hereinbelow were derived using the MatLab Image Processing Toolbox. The activity map corresponding to the proposed ASR was further utilized to predict the upcoming event in the next frame. The SAR framework was utilized for multi-object detection by the exemplary embodiment method described in Section 6; however, it can be utilized to enhance any performance parameter, as illustrated in Section 2.
True-positive pixel count (TPC), false-positive pixel count (FPC), false-positive pixel-detection rate (FPR), true-negative pixel count (TNC), true-negative pixel-detection rate (TNR), and false-negative pixel count (FNC) were used as primary performance parameters to obtain MOTA values using Equation (1).
MOTA (%) = (Total count of pixels − falsely detected pixels) × 100

7.1. Surveillance Dataset 1

A comparison of the performance of the system presented in [8] and the proposed ASR system, tested on surveillance dataset 1 in terms of MOTA, is presented in Table 9 and Figure 6. The predictions of the directions of vehicle movement in various random frames from surveillance dataset 1 based on the generated activity map are presented in Figure 7.

7.2. Surveillance Dataset 2

A comparison of the performances of the system in [8] and the proposed ASR, tested on surveillance dataset 2 in terms of MOTA, is presented in Table 10 and Figure 8. The predictions of the directions of vehicle movement in various random frames from surveillance dataset 2 based on the generated activity map are presented in Figure 9.
A comparison of the performances of the system presented in [8] and the proposed ASR for multi-object tracking applications are represented in Figure 6 and Figure 8. It must be noted that, as the reconfiguration is dependent on tracking accuracy, the reconfiguration presented by the distributed ASR model is better than that of the centralized reconfigurator model of [8].

8. Conclusions and Scope

The performance of active vision systems and reconfiguration of the sensors providing data to active vision systems are interdependent. However, the reconfiguration of the calibration space of an active vision system employing a network of sensors can be challenging. Most of the state-of-the-art active vision systems fail miserably in dealing with unforeseen conditions, as the reconfiguration model takes time to adapt to the new conditions to develop an understanding of an unforeseen event. Thus, reconfiguring such a system in real time is nearly impossible. Further, most active vision systems nowadays rely on artificial-intelligence-based models for the processing of sensor data to develop an understanding of an event. However, such models are prone to adversarial attacks and are thus threatened with data loss. Therefore, such systems cannot be relied on for making critical decisions in real time.
This article discusses the challenges at different operational levels in deploying an active vision system employing a camera network. This article presents a detailed description of the systems and methods proposed for addressing the challenges with respect to both data processing and reconfiguration, along with the state-of-the-art solutions for the same, proposing an adaptive self-reconfiguration (ASR) framework employing a blockchain-based distributed network for data processing and reconfiguration of the sensor network’s configuration space. To make the understanding of the framework easier, this article has briefly defined the concepts of self-adaptation and self-reconfiguration in systems prior to the ASR framework.
The blockchain network of the ASR-framework-based architecture acts as a system of systems, connecting a number of smart camera networks together in a distributed architecture. The blockchain further includes a number of datacenters configured to obtain data from the SCNs and perform data processing to obtain understanding about the events and activities in the scene. The datacenters further generate pattern vectors corresponding to the event or activity detected and distribute pattern vectors along with corresponding reconfiguration models to each datacenter in the distributed blockchain network, such that if a similar event is observed at any other SCN in the network, the reconfiguration model associated with a pre-existing pattern can be utilized for a much faster reconfiguration of the SCN. The blockchain network is founded on proof-of-stake and proof-of-active-participation consensus mechanisms to maintain an unbiased and smooth flow of operations between all the participating datacenters in the network. Further, due to the distributed architecture, critical data float in the blockchain network and thus the effect of adversarial attacks is minimized.
This article further compared the performance of a centralized active vision system with a distributed system based on the proposed ASR framework for multi-object tracking and showcased enhanced tracking performance in terms of multi-object tracking accuracy and low latency. The proposed framework was tested on a homogeneous environment, with some limitations and assumptions; however, the ASR framework aims to be developed for heterogeneous systems to enhance the scope of its applications in future.

Author Contributions

S.: Lead author of the manuscript (corresponding author), conceptualization and methodology, writing—original draft preparation, investigation, and editing; I.S.: Second author, research design, guidance, and reviewing. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Data Availability Statement

Training of both models (i.e., for [8] and the proposed ASR) was performed on a standard surveillance video dataset. The models were further tested on surveillance dataset 1 and surveillance dataset 2. The standard surveillance dataset, as well as surveillance dataset 1 and surveillance dataset 2, are non-restricted and free-to-use third-party datasets with 640 × 360-pixel resolutions. The datasets and the training and testing models can be accessed by clicking on this link: https://drive.google.com/drive/folders/1ZizuTfDZ2gT-ro1o2lAAo_SVsw4_MY-7 (accessed on 23 October 2022).

Acknowledgments

This work was carried out under the supervision of Indu Sreedevi in the Department of ECE, Delhi Technological University, New Delhi, India, and Shashank expresses immense gratitude to his guide and UGC for enlightening him throughout the process.

Conflicts of Interest

The authors declare that they have no conflicts of interest. The authors declare that they have no known competing financial interests or personal relationships that could have appeared to influence the work reported in this article.

References

  1. Reisslein, M.; Rinner, B.; Roy-Chowdhury, A. Smart Camera Networks. Computer 2014, 47, 23–25. [Google Scholar]
  2. Zhang, T.; Aftab, W.; Mihaylova, L.; Langran-Wheeler, C.; Rigby, S.; Fletcher, D.; Maddock, S.; Bosworth, G. Recent Advances in Video Analytics for Rail Network Surveillance for Security, Trespass and Suicide Prevention—A Survey. Sensors 2022, 22, 4324. [Google Scholar] [CrossRef] [PubMed]
  3. Theagarajan, R.; Pala, F.; Zhang, X.; Bhanu, B. Soccer: Who has the ball? Generating visual analytics and player statistics. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops, Salt Lake City, UT, USA, 18–23 June 2018; pp. 1749–1757. [Google Scholar]
  4. Wu, C.; Khalili, A.H.; Aghajan, H. Multiview activity recognition in smart homes with spatio-temporal features. In Proceedings of the Fourth ACM/IEEE International Conference on Distributed Smart Cameras, New York, NY, USA, 31 August–4 September 2010; pp. 142–149. [Google Scholar]
  5. Bharati, S.P.; Wu, Y.; Sui, Y.; Padgett, C.; Wang, G. Real-time obstacle detection and tracking for sense-and-avoid mechanism in UAVs. IEEE Trans. Intell. Veh. 2018, 3, 185–197. [Google Scholar] [CrossRef]
  6. Agarwal, M.; Parashar, P.; Mathur, A.; Utkarsh, K.; Sinha, A. Suspicious Activity Detection in Surveillance Applications Using Slow-Fast Convolutional Neural Network. In Advances in Data Computing, Communication and Security; Springer: Berlin/Heidelberg, Germany, 2022; pp. 647–658. [Google Scholar]
  7. Hanson, A. (Ed.) Computer Vision Systems; Elsevier: Amsterdam, The Netherlands, 1978. [Google Scholar]
  8. Piciarelli, C.; Esterle, L.; Khan, A.; Rinner, B.; Foresti, G.L. Dynamic reconfiguration in camera networks: A short survey. IEEE Trans. Circuits Syst. Video Technol. 2015, 26, 965–977. [Google Scholar] [CrossRef]
  9. Jesus, T.C.; Costa, D.G.; Portugal, P.; Vasques, F. A Survey on Monitoring Quality Assessment for Wireless Visual Sensor Networks. Future Internet 2022, 14, 213. [Google Scholar] [CrossRef]
  10. Indu, S.; Chaudhury, S.; Mittal, N.R.; Bhattacharyya, A. Optimal sensor placement for surveillance of large spaces. In Proceedings of the 3rd ACM/IEEE International Conference on Distributed Smart Cameras (ICDSC), Como, Italy, 30 August 2009–2 September 2009; pp. 1–8. [Google Scholar]
  11. Zhang, G.; Dong, B.; Zheng, J. Visual Sensor Placement and Orientation Optimization for Surveillance Systems. In Proceedings of the 10th International Conference on Broadband and Wireless Computing, Communication and Applications (BWCCA), Krakow, Poland, 4–6 November 2015; pp. 1–5. [Google Scholar]
  12. Da Silva, L.C.; Bernardo, R.M.; De Oliveira, H.A.; Rosa, P.F. Multi-UAV agent-based coordination for persistent surveillance with dynamic priorities. In Proceedings of the International Conference on Military Technologies (ICMT), Brno, Czech Republic, 31 May–2 June 2017; pp. 765–771. [Google Scholar]
  13. Jamshed, M.A.; Khan, M.F.; Rafique, K.; Khan, M.I.; Faheem, K.; Shah, S.M.; Rahim, A. An energy efficient priority based wireless multimedia sensor node dynamic scheduler. In Proceedings of the 12th International Conference on High-capacity Optical Networks and Enabling/Emerging Technologies (HONET), Islamabad, Pakistan, 21–23 December 2015; pp. 1–4. [Google Scholar]
  14. Vejdanparast, A. Improving the Fidelity of Abstract Camera Network Simulations. Ph.D. Thesis, Aston University, Birmingham, UK, 2020. [Google Scholar]
  15. Wang, X.; Zhang, H.; Gu, H. Solving Optimal Camera Placement Problems in IoT Using LH-RPSO. IEEE Access 2020, 8, 40881–40891. [Google Scholar] [CrossRef]
  16. Redding, N.J.; Ohmer, J.F.; Kelly, J.; Cooke, T. Cross-matching via feature matching for camera handover with non-overlapping fields of view. In Proceedings of the 2008 Digital Image Computing: Techniques and Applications, Canberra, ACT, Australia, 1–3 December 2008; pp. 343–350. [Google Scholar]
  17. Esterle, L.; Lewis, P.R.; Bogdanski, M.; Rinner, B.; Yao, X. A socio-economic approach to online vision graph generation and handover in distributed smart camera networks. In Proceedings of the 5th ACM/IEEE International Conference on Distributed Smart Cameras, Ghent, Belgium, 22–25 August 2011; pp. 1–6. [Google Scholar]
  18. Lin, J.L.; Hwang, K.S.; Huang, C.Y. Active and Seamless Handover Control of Multi-Camera Systems With 1-DoF Platforms. IEEE Syst. J. 2012, 8, 769–777. [Google Scholar]
  19. Hall, E.L.; Tio, J.B.; McPherson, C.A.; Sadjadi, F.A. Measuring curved surfaces for robot vision. Computer 1982, 5, 42–54. [Google Scholar] [CrossRef]
  20. Tsai, R. A versatile camera calibration technique for high-accuracy 3D machine vision metrology using off-the-shelf TV cameras and lenses. IEEE J. Robot. Autom. 1987, 3, 323–344. [Google Scholar] [CrossRef] [Green Version]
  21. Faugeras, O.D. The Calibration Problem for Stereo; CVPR: Miami, FL, USA, 1986; pp. 15–20. [Google Scholar]
  22. Weng, J.; Cohen, P.; Herniou, M. Camera calibration with distortion models and accuracy evaluation. IEEE Trans. Pattern Anal. Mach. Intell. 1992, 10, 965–980. [Google Scholar] [CrossRef] [Green Version]
  23. Durrant-Whyte, H.; Bailey, T. Simultaneous localization and mapping: Part I. IEEE Robot. Autom. Mag. 2006, 13, 99–110. [Google Scholar] [CrossRef] [Green Version]
  24. Durrant-Whyte, H.; Bailey, T. Simultaneous localization and mapping (SLAM): Part II. IEEE Robot. Autom. Mag. 2006, 13, 108–117. [Google Scholar] [CrossRef]
  25. Özyeşil, O.; Voroninski, V.; Basri, R.; Singer, A. A survey of structure from motion. Acta Numer. 2017, 26, 305–364. [Google Scholar] [CrossRef]
  26. Fox, D.; Burgard, W.; Dellaert, F.; Thrun, S. Monte carlo localization: Efficient position estimation for mobile robots. Am. Assoc. Artif. Intell. 1999, 1999, 1–7. [Google Scholar]
  27. Mantzel, W.E.; Hyeokho, C.; Richard, G.B. Distributed camera network localization. In Proceedings of the Conference Record of the Thirty-Eighth Asilomar Conference on Signals, Systems and Computers, Pacific Grove, CA, USA, 7–10 November 2004; Volume 2, pp. 1381–1386. [Google Scholar]
  28. Brachmann, E.; Rother, C. Learning less is more-6d camera localization via 3d surface regression. In Proceedings of the Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–22 June 2018; pp. 4654–4662. [Google Scholar]
  29. Tang, Z.; Lin, Y.S.; Lee, K.H.; Hwang, J.N.; Chuang, J.H.; Fang, Z. Camera self-calibration from tracking of moving persons. In Proceedings of the 23rd International Conference on Pattern Recognition (ICPR), Cancun, Maxico, 4–8 December 2016; pp. 265–270. [Google Scholar]
  30. Zheng, C.; Qiu, H.; Liu, C.; Zheng, X.; Zhou, C.; Liu, Z.; Yang, J. A Fast Method to Extract Focal Length of Camera Based on Parallel Particle Swarm Optimization. In Proceedings of the 37th Chinese Control Conference (CCC), Wuhan, China, 25–27 July 2018; pp. 9550–9555. [Google Scholar]
  31. Führ, G.; Jung, C.R. Camera self-calibration based on nonlinear optimization and applications in surveillance systems. IEEE Trans. Circuits Syst. Video Technol. 2015, 27, 1132–1142. [Google Scholar] [CrossRef]
  32. Yao, Q.; Sankoh, H.; Nonaka, K.; Naito, S. Automatic camera self-calibration for immersive navigation of free viewpoint sports video. In Proceedings of the 18th International Workshop on Multimedia Signal Processing (MMSP), Montreal, QC, Canada, 21–23 September 2016; pp. 1–6. [Google Scholar]
  33. Li, F.; Sekkati, H.; Deglint, J.; Scharfenberger, C.; Lamm, M.; Clausi, D.; Zelek, J.; Wong, A. Simultaneous projector-camera self-calibration for three-dimensional reconstruction and projection mapping. IEEE Trans. Comput. Imaging 2017, 3, 74–83. [Google Scholar] [CrossRef]
  34. Heikkila, J. Using sparse elimination for solving minimal problems in computer vision. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 76–84. [Google Scholar]
  35. Tang, Z.; Lin, Y.S.; Lee, K.H.; Hwang, J.N.; Chuang, J.H. ESTHER: Joint camera self-calibration and automatic radial distortion correction from tracking of walking humans. IEEE Access 2019, 7, 10754–10766. [Google Scholar] [CrossRef]
  36. Marinakis, D.; Dudek, G. Topology inference for a vision-based sensor network. In Proceedings of the 2nd Canadian Conference on Computer and Robot Vision (CRV’05), Victoria, BC, Canada, 9–11 May 2005; pp. 121–128. [Google Scholar]
  37. Van Den Hengel, A.; Dick, A.; Hill, R. Activity topology estimation for large networks of cameras. In Proceedings of the IEEE International Conference on Video and Signal Based Surveillance, Sydney, Australia, 22–24 November 2006; p. 44. [Google Scholar]
  38. Detmold, H.; Van Den Hengel, A.; Dick, A.; Cichowski, A.; Hill, R.; Kocadag, E.; Falkner, K.; Munro, D.S. Topology estimation for thousand-camera surveillance networks. In Proceedings of the 1st ACM/IEEE International Conference on Distributed Smart Cameras, Vienna, Austria, 25–28 September 2007; pp. 195–202. [Google Scholar]
  39. Clarot, P.; Ermis, E.B.; Jodoin, P.M.; Saligrama, V. Unsupervised camera network structure estimation based on activity. In Proceedings of the 3rd ACM/IEEE International Conference on Distributed Smart Cameras (ICDSC), Como, Italy, 30 August–2 September 2009; pp. 1–8. [Google Scholar]
  40. Zou, X.; Bhanu, B.; Song, B.; Roy-Chowdhury, A.K. Determining topology in a distributed camera network. In Proceedings of the IEEE International Conference on Image Processing, San Antonio, TX, USA, 16–19 September 2007; Volume 5, p. V-133. [Google Scholar]
  41. Farrell, R.; Davis, L.S. Decentralized discovery of camera network topology. In Proceedings of the 2nd ACM/IEEE International Conference on Distributed Smart Cameras, Palo Alto, CA, USA, 7–11 September 2008; pp. 1–10. [Google Scholar]
  42. Zhu, M.; Dick, A.; van den Hengel, A. Camera network topology estimation by lighting variation. In Proceedings of the International Conference on Digital Image Computing: Techniques and Applications (DICTA), Adelaide, Australia, 23–25 November 2015; pp. 1–6. [Google Scholar]
  43. Mali, G.; Sudip, M. TRAST: Trust-based distributed topology management for wireless multimedia sensor networks. IEEE Trans. Comput. 2015, 65, 1978–1991. [Google Scholar] [CrossRef]
  44. Feigang, T.; Xiaoju, Z.; Quanmi, L.; Jianyi, L. A Camera Network Topology Estimation Based on Blind Distance. In Proceedings of the 11th International Conference on Intelligent Computation Technology and Automation (ICICTA), Changsha, China, 22–23 September 2018; pp. 138–140. [Google Scholar]
  45. Li, Z.; Wang, J.; Chen, J. Estimating Path in camera network with non-overlapping FOVs. In Proceedings of the 5th International Conference on Systems and Informatics (ICSAI), Nanjing, China, 10–12 November 2018; pp. 604–609. [Google Scholar]
  46. Kansal, A.; Srivastava, M.B. An environmental energy harvesting framework for sensor networks. In Proceedings of the 2003 International Symposium on Low Power Electronics and Design, Seoul, Korea, 25–27 August 2003; pp. 481–486. [Google Scholar]
  47. Bramberger, M.; Quaritsch, M.; Winkler, T.; Rinner, B.; Schwabach, H. Integrating multi-camera tracking into a dynamic task allocation system for smart cameras. In Proceedings of the IEEE Conference on Advanced Video and Signal Based Surveillance, Como Italy, 15–16 September 2005; pp. 474–479. [Google Scholar]
  48. Bramberger, M.; Rinner, B.; Schwabach, H. A method for dynamic allocation of tasks in clusters of embedded smart cameras. In Proceedings of the IEEE International Conference on Systems, Man and Cybernetics, Waikolova, HI, USA, 12 October 2005; Volume 3, pp. 2595–2600. [Google Scholar]
  49. Karuppiah, D.R.; Grupen, R.A.; Zhu, Z.; Hanson, A.R. Automatic resource allocation in a distributed camera network. Mach. Vis. Appl. 2010, 21, 517–528. [Google Scholar] [CrossRef]
  50. Dieber, B.; Micheloni, C.; Rinner, B. Resource-aware coverage and task assignment in visual sensor networks. IEEE Trans. Circuits Syst. Video Technol. 2011, 21, 1424–1437. [Google Scholar] [CrossRef]
  51. Dieber, B.; Esterle, L.; Rinner, B. Distributed resource-aware task assignment for complex monitoring scenarios in visual sensor networks. In Proceedings of the 6th International Conference on Distributed Smart Cameras (ICDSC), Hong Kong, China, 30 October–2 November 2012; pp. 1–6. [Google Scholar]
  52. Kyrkou, C.; Laoudias, C.; Theocharides, T.; Panayiotou, C.G.; Polycarpou, M. Adaptive energy-oriented multitask allocation in smart camera networks. IEEE Embed. Syst. Lett. 2016, 8, 37–40. [Google Scholar] [CrossRef]
  53. Wang, Y.; Zhang, J.; Liu, Z.; Wu, Q.; Chou, P.A.; Zhang, Z.; Jia, Y. Handling occlusion and large displacement through improved RGB-D scene flow estimation. IEEE Trans. Circuits Syst. Video Technol. 2015, 26, 1265–1278. [Google Scholar] [CrossRef] [Green Version]
  54. Ouyang, W.; Zeng, X.; Wang, X. Partial occlusion handling in pedestrian detection with a deep model. IEEE Trans. Circuits Syst. Video Technol. 2015, 26, 2123–2137. [Google Scholar] [CrossRef]
  55. Shehzad, M.I.; Shah, Y.A.; Mehmood, Z.; Malik, A.W.; Azmat, S. K-means based multiple objects tracking with long-term occlusion handling. IET Comput. Vis. 2016, 11, 68–77. [Google Scholar] [CrossRef]
  56. Ur-Rehman, A.; Naqvi, S.M.; Mihaylova, L.; Chambers, J.A. Multi-target tracking and occlusion handling with learned variational Bayesian clusters and a social force model. IEEE Trans. Signal Processing 2015, 64, 1320–1335. [Google Scholar] [CrossRef] [Green Version]
  57. Chang, J.; Wang, L.; Meng, G.; Xiang, S.; Pan, C. Vision-based occlusion handling and vehicle classification for traffic surveillance systems. IEEE Intell. Transp. Syst. Mag. 2018, 10, 80–92. [Google Scholar] [CrossRef]
  58. Zhao, S.; Zhang, S.; Zhang, L. Towards occlusion handling: Object tracking with background estimation. IEEE Trans. Cybern. 2017, 48, 2086–2100. [Google Scholar] [CrossRef]
  59. Liu, Y.; Jing, X.Y.; Nie, J.; Gao, H.; Liu, J.; Jiang, G.P. Context-Aware Three-Dimensional Mean-Shift with Occlusion Handling for Robust Object Tracking in RGB-D Videos. IEEE Trans. Multimed. 2018, 21, 664–677. [Google Scholar] [CrossRef]
  60. Feng, X.; Jiang, Y.; Yang, X.; Du, M.; Li, X. Computer vision algorithms and hardware implementations: A survey. Integration 2019, 69, 309–320. [Google Scholar] [CrossRef]
  61. Hørup, S.A.; Juul, S.A. General-Purpose Computations; Aalborg University: Aalborg, Denmark, 2011. [Google Scholar]
  62. Guo, Y.; Liu, J.; Li, G.; Mai, L.; Dong, H. Fast and Flexible Human Pose Estimation with HyperPose. In Proceedings of the 29th ACM International Conference on Multimedia, Chengdu, China, 20–24 October 2021; pp. 3763–3766. [Google Scholar]
  63. Tan, S.; Knott, B.; Tian, Y.; Wu, D.J. CryptGPU: Fast privacy-preserving machine learning on the GPU. In Proceedings of the IEEE Symposium on Security and Privacy (SP), San Francisco, CA, USA, 24–27 May 2021; pp. 1021–1038. [Google Scholar]
  64. Irmak, H.; Ziener, D.; Alachiotis, N. Increasing Flexibility of FPGA-based CNN Accelerators with Dynamic Partial Reconfiguration. In Proceedings of the 31st International Conference on Field-Programmable Logic and Applications (FPL), Dredsen, Germany, 1–3 September 2021; pp. 306–311. [Google Scholar]
  65. Costa, A.; Corna, N.; Garzetti, F.; Lusardi, N.; Ronconi, E.; Geraci, A. High-Performance Computing of Real-Time and Multichannel Histograms: A Full FPGA Approach. IEEE Access 2022, 10, 47524–47540. [Google Scholar] [CrossRef]
  66. Carbajal, M.A.; Villa, R.P.; Palazuelos, D.E.; Astorga, G.J. Rubio Astorga; Reconfigurable Digital FPGA Based Architecture for 2-Dimensional Linear Convolution Applications; Identitad Energetica: Madrid, Spain, 2021. [Google Scholar]
  67. Xiong, H.; Sun, K.; Zhang, B.; Yang, J.; Xu, H. Deep-Sea: A Reconfigurable Accelerator for Classic CNN. Wirel. Commun. Mob. Comput. 2022, 2022, 4726652. [Google Scholar] [CrossRef]
  68. Wei, L.; Peng, L. An Efficient OpenCL-Based FPGA Accelerator for MobileNet. J. Phys. Conf. Ser. 2021, 1883, 012086. [Google Scholar] [CrossRef]
  69. Szeliski, R. Scene Reconstruction from multiple cameras. In Proceedings of the International Conference on Image Processing (ICISP), Vancouver, BC, Canada, 10–13 September 2000; Volume 1, pp. 13–16. [Google Scholar]
  70. Micušık, B.; Martinec, D.; Pajdla, T. 3D metric reconstruction from uncalibrated omnidirectional images. In Proceedings of the Asian Conference on Computer Vision (ACCV’04), Jeju Island, Korea, January 2014. [Google Scholar]
  71. Peng, L.; Zhang, Y.; Zhou, H.; Lu, T. A robust method for estimating image geometry with local structure constraint. IEEE Access 2018, 6, 20734–20747. [Google Scholar] [CrossRef]
  72. Brito, D.N.; Nunes, C.F.; Padua, F.L.; Lacerda, A. Evaluation of interest point matching methods for projective reconstruction of 3d scenes. IEEE Lat. Am. Trans. 2016, 14, 1393–1400. [Google Scholar] [CrossRef]
  73. Milani, S. Three-dimensional reconstruction from heterogeneous video devices with camera-in-view information. In Proceedings of the IEEE International Conference on Image Processing (ICIP), Quebec, QC, Canada, 27–30 September 2015; pp. 2050–2054. [Google Scholar]
  74. Aliakbarpour, H.; Prasath, V.S.; Palaniappan, K.; Seetharaman, G.; Dias, J. Heterogeneous multi-view information fusion: Review of 3-D reconstruction methods and a new registration with uncertainty modeling. IEEE Access 2016, 4, 8264–8285. [Google Scholar] [CrossRef]
  75. Wang, C.; Guo, X. Plane-Based Optimization of Geometry and Texture for RGB-D Reconstruction of Indoor Scenes. In Proceedings of the International Conference on 3D Vision (3DV), Verona, Italy, 5–8 September 2018; pp. 533–541. [Google Scholar]
  76. Ma, D.; Li, G.; Wang, L. Rapid Reconstruction of a Three-Dimensional Mesh Model Based on Oblique Images in the Internet of Things. IEEE Access 2018, 6, 61686–61699. [Google Scholar] [CrossRef]
  77. Ichimaru, K.; Furukawa, R.; Kawasaki, H. CNN based dense underwater 3D scene reconstruction by transfer learning using bubble database. In Proceedings of the IEEE Winter Conference on Applications of Computer Vision (WACV), Waikolova Village, HI, USA, 7–9 January 2019; pp. 1543–1552. [Google Scholar]
  78. Viola, P.; Jones, M. Robust real-time object detection. Int. J. Comput. Vis. 2001, 4, 34–47. [Google Scholar]
  79. Piccinini, P.; Prati, A.; Cucchiara, R. Real-time object detection and localization with SIFT-based clustering. Image Vis. Comput. 2012, 30, 573–587. [Google Scholar] [CrossRef]
  80. Dalal, N.; Triggs, B. Histograms of oriented gradients for human detection. In Proceedings of the IEEE computer society conference on computer vision and pattern recognition (CVPR’05), San Diago, CA, USA, 20–26 June 2005; Volume 1, pp. 886–893. [Google Scholar]
  81. Aslani, S.; Mahdavi-Nasab, H. Optical flow based moving object detection and tracking for traffic surveillance. Int. J. Electr. Comput. Energetic Electron. Commun. Eng. 2013, 7, 1252–1256. [Google Scholar]
  82. Huang, J.; Zou, W.; Zhu, J.; Zhu, Z. Optical flow based real-time moving object detection in unconstrained scenes. arXiv 2018, arXiv:1807.04890. [Google Scholar]
  83. Tougaard, S. Practical algorithm for background subtraction. Surf. Sci. 1989, 216, 343–360. [Google Scholar] [CrossRef]
  84. Rieke, J. Object detection with neural networks-a simple tutorial using keras. Towards Data Sci. 2017, 6, 12. [Google Scholar]
  85. Redmon, J.; Divvala, S.; Girshick, R.; Farhadi, A. You only look once: Unified, real-time object detection. In Proceedings of the Computer Vision and Pattern Recognition, Las Vegas, USA, 27–30 June 2016; pp. 779–788. [Google Scholar]
  86. Girshick, R.; Donahue, J.; Darrell, T.; Malik, J. Rich feature hierarchies for accurate object detection and semantic segmentation. In Proceedings of the Computer Vision and Pattern Recognition, Columbus, OH, USA, 23–28 June 2014; pp. 580–587. [Google Scholar]
  87. Zhang, S.; Wen, L.; Bian, X.; Lei, Z.; Li, S.Z. Single-shot refinement neural network for object detection. In Proceedings of the Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 4203–4212. [Google Scholar]
  88. Pang, J.; Chen, K.; Shi, J.; Feng, H.; Ouyang, W.; Lin, D. Libra r-cnn: Towards balanced learning for object detection. In Proceedings of the Computer vision and pattern recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 821–830. [Google Scholar]
  89. Liu, W.; Anguelov, D.; Erhan, D.; Szegedy, C.; Reed, S.; Fu, C.Y.; Berg, A.C. Ssd: Single shot multibox detector. In Proceedings of the European Conference on Computer Vision, Amsterdam, The Netherlands, 11–14 October 2016; pp. 21–37. [Google Scholar]
  90. Roy, S.M.; Ghosh, A. Real-time adaptive Histogram Min-Max Bucket (HMMB) model for background subtraction. IEEE Trans. Circuits Syst. Video Technol. 2017, 28, 1513–1525. [Google Scholar] [CrossRef]
  91. Min, W.; Fan, M.; Guo, X.; Han, Q. A new approach to track multiple vehicles with the combination of robust detection and two classifiers. IEEE Trans. Intell. Transp. Syst. 2017, 19, 174–186. [Google Scholar] [CrossRef]
  92. Wu, Y.; He, X.; Nguyen, T.Q. Moving object detection with a freely moving camera via background motion subtraction. IEEE Trans. Circuits Syst. Video Technol. 2015, 27, 236–248. [Google Scholar] [CrossRef]
  93. Hu, W.; Yang, Y.; Zhang, W.; Xie, Y. Moving object detection using tensor-based low-rank and saliently fused-sparse decomposition. IEEE Trans. Image Processing 2016, 26, 724–737. [Google Scholar] [CrossRef]
  94. Parekh, H.S.; Thakore, D.G.; Jaliya, U.K. A survey on object detection and tracking methods. Int. J. Innov. Res. Comput. Commun. Eng. 2014, 2, 2970–2979. [Google Scholar]
  95. Yilmaz, A.; Javed, O.; Shah, M. Object tracking: A survey. Acm. Comput. Surv. 2006, 38, 13. [Google Scholar] [CrossRef]
  96. Du, C.-J.; Sun, D.-W. Object Classification Methods. In Computer Vision Technology for Food Quality Evaluation; Academic Press: Cambridge, MA, USA, 2016; pp. 87–110. [Google Scholar]
  97. Ankerst, M.; Elsen, C.; Ester, M.; Kriegel, H.P. Visual classification: An interactive approach to decision tree construction. In Proceedings of the 5th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, San Diago, CA, USA, 15–18 August 1999; pp. 392–396. [Google Scholar]
  98. Anurag, S.; Han, E.; Kumar, V.; Singh, V. Parallel formulations of decision-tree classification algorithms. In High Performance Data Mining; Springer: Boston, MA, USA, 1999; pp. 237–261. [Google Scholar]
  99. Schroff, F.; Criminisi, A.; Zisserman, A. Object Class Segmentation using Random Forests. In Proceedings of the British Machine Vision Conference, University of Leeds, Leeds, UK, 1–4 September 2008; pp. 1–10. [Google Scholar]
  100. Bayes, T.; LII. An essay towards solving a problem in the doctrine of chances. By the late Rev. Mr. Bayes, FRS communicated by Mr. Price, in a letter to John Canton, AMFR S. Philos. Trans. R. Soc. Lond. 1763, 53, 370–418. [Google Scholar]
  101. Leung, K.M. Naive Bayesian Classifier; Polytechnic University Department of Computer Science/Finance and Risk Engineering: New York, NY, USA, 2007; pp. 123–156. [Google Scholar]
  102. Kononenko, I. Semi-Naive Bayesian Classifier. In European Working Session on Learning; Springer: Berlin/Heidelberg, Germany, 1991; pp. 206–219. [Google Scholar]
  103. Klecka, W.R.; Gudmund, R.I.; Klecka, W.R. Discriminant Analysis; Sage: New York, NY, USA, 1980; Volume 19. [Google Scholar]
  104. Menard, S. Applied Logistic Regression Analysis; Sage: New York, NY, USA, 2002; Volume 106. [Google Scholar]
  105. Hastie, T.; Tibshirani, R. Discriminant adaptive nearest neighbor classification. IEEE Trans. Pattern Anal. Mach. Intell. 1996, 18, 607–616. [Google Scholar] [CrossRef] [Green Version]
  106. Durgesh, K.S.; Lekha, B. Data classification using support vector machine. J. Theor. Appl. Inf. Technol. 2010, 12, 1–7. [Google Scholar]
  107. Lawrence, S.; Giles, C.L.; Tsoi, A.C.; Back, A.D. Face recognition: A convolutional neural-network approach. IEEE Trans. Neural Netw. 1997, 8, 98–113. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  108. Murtagh, F. Multilayer perceptrons for classification and regression. Neurocomputing 1991, 2, 183–197. [Google Scholar] [CrossRef]
  109. Jmour, N.; Zayen, S.; Abdelkrim, A. Convolutional neural networks for image classification. In Proceedings of the International Conference on Advanced Systems and Electric Technologies (IC_ASET), Hammamet, Tinisia, 22–25 March 2018; pp. 397–402. [Google Scholar]
  110. De Villiers, J.P.; Leuschner, F.W.; Geldenhuys, R. Centi-pixel accurate real-time inverse distortion correction. In Proceedings of the International Symposium on Optomechatronic Technologies, West Harbor, San Diago, CA, USA, 17–19 November 2008; Volume 7266, p. 726611. [Google Scholar]
  111. Caprile, B.; Torre, V. Using vanishing points for camera calibration. Int. J. Comput. Vis. 1990, 4, 127–139. [Google Scholar] [CrossRef]
  112. Wang, A.; Qiu, T.; Shao, L. A simple method of radial distortion correction with centre of distortion estimation. J. Math. Imaging Vis. 2009, 35, 165–172. [Google Scholar] [CrossRef]
  113. Hartley, R.; Kang, S.B. Parameter-free radial distortion correction with center of distortion estimation. IEEE Trans. Pattern Anal. Mach. Intell. 2007, 29, 1309–1321. [Google Scholar] [CrossRef] [Green Version]
  114. Huang, K.; Ziauddin, S.; Zand, M.; Greenspan, M. One Shot Radial Distortion Correction by Direct Linear Transformation. In Proceedings of the IEEE International Conference on Image Processing (ICIP), Abu Dhabi, Dubai, 25–28 October 2020; pp. 473–477. [Google Scholar]
  115. Zhao, H.; Shi, Y.; Tong, X.; Ying, X.; Zha, H. A Simple Yet Effective Pipeline For Radial Distortion Correction. In Proceedings of the IEEE International Conference on Image Processing (ICIP), Abu Dhabi, Dubai, 25–28 October 2020; pp. 878–882. [Google Scholar]
  116. Wang, Y.M.; Li, Y.; Zheng, J.B. A camera calibration technique based on OpenCV. In Proceedings of the 3rd International Conference on Information Sciences and Interaction Sciences, Chengdu, China, 23–25 June 2010; pp. 403–406. [Google Scholar]
  117. Lee, S.; Hong, H. A robust camera-based method for optical distortion calibration of head-mounted displays. J. Disp. Technol. 2014, 11, 845–853. [Google Scholar] [CrossRef]
  118. Wang, Z.; Liu, M.; Yang, S.; Huang, S.; Bai, X.; Liu, X.; Zhu, J.; Liu, X.; Zhang, Z. Precise full-field distortion rectification and evaluation method for a digital projector. Opt. Rev. 2016, 23, 746–752. [Google Scholar] [CrossRef]
  119. Yang, S.; Srikanth, M.; Lelescu, D.; Venkataraman, K. Systems and Methods for Depth-Assisted Perspective Distortion Correction. U.S. Patent 9,898,856, 20 February 2018. [Google Scholar]
  120. Finlayson, G.; Gong, H.; Fisher, R.B. Color homography: Theory and applications. IEEE Trans. Pattern Anal. Mach. Intell. 2017, 41, 20–33. [Google Scholar] [CrossRef] [Green Version]
  121. Wang, H.; Yang, J.; Xue, B.; Yan, X.; Tao, J. A novel color calibration method of multi-spectral camera based on normalized RGB color model. Results Phys. 2020, 19, 103498. [Google Scholar] [CrossRef]
  122. Han, S.; Huang, P.; Wang, H.; Yu, E.; Liu, D.; Pan, X. Mat: Motion-aware multi-object tracking. Neurocomputing 2022, 476, 75–86. [Google Scholar] [CrossRef]
  123. Meinhardt, T.; Kirillov, A.; Leal-Taixe, L.; Feichtenhofer, C. Trackformer: Multi-object tracking with transformers. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–22 June 2022; pp. 8844–8854. [Google Scholar]
  124. Cucchiara, R.; Grana, C.; Piccardi, M.; Prati, A. Detecting moving objects, ghosts, and shadows in video streams. IEEE Trans. Pattern Anal. Mach. Intell. 2003, 25, 1337–1342. [Google Scholar] [CrossRef] [Green Version]
  125. Zhang, W.; Ma, B.; Liu, K.; Huang, R. Video-based pedestrian re-identification by adaptive spatio-temporal appearance model. IEEE Trans. Image Processing 2017, 26, 2042–2054. [Google Scholar] [CrossRef]
  126. Yang, X.; Wang, M.; Tao, D. Person re-identification with metric learning using privileged information. IEEE Trans. Image Processing 2017, 27, 791–805. [Google Scholar] [CrossRef] [Green Version]
  127. Geng, S.; Yu, M.; Guo, Y.; Yu, Y. A Weighted Center Graph Fusion Method for Person Re-Identification. IEEE Access 2019, 7, 23329–23342. [Google Scholar] [CrossRef]
  128. Yang, X.; Tang, Y.; Wang, N.; Song, B.; Gao, X. An End-to-End Noise-Weakened Person Re-Identification and Tracking with Adaptive Partial Information. IEEE Access 2019, 7, 20984–20995. [Google Scholar] [CrossRef]
  129. Chen, T.; Fang, C.; Shen, X.; Zhu, Y.; Chen, Z.; Luo, J. Anatomy-aware 3d human pose estimation with bone-based pose decomposition. IEEE Trans. Circuits Syst. Video Technol. 2021, 32, 198–209. [Google Scholar] [CrossRef]
  130. Straka, M.; Hauswiesner, S.; Rüther, M.; Bischof, H. Skeletal Graph Based Human Pose Estimation in Real-Time. In Proceedings of the BMVC, Dundee, UK, 29 August–2 September 2011; pp. 1–12. [Google Scholar]
  131. Campbell, L.W.; Bobick, A.F. Using phase space constraints to represent human body motion. In Proceedings of the International Workshop on Automatic Face and Gesture Recognition, Zurich, Switzerland, 26–28 June 1995; pp. 338–343. [Google Scholar]
  132. Oren, M.; Papageorgiou, C.; Sinha, P.; Osuna, E.; Poggio, T. Pedestrian detection using wavelet templates. In Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition, Puerto Rico, USA, 17–19 June 1997; Volume 97, pp. 193–199. [Google Scholar]
  133. You, Q.; Jin, H.; Wang, Z.; Fang, C.; Luo, J. Image captioning with semantic attention. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, GA, USA, 27–30 June 2016; pp. 4651–4659. [Google Scholar]
  134. Wcg, P. Role of the manuscript reviewer. Singap. Med. J. 2009, 50, 931–934. [Google Scholar]
  135. Polak, J.F. The role of the manuscript reviewer in the peer review process. Am. J. Roentgenol. 1995, 165, 685–688. [Google Scholar] [CrossRef] [Green Version]
  136. Nguyen, H.; Bhanu, B.; Patel, A.; Diaz, R. VideoWeb: Design of a wireless camera network for real-time monitoring of activities. In Proceedings of the 3rd ACM/IEEE International Conference on Distributed Smart Cameras (ICDSC), Como, Italy, 30 August –2 September 2009; pp. 1–8. [Google Scholar]
  137. Ibraheem, O.W.; Irwansyah, A.; Hagemeyer, J.; Porrmann, M.; Rueckert, U. Reconfigurable vision processing system for player tracking in indoor sports. In Proceedings of the Conference on Design and Architectures for Signal and Image Processing (DASIP), Dresden, Germany, 27–29 September 2017; pp. 1–6. [Google Scholar]
  138. Xiang, Y.; Alahi, A.; Savarese, S. Learning to Track: Online multi-object tracking by decision making. In Proceedings of the IEEE International Conference on Computer Vision, Santiago, Chile, 7–13 December 2015; pp. 4705–4713. [Google Scholar]
  139. Laptev, I.; Caputo, B. Recognizing human actions: A local SVM approach. In Proceedings of the 17th International Conference on Pattern Recognition, Washington, DC, USA, 23–26 August 2004; pp. 32–36. [Google Scholar]
  140. Duy, A.N.; Yoo, M. Calibration-Net: LiDAR and Camera Auto-Calibration using Cost Volume and Convolutional Neural Network. In Proceedings of the 2022 International Conference on Artificial Intelligence in Information and Communication (ICAIIC), Jeju Island, Korea, 21–24 February 2022; pp. 141–144. [Google Scholar]
  141. Cao, Y.; Wang, H.; Zhao, H.; Yang, X. Neural-Network-Based Model-Free Calibration Method for Stereo Fisheye Camera. Front. Bioeng. Biotechnol. 2022, 10, 955233. [Google Scholar] [CrossRef]
  142. Chen, H.; Munir, S.; Lin, S. RFCam: Uncertainty-aware Fusion of Camera and Wi-Fi for Real-time Human Identification with Mobile Devices. Proc. ACM Interact. Mob. Wearable Ubiquitous Technol. 2022, 6, 1–29. [Google Scholar] [CrossRef]
  143. Dufera, T.T.; Seboka, Y.C.; Portillo, C.F. Parameter Estimation for Dynamical Systems Using a Deep Neural Network. Appl. Comput. Intell. Soft Comput. 2022, 2022, 2014510. [Google Scholar] [CrossRef]
  144. Doula, A.; Sanchez Guinea, A.; Mühlhäuser, M. VR-Surv: A VR-Based Privacy Preserving Surveillance System. In Proceedings of the CHI Conference on Human Factors in Computing Systems Extended Abstracts, New Orleans, NY, USA, 29 April–5 May 2022; pp. 1–7. [Google Scholar]
  145. Pooja, C.; Jaisharma, K. Novel Framework for the Improvement of Object Detection Accuracy of Smart Surveillance Camera Visuals Using Modified Convolutional Neural Network Technique Compared with Global Color Histogram. ECS Trans. 2022, 107, 18823. [Google Scholar] [CrossRef]
  146. Jiang, T.; Zhang, Q.; Yuan, J.; Wang, C.; Li, C. Multi-Type Object Tracking Based on Residual Neural Network Model. Symmetry 2022, 14, 1689. [Google Scholar] [CrossRef]
  147. Jaganathan, T.; Panneerselvam, A.; Kumaraswamy, S.K. Object detection and multi-object tracking based on optimized deep convolutional neural network and unscented Kalman filtering. Concurr. Comput. Pr. Exp. 2022, 34, e7245. [Google Scholar] [CrossRef]
  148. Deshpande, T.R.; Sapkal, S.U. Development of Object Tracking System Utilizing Camera Movement and Deep Neural Network. In Proceedings of the 2022 IEEE Region 10 Symposium (TENSYMP), Mumbai, India, 1–3 July 2022; pp. 1–6. [Google Scholar]
  149. Praveenkumar, S.M.; Patil, P.; Hiremath, P.S. Real-Time Multi-Object Tracking of Pedestrians in a Video Using Convolution Neural Network and Deep SORT. In Proceedings of the ICT Systems and Sustainability, (ICT4SD), Goa, India, 23–24 July 2022; pp. 725–736. [Google Scholar]
  150. Jhansi, M.; Bachu, S.; Kumar, N.U.; Kumar, M.A. IODTDLCNN: Implementation of Object Detection and Tracking by using Deep Learning based Convolutional Neural Network. In Proceedings of the 2022 First International Conference on Electrical, Electronics, Information and Communication Technologies (ICEEICT), Trichy, India, 16–18 February 2022; pp. 1–6. [Google Scholar]
  151. Barazande, J.; Farzaneh, N. WSAMLP: Water Strider Algorithm and Artificial Neural Network-based Activity Detection Method in Smart Homes. J. AI Data Min. 2022, 10, 1–13. [Google Scholar]
  152. Wong, P.K.Y.; Luo, H.; Wang, M.; Cheng, J.C. Enriched and discriminative convolutional neural network features for pedestrian re-identification and trajectory modelling. Comput.-Aided Civ. Infrastruct. Eng. 2022, 37, 573–592. [Google Scholar] [CrossRef]
  153. Yao, Y.; Jiang, X.; Fujita, H.; Fang, Z. A sparse graph wavelet convolution neural network for video-based person re-identification. Pattern Recognit. 2022, 129, 108708. [Google Scholar] [CrossRef]
  154. Mohana, M.; Alelyani, S.; Alsaqer, M.S. Fused Deep Neural Network based Transfer Learning in Occluded Face Classification and Person re-Identification. arXiv 2022, arXiv:2205.07203. [Google Scholar]
  155. You, C.; Zheng, H.; Guo, Z.; Wang, T.; Wu, X. Tampering detection and localization base on sample guidance and individual camera device convolutional neural network features. Expert Syst. 2022, e13102. [Google Scholar] [CrossRef]
  156. Karamchandani, S.; Bhattacharjee, S.; Issrani, D.; Dhar, R. SLAM Using Neural Network-Based Depth Estimation for Auto Vehicle Parking. In IOT with Smart Systems; Springer: Berlin/Heidelberg, Germany, 2022; pp. 37–44. [Google Scholar]
  157. AI In Computer Vision Market Research Report by Component (Hardware, Software), Vertical (Healthcare, Security, Automotive, Agriculture, Sports & Entertainment, and Others), and Region–Global Forecast to 2027. Available online: https://www.expertmarketresearch.com/reports/ai-in-computer-vision-market (accessed on 22 August 2022).
  158. Andriyanov, N. Methods for preventing visual attacks in convolutional neural networks based on data discard and dimensionality reduction. Appl. Sci. 2021, 11, 5235. [Google Scholar] [CrossRef]
  159. Wang, B.; Zhao, M.; Wang, W.; Dai, X.; Li, Y.; Guo, Y. Adversarial Analysis for Source Camera Identification. IEEE Trans. Circuits Syst. Video Technol. 2021, 31, 4174–4186. [Google Scholar] [CrossRef]
  160. Zhang, C.; Benz, P.; Lin, C.; Karjauv, A.; Wu, J.; Kweon, I.S. A survey on universal adversarial attack. arXiv 2021, arXiv:2103.01498. [Google Scholar]
  161. Edwards DRawat, D.B. Study of Adversarial Machine Learning with Infrared Examples for Surveillance Applications. Electronics 2020, 9, 1284. [Google Scholar] [CrossRef]
  162. Chakraborty, A.; Alam, M.; Dey, V.; Chattopadhyay, A.; Mukhopadhyay, D. A survey on adversarial attacks and defences. CAAI Trans. Intell. Technol. 2021, 6, 25–45. [Google Scholar] [CrossRef]
  163. Akhtar, N.; Mian, A.; Kardan, N.; Shah, M. Advances in adversarial attacks and defenses in computer vision: A survey. IEEE Access 2021, 9, 155161–155196. [Google Scholar] [CrossRef]
  164. SanMiguel, J.C.; Micheloni, C.; Shoop, K.; Foresti, G.L.; Cavallaro, A. Self-reconfigurable smart camera networks. Computer 2014, 47, 67–73. [Google Scholar]
  165. Leong, W.L.; Martinel, N.; Huang, S.; Micheloni, C.; Foresti, G.L.; Teo, R.S. An Intelligent Auto-Organizing Aerial Robotic Sensor Network System for Urban Surveillance. J. Intell. Robot. Syst. 2021, 102, 33. [Google Scholar] [CrossRef]
  166. Natarajan, P.; Atrey, P.K.; Kankanhalli, M. Multi-camera coordination and control in surveillance systems: A survey. ACM Trans. Multimed. Comput. Commun. Appl. 2015, 11, 1–30. [Google Scholar] [CrossRef]
  167. Martinel, N.; Dunnhofer, M.; Pucci, R.; Foresti, G.L.; Micheloni, C. Lord of the rings: Hanoi pooling and self-knowledge distillation for fast and accurate vehicle reidentification. IEEE Trans. Ind. Inform. 2021, 18, 87–96. [Google Scholar] [CrossRef]
  168. Rinner, B.; Esterle, L.; Simonjan, J.; Nebehay, G.; Pflugfelder, R.; Dominguez, G.F.; Lewis, P.R. Self-aware and self-expressive camera networks. Computer 2015, 48, 21–28. [Google Scholar] [CrossRef]
  169. Lewis, P.R.; Chandra, A.; Glette, K. Self-awareness and Self-expression: Inspiration from Psychology. In Self-Aware Computing Systems; Springer: Berlin/Heidelberg, Germany, 2016; pp. 9–21. [Google Scholar]
  170. Glette, K.; Lewis, P.R.; Chandra, A. Relationships to Other Concepts. In Self-aware Computing Systems; Springer: Berlin/Heidelberg, Germany, 2016; pp. 23–35. [Google Scholar]
  171. Wang, S.; Nebehay, G.; Esterle, L.; Nymoen, K.; Minku, L.L. Common Techniques for Self-awareness and Self-expression. In Self-Aware Computing Systems; Natural Computing Series; Springer: Berlin/Heidelberg, Germany, 2016. [Google Scholar]
  172. Isavudeen, A.; Ngan, N.; Dokladalova, E.; Akil, M. Auto-adaptive multi-sensor architecture. In Proceedings of the International Symposium on Circuits and Systems (ISCAS), Montreal, QC, Canada, 22–26 May 2016; pp. 2198–2201. [Google Scholar]
  173. Guettatfi, Z.; Hübner, P.; Platzner, M.; Rinner, B. Computational self-awareness as design approach for visual sensor nodes. In Proceedings of the 12th International Symposium on Reconfigurable Communication-centric Systems-on-Chip (ReCoSoC), Madrid, Spain, 12–17 July 2017; pp. 1–8. [Google Scholar]
  174. Zhu, Z.; Luo, Y.; Chen, S.; Qi, G.; Mazur, N.; Zhong, C.; Li, Q. Camera style transformation with preserved self-similarity and domain-dissimilarity in unsupervised person re-identification. J. Vis. Commun. Image Represent. 2021, 80, 103303. [Google Scholar] [CrossRef]
  175. Lin, S.; Lv, J.; Yang, Z.; Li, Q.; Zheng, W.S. Heterogeneous graph driven unsupervised domain adaptation of person re-identification. Neurocomputing 2022, 471, 1–11. [Google Scholar] [CrossRef]
  176. Wu, M.; Li, C.; Yao, Z. Deep Active Learning for Computer Vision Tasks: Methodologies, Applications, and Challenges. Appl. Sci. 2022, 12, 8103. [Google Scholar] [CrossRef]
  177. Rudolph, S.; Tomforde, S.; Hähner, J. On the Detection of Mutual Influences and Their Consideration in Reinforcement Learning Processes. arXiv 2019, arXiv:1905.04205. [Google Scholar]
  178. Cai, L.; Ma, H.; Liu, Z.; Li, Z.; Zhou, Z. Coverage Control for PTZ Camera Networks Using Scene Potential Map. In Proceedings of the IEEE International Conference on Multimedia and Expo (ICME), Taipei, Taiwan, 11–15 July 2022; pp. 1–6. [Google Scholar]
  179. Suresh, S.; Menon, V. An Efficient Graph Based Approach for Reducing Coverage Loss From Failed Cameras of a Surveillance Network. IEEE Sens. J. 2022, 22, 8155–8163. [Google Scholar]
  180. Liang, K. Fission: A Provably Fast, Scalable, and Secure Permissionless Blockchain. arXiv 2018, arXiv:1812.05032. [Google Scholar]
  181. Zhao, W. On Nxt Proof of Stake Algorithm: A Simulation Study. IEEE Trans. Dependable Secur. Comput. 2022. Available online: https://ieeexplore.ieee.org/document/9837462 (accessed on 22 August 2022). [CrossRef]
Figure 1. A general process flow of a computer vision system employing a smart camera network.
Figure 1. A general process flow of a computer vision system employing a smart camera network.
Symmetry 14 02281 g001
Figure 2. Taxonomy of challenges in active vision applications employing camera networks.
Figure 2. Taxonomy of challenges in active vision applications employing camera networks.
Symmetry 14 02281 g002
Figure 3. Self-reconfiguration framework as proposed in [8].
Figure 3. Self-reconfiguration framework as proposed in [8].
Symmetry 14 02281 g003
Figure 4. Adaptive self-reconfiguration framework for SCN-enabled active vision systems.
Figure 4. Adaptive self-reconfiguration framework for SCN-enabled active vision systems.
Symmetry 14 02281 g004
Figure 5. Flow diagram illustrating the adaptive reconfiguration using the ASR framework.
Figure 5. Flow diagram illustrating the adaptive reconfiguration using the ASR framework.
Symmetry 14 02281 g005
Figure 6. Comparison of the performances of the system in [8] and the proposed ASR for surveillance dataset 1.
Figure 6. Comparison of the performances of the system in [8] and the proposed ASR for surveillance dataset 1.
Symmetry 14 02281 g006
Figure 7. Predicted motion of multiple objects in random frames from surveillance dataset 1.
Figure 7. Predicted motion of multiple objects in random frames from surveillance dataset 1.
Symmetry 14 02281 g007
Figure 8. Comparison of the performances of the system in [8] and the proposed ASR for surveillance dataset 2.
Figure 8. Comparison of the performances of the system in [8] and the proposed ASR for surveillance dataset 2.
Symmetry 14 02281 g008
Figure 9. Predicted motions of multiple objects in random frames from surveillance dataset 2.
Figure 9. Predicted motions of multiple objects in random frames from surveillance dataset 2.
Symmetry 14 02281 g009
Table 1. Evolution of sensor-placement techniques.
Table 1. Evolution of sensor-placement techniques.
Ref.YearMethodologyAdvantages
[16]2008Online system for tracking multiple people in an SCN with overlapping and non-overlapping viewsDevelopment of a larger, more capable, and fully automatic system without prior localization information
[10]2009Genetic
algorithm
Maximum coverage of users;
Defined priority areas with optimum values of parameters;
The proposed algorithm works offline
and does not require camera calibration;
Minimizes the probability of occlusion due to randomly moving objects
[17]2011Ant-colony-inspired mechanism used to grow the vision graph during runtimeGenerates a vision graph online;
Increased autonomy, robustness, and flexibility in smart
camera networks
[18]2012Approach to construct the automatic co-operative handover of multiple cameras for real-time tracking Tracking a moving target quickly
and keeping the target within
the viewing scope at all times
[11]2015Novel model with non-uniformly distributed
detection capability (DC)
Orientation of each visual sensor can be optimized through a least-squares problem;
More efficient with an averaged relative error of about 3.4%
[13]2015Node-level optimal real-time priority-based dynamic scheduling algorithmPortable system with ease of access in hard-to-access areas
[12]2017Coordination of embedded agents using spatial coordination on strategical positioning and role exchangePersistent surveillance with dynamic priorities
[14]2020Novel decomposition method with an intermediate point of representationLow computational expense;
Higher fidelity of the outcomes
[15]2020Latin-Hypercube-based Resampling Particle Swarm Optimization (LH-RPSO)LH-RPSO has higher performance than the PSO and the RPSO;
LH-RPSO is more stable and has a higher
probability of obtaining the optimal solution
Table 2. Evolution of localization techniques.
Table 2. Evolution of localization techniques.
Ref.YearMethodologyAdvantages
[26]1999Online system for tracking multiple people in an SCN with overlapping and non-overlapping viewsDevelopment of a larger, more capable, and fully automatic system without prior localization information
[27]2004Sparse overlappingBetter energy efficiency and able to cope with networking dynamics
[23]2006SLAMLocally optimal maps with computational complexity independent of the size of the map
[24]2006SLAMLocally optimal maps with computational complexity independent of the size of the map
[29]2016Estimated distribution algorithm (EDA)Accurate estimation of the features of moving objects (person)
[25]2017SFMBetter ambiguity handling in 3D environments
[28]20186D pose estimation using an end-to-end localization pipelineEfficient, highly accurate, robust in training, and exhibits outstanding generalization capabilities
Table 3. Parameter-estimation techniques.
Table 3. Parameter-estimation techniques.
Ref.YearMethodologyAdvantages
[31]2015Projection matrix obtained from non-linear
optimization
Better accuracy
[32]2016Field modelAutomatic estimation of camera parameters with high accuracy
[33]2017Greedy descent
optimization
Stable and robust automatic geometric projector camera calibration with high accuracy;
Efficient in tele-immersion applications
[34]2017Homography from
unknown planar scenes
Highly stable
[30]2018Parallel particle swarm optimization (PSO)Low time complexity and efficient
performance
[35]2019Evolutionary optimization scheme on an EDACapability of reliably converting 2D
object tracking into 3D space
Table 4. Topology-estimation techniques.
Table 4. Topology-estimation techniques.
Ref.YearMethodologyAdvantages
[36]2005Monte Carlo expectation maximization and samplingMinimum effects of noise and delay
[37]2006Window-occupancy-based methodEfficient and effective way to learn an activity topology for a large network of cameras with a limited number of data
[38]2007Exclusion algorithm in distributed clustersHigh scalability
[40]2007Statistical approach in distributed network environmentRobustness with respect to appearance changes and better estimation in a time varying network
[41]2008Decentralized data processingRobustness with respect to variable appearance and better scalability
[39]2009Activity-based multi-camera matching procedureFlexible and scalable
[42]2015Pipeline processing of lightning variationsAutomated tracking and re-identification across large camera networks
[43]2015Trust-based topology management systemHigher average coverage ratio and average packet delivery ratio
[44]2018Blind-area distance
estimation
Finer granularity and high accuracy
[45]2018Gaussian and mean cross-correlationsBetter target tracking under a single region and better interference in multi-view regions
Table 5. Task-load balancing techniques.
Table 5. Task-load balancing techniques.
Ref.YearMethodologyAdvantages
[46]2003Method for distributed adaptive task-load assignmentBetter resource efficiency
[47]2005Multiple-mobile-agent-based task-allocation frameworkSelective operation of the tracking algorithm to reduce the resource utilization
[48]2005Multiple-mobile-agent-based task-allocation frameworkSelective operation of the tracking algorithm to reduce the resource utilization
[49]2010Hierarchy-based automatic resource allotmentRobust tracking
[50]2011Expectation-maximization-based approximationEfficient approximation method for optimizing the coverage and resource allocation
[51]2012Market-based handoverImproved quality of surveillance with optimized resources
[52]2016Market-based handoverImproved quality of surveillance with optimized resources
Table 6. Occlusion-handling techniques.
Table 6. Occlusion-handling techniques.
Ref.YearMethodologyAdvantages
[53]2015Patch-match optimizationReduced computational complexity by large displacement motion
[54]2015Part-based deep modelHandles illumination changes, appearance change, abnormal deformation, and occlusions effectively
[56]2015Social force modelImproved tracking performance in the presence of complex occlusions
[55]2016K-means algorithm and statistical approachCost-effective in terms of resources (memory and computation)
[58]2017Gaussian model for occlusion handlingHandles appearance changes and is capable of dealing with complex occlusions
[57]2018CNNHigh performance with a limited labelled training dataset
[59]2018Distraction-aware tracking systemEffective and computationally efficient occlusion handling
Table 7. Object-detection techniques.
Table 7. Object-detection techniques.
Ref.YearMethodologyAdvantages
[83]1989Background subtractionLow computational complexity
[78]2001Viola and Jones technique Low processing latency with high
detection rate
[80]2005HOG-based detection Precise object detection and
classification
[79]2012Scale-invariant feature transformation Efficient detection and
localization of duplicate objects
under extreme occlusion
[81]2013Optical flowAccurate detection of moving objects
[86]2014Region proposals (R-CNNs) High accuracy and precision for object detection
[92]2015Background subtraction and mean shiftRefined and precise foreground detection
[85]2016“You only look once”Low latency multi-object detection
[89]2016Deep-neural-network-based SSDPrediction-based detection for variable shapes of objects
[93]2016Tensor flowDetection of mobile objects in FOVs
[84]2017Neural network Multi-object detection with
variable shapes
[90]2017Adaptive background subtraction modelBetter accuracy as compared to traditional background subtraction
[91]2017State-vector machine and CNN-based classifierMultiple-object-detection approach to detect ghost shadows and avoid occlusions
[82]2018Optical flowAccurate detection of moving objects
[87]2018Single-shot refinement neural network High detection accuracy
[5]2018Kernelized correlation framework Real-time occlusion handling
[88]2019Retina-Net Balanced detection performance in terms of latency, accuracy, and precision of detection
Table 8. State-of-the-art AI-based approaches addressing reconfiguration and active-vision challenges.
Table 8. State-of-the-art AI-based approaches addressing reconfiguration and active-vision challenges.
Ref.Challenge AddressedAI-Based Approach Used
[140]Camera calibrationConvolutional neural network (CNN)
[141]Neural network
[142]Parameter estimationConvolutional neural network (CNN)
[143]Deep neural network (DNN)
[144]Pose estimationNeural network
[145]Object detectionModified CNN
[146]Object trackingResidual neural network
[147]Deep CNN and Kalman filter
[148]Deep neural network (DNN)
[149]CNN and deep sort
[150]Deep-learning-based CNN
[6]Activity detectionSlow–fast CNN
[151]Neural network and strider algorithm
[152]Object re-identificationCNN
[153]Sparse graph-wavelet-based CNN
[154]Object re-identification and
occlusion handling
Deep-neural-network-based transfer learning
[155]LocalizationCNN
[156]Neural network
Table 9. Comparison of the performance for the system in [8] and the proposed ASR for surveillance dataset 1.
Table 9. Comparison of the performance for the system in [8] and the proposed ASR for surveillance dataset 1.
Pixels: 640 × 360T1T2T3T4T5T6T7
[8]TPC30,54141,55652,95656,87160,27974,63876,267
TNC18,71920,26828,51433,40039,27735,09841,905
FPC80,27174,35269,18366,41863,11660,94759,104
FNC1,00,86994,02479,74773,71167,72859,71753,124
MOTA (%)21.3826.9235.3639.1843.2147.6351.29
TRAINING CYCLES TO OBTAIN ABOVE 80% MOTA: 18
ASRTPC63,01869,21772,14176,23878,90879,51986,211
TNC33,12837,32041,16946,47348,04252,79351,775
FPC47,98243,75540,07336,11736,43131,82429,273
FNC86,27280,10877,01771,51267,01966,26463,141
MOTA (%)41.7346.2449.1853.2655.1057.3459.89
TRAINING CYCLES TO OBTAIN ABOVE 80% MOTA: 12
Table 10. Comparison of the performances of the system in [8] and the proposed ASR for surveillance dataset 2.
Table 10. Comparison of the performances of the system in [8] and the proposed ASR for surveillance dataset 2.
T1T2T3T4T5T6T7
[8]TPC23,21127,32429,84133,26636,42139,97241,101
TNC50,08058,47766,58169,51576,45177,60183,591
FPC89,23385,16182,68679,95775,27772,09869,035
FNC67,87659,43851,29247,66242,25140,72936,673
MOTA (%)31.8137.2441.8544.6148.9951.0354.12
TRAINING CYCLES TO OBTAIN ABOVE 80% MOTA: 15
ASRTPC38,21140,12841,00742,09143,10843,23643,901
TNC70,86077,65283,01784,95290,31792,88496,666
FPC77,10271,92869,98267,04166,10165,38463,687
FNC44,22740,69236,39436,31630,87428,89626,146
MOTA (%)47.3451.1253.8355.1457.9159.0861.01
TRAINING CYCLES TO OBTAIN ABOVE 80% MOTA: 11
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Shashank; Sreedevi, I. Distributed Network of Adaptive and Self-Reconfigurable Active Vision Systems. Symmetry 2022, 14, 2281. https://doi.org/10.3390/sym14112281

AMA Style

Shashank, Sreedevi I. Distributed Network of Adaptive and Self-Reconfigurable Active Vision Systems. Symmetry. 2022; 14(11):2281. https://doi.org/10.3390/sym14112281

Chicago/Turabian Style

Shashank, and Indu Sreedevi. 2022. "Distributed Network of Adaptive and Self-Reconfigurable Active Vision Systems" Symmetry 14, no. 11: 2281. https://doi.org/10.3390/sym14112281

APA Style

Shashank, & Sreedevi, I. (2022). Distributed Network of Adaptive and Self-Reconfigurable Active Vision Systems. Symmetry, 14(11), 2281. https://doi.org/10.3390/sym14112281

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop