Next Article in Journal
Identification of Driver Status Hazard Level and the System
Next Article in Special Issue
Scale-Hybrid Group Distillation with Knowledge Disentangling for Continual Semantic Segmentation
Previous Article in Journal
Selective Detection of Erythrocytes with QCMs—ABO Blood Group Typing
Previous Article in Special Issue
Small Object Detection and Tracking: A Comprehensive Review
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Deep Sensing for Compressive Video Acquisition †

1
Japan Society for the Promotion of Science, Shizuoka University, Hamamatsu 102-0083, Japan
2
Department of Systems and Control Engineering, School of Engineering, Tokyo Institute of Technology, Tokyo 152-8550, Japan
3
Faculty of Information Science and Electrical Engineering, Kyushu University, Fukuoka 819-0395, Japan
4
Institute of Datability Science, Osaka University, Suita 565-0871, Japan
*
Author to whom correspondence should be addressed.
This paper is an extension version of the conference paper: M. Yoshida, A. Torii, M. Okutomi, K. Endo, Y. Sugiyama, R. Taniguchi, and H. Nagahara. Joint optimization for compressive video sensing and reconstruction under hardware constraints. In Proceedings of European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018.
Sensors 2023, 23(17), 7535; https://doi.org/10.3390/s23177535
Submission received: 19 July 2023 / Revised: 25 August 2023 / Accepted: 25 August 2023 / Published: 30 August 2023

Abstract

:
A camera captures multidimensional information of the real world by convolving it into two dimensions using a sensing matrix. The original multidimensional information is then reconstructed from captured images. Traditionally, multidimensional information has been captured by uniform sampling, but by optimizing the sensing matrix, we can capture images more efficiently and reconstruct multidimensional information with high quality. Although compressive video sensing requires random sampling as a theoretical optimum, when designing the sensing matrix in practice, there are many hardware limitations (such as exposure and color filter patterns). Existing studies have found random sampling is not always the best solution for compressive sensing because the optimal sampling pattern is related to the scene context, and it is hard to manually design a sampling pattern and reconstruction algorithm. In this paper, we propose an end-to-end learning approach that jointly optimizes the sampling pattern as well as the reconstruction decoder. We applied this deep sensing approach to the video compressive sensing problem. We modeled the spatio–temporal sampling and color filter pattern using a convolutional neural network constrained by hardware limitations during network training. We demonstrated that the proposed method performs better than the manually designed method in gray-scale video and color video acquisitions.

1. Introduction

Digital cameras use light to capture information about a scene in the real world. However, image sensors usually consist of photodiodes arranged in two dimensions, and it is not possible to directly capture the scene information, because the information about the scene has more dimensions, such as space, time, and color. Therefore, it is necessary to map this high-dimensional information into two dimensions using a sensing matrix. Formally, image capture using a sensing matrix is expressed as follows:
y ( w , h ) = c , t C , T ϕ ( w , h , c , t ) x ( w , h , c , t ) ,
where y R W × H is the two-dimensional (2D) captured image, ϕ R W × H × C × T is the sensing matrix, ⊙ is the element-wise product, and x R W × H × C × T is the target scene (W: width of image, H: height of image, C: number of channels, and T: number of frames). The color and time information of x often are mapped to the neighbors of the pixel y using the sensing matrix ϕ . The target scene x is reconstructed from captured image y , but the captured image is less dimensional than the scene, making it an ill-posed problem. Traditional methods reconstruct the target scene by interpolation [1,2], sparse reconstruction [3,4,5,6,7], or a Gaussian mixture model (GMM) [8]; or, recently, deep neural network (DNN)-based models [9,10,11,12,13,14,15] have been used for higher speed and quality. Here, the sensing matrix and the reconstruction method are related. The sensing matrix can be described as the convolution of scene information into the captured image and the reconstruction as the deconvolution from the captured image. The task of convolving the data into features and recovering the original data by deconvolution is the same as that of an autoencoder [16]. Therefore, methods have been proposed to jointly optimize the sensing matrix and reconstruction using deep learning, as in an autoencoder [9,10,15,17,18,19,20,21]. This approach is called deep sensing, and it obtains better results than optimizing sensing and reconstruction separately. We adopt the deep sensing approach to color video sensing. While previous methods using deep sensing optimize the measurement matrix in one dimension, our proposed method optimizes the observation matrix in the color and temporal dimensions with reconstruction.
The color information is usually obtained by distributing it to neighboring pixels using a uniform color filter array (CFA). The Bayer filter [1] is often used for this color filter, and it consists of red (R), green (G), and blue (B) filters (as shown in Figure 1a). Because the RGB filter cuts out a lot of light, filters such as RGB-white (W) and cyan (C), magenta (M), and yellow (Y) filters have also been proposed to increase sensitivity [1,22,23]. When a uniform CFA is used, the spatial resolution of each color is decreased to 1 / c (c is the number of color channels). The reconstruction of color information from uniformly sampled images is called demosaicing. The existing demosaicing methods are just improvements of linear interpolation, and the reconstruction quality is limited because of the lack of captured information [24]. Therefore, compressive sensing has been used to obtain color information using a random CFA [2,25] (Figure 1b). The compressive sensing theory assumes that information is represented by a small number of coefficients (or a sparse representation vector) on a given basis. Moreover, the original signal can be recovered from a small number of observations. If the target scene is unknown, a random CFA is optimal, but if the target scene is known, there exists an optimal CFA that is better than a random one [9,10,26]. Furthermore, because a CFA is a matrix that selects the channel, it is easy to represent it in a DNN, and it is hence possible to optimize the CFA using a DNN. This approach is able to jointly optimize CFA and reconstruction using an end-to-end DNN because sensing can be represented using convolution, and reconstruction can be represented by deconvolution [9,10]. Methods for compressing and acquiring spatio–temporal information can be divided into the temporal compression method and the spatial compression method [27]. The spatial compression method [28] is similar to a single-pixel camera [29]. It acquires an image as a single observation by using a high-speed encoding sensor. On the other hand, the temporal compression method compresses spatio–temporal information and acquires the video with a single image. Traditionally, it captured spatio–temporal information by shifting the exposure times during the capture of a single frame in the neighboring pixels and sacrificing spatial resolution [30,31] (Figure 1c). This capturing method reduces the spatial resolution to 1 / T (T: number of frames). For example, the first sub-pixel is exposed at the first sub-frame, the second sub-pixel is exposed at the second sub-frame, and the last sub-pixel is exposed at the last sub-frame. Although it is possible to recover the original resolution using a super-resolution method, such methods are only an estimation using the few data captured by uniform sampling, and they do not always reconstruct the original scene. Gupta et al. [31] chose between capturing space or time information for each region but were not able to overcome the trade-off between the spatial and temporal dimensions. In contrast, compressed sensing can be used to acquire spatial and temporal information simultaneously without reducing spatial resolution or color information. By capturing coded images using spatio–temporal random exposure, the spatio–temporal information is convoluted into a single image, and the video is reconstructed in post-processing (Figure 1d). Before the appearance of deep learning, videos were reconstructed by iterative optimization such as the L0 norm minimization problem [3,4], the L1 norm minimization problem [5,6,7], a framework that combines motion estimation and the Kalman filter [32], and Gaussian mixture model-based methods [8]. In recent years, the introduction of deep learning has improved the speed and quality of reconstruction [11,12,13,14,15,33]. Furthermore, if the reconstruction is represented by a neural network, the encoder can also be represented by a neural network. Because coded imaging achieved by shifting the exposure timing can be expressed as a convolution in the time dimension and the reconstruction of video from a captured image can be expressed as deconvolution, we can use deep sensing to jointly optimize the sensing matrix and reconstruction [15,21]. However, a fully controllable complementary metal oxide semiconductor (CMOS) sensor is not realistic owing to hardware limitations [34]. Normal image sensors are designed to capture images in a uniform manner, similar to a global or rolling shutter, and it is difficult for these special sensors to achieve compatibility between sensitivity and resolution because they typically have additional complicated circuits in each pixel, and this decreases the size of the photodiode [35]. Additionally, standard commercial CMOS sensors, e.g., three-transistor CMOS sensors, do not have a per-pixel frame buffer on the chip. Thus, such sensors are incapable of multiple exposures in a non-destructive manner [3]. There is an advanced prototype CMOS sensor [4] that can control the exposure time more flexibly. However, its spatial control is limited to per-line (column and row) operations. Therefore, it is necessary to optimize the exposure patterns by considering the hardware constraints.
In color video sensing, it is necessary to optimize color filters and exposure patterns simultaneously because these encodings are dependent on each other (Figure 1e). Therefore, we propose jointly optimizing the sensing matrix and decoder for compressive color video sensing. The contributions of our paper are summarized as follows:
  • We propose a new pipeline to optimize the color filter, exposure pattern, and reconstruction decoder in compressive color video sensing using a DNN framework [36]. To the best of our knowledge, ours is the first study that considers the actual hardware sensor constraints and jointly optimizes the color filter, exposure patterns, and decoder in an end-to-end manner.
  • The proposed method is a general framework for optimizing the exposure patterns with and without hardware constraints. It is difficult to learn the sensing matrix with a DNN because DNNs can only handle differentiable functions. The optimization of the color filter is an arrangement of the RGB filter, but the exposure pattern is binary and needs to take into account the constraints of the sensor.
  • We demonstrate that the learned exposure pattern can recover high-frame-rate videos with better quality than existing handcrafted and random patterns. Moreover, we demonstrate the effectiveness of our method with images captured by an actual sensor.
This paper is an extended version of the paper [17]. In our ECCV2018 paper [17], we only optimized the exposure pattern, and we used a method of compressing three dimensions into two dimensions. In this paper, the proposed method is a more general method of compressing four dimensions into two dimensions by adding the optimization of color filters and extending it to compressed sensing in space, time, and color.
The other parts of this paper are organized as follows. Section 2 reviews the methods for deep sensing and deep optics, optimization of exposure patterns and reconstruction of video, and optimization of chromatic filter patterns and demosaicing. Section 3 describes the main contribution of this paper: deep sensing for compressive video acquisition under hardware constraints. Section 4 presents the results of the method with simulation and a prototype sensor. Section 5 offers some conclusions.

2. Related Studies

2.1. Deep Sensing and Deep Optics

In recent years, increasing numbers of studies have been published that implement sensing and reconstruction models using DNNs similar to an encoder–decoder network and jointly optimize the sensing parameters and a reconstruction decoder by end-to-end learning. Such techniques are called deep sensing/optics and have been applied in various domains. Inagaki et al. [18] proposed compressive light-field sensing using a coded aperture camera. They modeled the image acquisition and reconstruction process using a neural network and optimized the aperture patterns as well as a reconstruction decoder with end-to-end training. They were able to reconstruct the 64 × 64 sub-images of the light field from only a few images that were captured with optimized coded aperture patterns. Wu et al. [19] demonstrated a single-frame, single-view, passive 3D imaging technique using a phase mask at the camera aperture. By optimizing the phase mask and reconstruction simultaneously, they designed a phase mask with better performance than related methods. Sun et al. [20] proposed a method for training an autoencoder consisting of an encoder that selects the sampling at the sensor and a decoder for very-long-baseline interferometry (VLBI). The encoder represents the sensor selection as a binary using a fully connected Ising model [37,38]. They demonstrated that their method selects the optimal sensor in the context of a very-long-baseline interferometry array design task. Similar approaches have also been used in compressive video sensing [15,21] and compressive spectral sensing [9,10] tasks, and they provide better performance than previous methods.

2.2. Optimization of Exposure Patterns and Reconstruction of Video

Ideal compressive video sensing requires an image captured with a random exposure (Figure 2a), as expressed in Equation (1) and shown in Figure 1d. Traditionally, random exposure patterns [3,4,8,11,12,33] and model-based reconstruction methods, such as dictionary-based methods [3,4], Gaussian mixture model-based methods [8], the L1 norm minimization problem [5,6,7], and the total variation minimization problem [39], have been used. However, reconstruction in model-based methods with iterative optimization is slow, and with the development of DNNs, DNN-based methods have been mainly used instead. Iliadis et al. [11] proposed a network that is composed of fully connected layers and learns the nonlinear mapping between a video sequence and a captured image. In recent years, researchers have started using convolutional neural network (CNN)-based unrolled networks [12,13,14,15]. Ma et al. [12] proposed a decoding convolutional neural network based on the alternating direction method of multipliers (ADMM). By representing ADMM as a convolutional neural network, they achieved a fast and high-quality reconstruction. Cheng et al. [40] proposed RevSCI-net that is based on multi-group reversible 3D convolutional neural networks. They reduced training time and memory usage by using an end-to-end, multi-loop network. Sun et al. [41] used a residual ensemble network to balance reconstruction time and reconstruction quality. This network aimed to improve reconstruction quality by exploiting spatio–temporal correlations between video frames. Wen et al. [33] applied a transformer base reconstruction network to snapshot compressive video sensing. They solved the problems of CNN-based methods by using a transformer base network to acquire long-range features effectively. Very recently, an end-to-end optimization of the exposure pattern and reconstruction has been proposed because a DNN model performs not only reconstruction but also image capture. Li et al. [15] proposed a network architecture to jointly design the encoding masks and reconstruction method for compressive video sensing. Their network optimizes the full-resolution mask, and they applied Anderson acceleration to the iterative reconstruction network. Martel et al. [42] presented an end-to-end network to jointly optimize a differentiable sensor exposure model and image processing network. By using a function to decode the control signal of the sensor into an exposure pattern, they were able to optimize the shutter function of the sensor with various hardware constraints. However, their sensor model optimizes the exposure control signal for each pixel and is not applicable to sensors with spatial exposure constraints like [4].
Conventional charge-coupled device (CCD) and CMOS sensors either have a global or a rolling shutter. A global shutter exposes all of the pixels concurrently, whereas a rolling shutter exposes every pixel row/column sequentially. A commercial sensor capable of capturing an image with random exposure does not exist. Therefore, most existing studies have only evaluated simulated data [11] or optically emulated implementations [3]. However, Hitomi et al. [3] used an exposure pattern that assumed actual sensor constraints. The sensors that they consider are single-bump exposure (SBE) sensors, which can only control the start time of a single exposure for each pixel and cannot split the exposure duration into multiple exposures in one frame even though the exposure time would be controllable because of the limited exposure control line and the dynamic range, as shown in Figure 2b. There have been only a few attempts to implement sensors for compressive video sensing [34]. Sonoda et al. [4] used a prototype CMOS sensor with exposure control capabilities. The basic structure of this row- and column-wise exposure (RCE) sensor is similar to that of a standard CMOS sensor, although separate reset and transfer signals control the start and end times of the exposure. Because the pixels in a column and row share the reset and transfer signal, respectively, the exposure pattern has row- and column-wise dependency (Figure 2c). These researchers also proposed increasing the randomness of the exposure pattern. However, the method could not completely solve the pattern’s row- and column-wise dependencies.

2.3. Optimization of Chromatic Filter Patterns and Demosaicing

A color camera with a single sensor has a chromatic filter array in front of the sensor because a CMOS imager only measures the intensity of the light and cannot discriminate the spectral information. Bayer color filters [1] have been commonly used as color filters. A Bayer filter consists of two green, one blue, and one red pixel in the four-pixel neighborhood, and the pattern uniformly repeats like a checker pattern. This color sensor has only single-channel color information, and the other channels can be interpolated from the neighbor pixels. The color interpolation is usually called demosaicing, and it is well known that demosaicing does not perfectly recover the spatial resolution. Hence, color imaging with a single sensor always sacrifices spatial resolution to obtain colors.
Therefore, a method to obtain color information by compressed sensing using a random CFA was proposed [2,25]. Condat [2] proposed using a random CFA to reduce false colors and the moiré effect. A Bayer filter consists of a checkerboard pattern, and it makes false colors and the moiré effect when capturing scenes with spatial frequencies higher than the sampling frequency. Therefore, by randomizing the arrangement of the CFA, he suppressed the artifacts in scenes containing high spatial frequencies. Moreover, Hirakawa et al. [26] optimized the CFA by analyzing the spatial frequency of the scene. However, the conventional reconstruction method by interpolation also causes artifacts. Therefore, using the compressed sensing theory, Sato et al. [25] reconstructed color images from images acquired with a random CFA using ADMM. These methods deal with encoding and decoding separately, but CFA optimization is a nonlinear and difficult problem. Hence, there have been studies that jointly optimize the color filter and demosaicing model.
Chakrabarti [9] proposed a method that learns optimal color filter patterns jointly with the demosaicing network. They developed an autoencoder that selects a color channel at the encoder and reconstructs it at the decoder. The learned color filters yielded better results than the traditional Bayer filter. There are also studies that, instead of using an RGB filter pattern arrangement, jointly optimize the color filter responses for hyperspectral imaging [10] and material classification [43] from multispectral images.
Since images captured with such color filters have only one channel of information per pixel; it is necessary to complement the missing channel information in order to obtain RGB images. Traditionally, simple interpolation including bicubic interpolation and spline interpolation has been used. However, while simple interpolation is useful when computational resources are limited, it is not suitable for applications that require high-quality images. Therefore, methods for demosaicing by interpolation and refining with CNNs [44,45] and methods for demosaicing using CNNs [46,47,48] have been proposed.
Finally, we summarize the related studies in Table 1.

3. Deep Sensing for Compressive Video Acquisition under Hardware Constraints

In this section, we describe the proposed method that jointly optimizes the color filter and exposure pattern for compressive video sensing and then performs reconstruction using a DNN. The proposed DNN consists of two main parts. The first part is the sensing (encoding) layer, which optimizes the arrangement of the color filter and exposure pattern as learnable parameters of a DNN under the constraint imposed by the hardware structure, as described in Section 2.2. The second part is the reconstruction (decoding) layer, which recovers multiple channels and multiple sub-frames from a single captured image that was compressed by using the optimized color filter and exposure pattern. The overall framework is shown in Figure 3.

3.1. Sensing Layer

Although color and temporal information are encoded with ϕ in Equation (1), this process is equivalent to encoding using separate matrices. Due to implementation constraints, the actual color filters and exposure patterns are binary. Therefore, Equation (1) can be rewritten as follows:
y ( w , h ) = t T c C ϕ C ( w , h , c ) ( ϕ T ( w , h , t ) x ( w , h , c , t ) ) ,
where ϕ C [ 0 , 1 ] W × H × C is a component of the color filter of the sensing matrix, and ϕ T [ 0 , 1 ] W × H × T is a component of the exposure pattern of the sensing matrix. In addition, when considering implementation, color information is encoded by the color filter and temporal information is encoded by the exposure pattern, so they must be encoded separately. Therefore, the sensing layer consists of a color-coding layer and a time-coding layer. The sensing matrix is optimized as learnable parameters of the color filter layer and the exposure pattern layer. Usually, the learnable parameters (weight matrix) ψ of a neural network are continuous values, but the sensing matrix must be binary values. However, binary values cannot be updated because the gradient cannot be calculated. Therefore, the sensing layer has binary parameters for forward and continuous parameters for backward. The continuous parameters are gradually updated by backward, and the binary parameters are updated when the amount of updates is enough to update the binary parameters. This behavior is similar to the sigmoid function commonly used in CNNs, and the gap between binary and continuous parameters has less effect on learning. As the reconstruction layer converges in the later stages of learning, the error propagated to the sensing layer also converges, and the updates to the sensing layer also converge.
Because human vision acquires only the three RGB channels, the camera only needs to acquire RGB channels in many situations. In addition, most of the color filter is placed on the surface of the image sensor for each pixel, and the mechanism for changing it is complicated and impractical. Hence, we use Chakrabarti’s method [9] to optimize the arrangement of a static RGB CFA. Each weight matrix ψ learned by the DNN is usually a continuous value, but a CFA must select one channel. It is possible to select one channel by thresholding the weight matrix ψ of the color filter layer after applying the softmax function to the channel dimension, but the accuracy is not so high. Therefore, we use a softmax function with a parameter β that increases with the number of iterations as follows:
ϕ C ( i , j , c ) = Softmax c [ β ψ ( i , j , c ) ] .
where Softmax c is the Softmax function along the c dimension. As β increases, the output distribution of the softmax function becomes peaky and finally binary [0,1]. Therefore, we binarize the weight ψ of the CFA while updating it along the gradient by gradually increasing β at each iteration.
We sought an exposure pattern that would be capable of reconstructing video frames with high quality when trained along with the reconstruction (decoding) layer. More importantly, the compressive sensing layer must be capable of handling the exposure pattern constraints imposed by actual hardware architectures. Because the exposure is controlled by the ON/OFF switch on the sensor, the exposure pattern needs to be binary. Further, there are some constraints depending on each sensor due to the arrangement of the signal lines for controlling the exposure, as discussed in the literature ([3,4]).Therefore, we use the method inspired by Martel et al. [42] to optimize the exposure patterns. We consider the following three constraints in this paper.
1.
An ideal sensor with a binary exposure pattern but without temporal or spatial constraints (unconstrained sensor);
2.
A sensor with a binary exposure pattern and with temporal constraints (SBE sensor);
3.
A sensor with a binary exposure pattern and spatial constraints (RCE sensor).
It is difficult to train a normal neural network with all these constraints. The reason is that the exposure pattern is binary because the exposure is controlled by the electronic shutter, but the weights of the neural network must be continuous values in order to calculate the gradient in the backward propagation. Therefore, the exposure pattern optimization layer learns constrained exposure patterns by changing the weights used for forward and backward propagation. The learning steps for the exposure pattern optimization layer are as follows:
1.
Initialize the forward weights (binary weights) and the backward weights (continuous weights) with a random pattern that satisfies sensor constraints.
2.
In the forward propagation, the exposure pattern optimization layer uses the binary weights to simulate a real camera.
3.
Before backward propagation, the weights of the exposure pattern optimization layer are switched to continuous weights.
4.
The continuous weights are updated by backward propagation.
5.
Update the binary weights by using a function that considers the constraints of each sensor on the updated continuous weights.
6.
Repeat Steps 2 to 5 until the network converges.
In the unconstrained sensor, each pixel can independently control the shutter on/off in each sub-frame. Since the exposure is controlled by the electronic shutter, the exposure pattern is binary. The update function of the binary weight uses a thresholding process because there are no temporal or spatial constraints in the exposure pattern.
ψ b ( i , j , t ) = 1 ( ψ c ( i , j , t ) T h r e s h o l d ) 0 ( ψ c ( i , j , t ) < T h r e s h o l d )
where ψ b ( i , j , t ) are binary weights of the exposure layer in pixel position ( i , j ) and sub-frame t, and ψ c ( i , j , t ) are continuous weights of the exposure layer in pixel position ( i , j ) and sub-frame t. This thresholding function updates the binary weights using continuous weights.
In the SBE sensor, each pixel can control the shutter on/off independently, but this sensor does not have a non-destructive readout and can only expose once in a frame. The dynamic range of the photodiode is limited, so the exposure time must be the same for all pixels. The following function is used for the binary weight update function to consider the temporal constraint.
ψ b ( i , j , t ) = 1 ( τ t < τ + N ) 0 ( o t h e r )
τ = arg max τ [ t = τ τ + N 1 ψ c ( i , j , t ) ]
where N is the number of sub-frames to be exposed. We use Equation (5) to find the frame with the largest total of the continuous weights and Equation (6) to update the binary weights.
In the RCE sensor, the signal lines that control the exposure are shared by the vertical or horizontal columns, so the pixels to be exposed are the vertical or horizontal lines. Therefore, the binary weight update function decides whether to expose or not to expose each vertical or horizontal line by convolving the continuous weights for each line. Since there is no time constraint for the RCE sensor, this process is carried out independently for each sub-frame.
ψ b ( i , j , t ) = 1 ( ψ v ( i , t ) > 0 o r ψ h ( j , t ) > 0 ) 0 ( o t h e r )
ψ v ( i , t ) = j W ψ c ( i , j , t ) , ψ h ( j , t ) = i H ψ c ( i , j , t )
where ψ v ( i , t ) is the weight corresponding to the vertical exposure control signal, and ψ h ( j , t ) is the weight corresponding to the horizontal exposure control signal. We obtain the vertical and horizontal exposure control signals from the continuous weights using Equation (8) and update the binary weights using Equation (7) with the exposure control signals.

3.2. Reconstruction Layer

We use RevSCI-net [40] as the reconstruction layer. RevSCI-net is composed of three modules: a feature extraction layer F F , a feature-level nonlinear mapping layer F R , and a reconstruction layer F M . The feature extraction layer F F extracts high-dimensional features of the input using 3D CNN layers (the kernel size is 5 × 5 × 5 , 3 × 3 × 3 , 1 × 1 × 1 , and 3 × 3 × 3 ). The input of F F is normalized and expanded to W × H × T using the exposure pattern as follows:
y ¯ = y t = 1 T ϕ T , x ¯ = y ¯ ϕ T
where ⊘ is element-wise division, ⊙ is the element-wise product, y ¯ R W × H is the normalized input, and x ¯ R W × H × T is the extended input. The expanded input x ¯ is converted into 4D features ( R W × H × C × T ) by F F .
The nonlinear mapping layer F R , which converts the feature extraction layer F F output into video features, uses stacked reversible blocks to reduce memory usage. The original Rev-Net [49] splits the features into two channels, but in RevSCI-net, the authors improved the performance of the video reconstruction task by splitting the features into multiple channels. The ResNet block requires a lot of memory to record all the activations in all the intermediate layers when calculating the gradient by backpropagation, but by dividing the channels in this way, only the last activation needs to be recorded to calculate the gradient, thus saving a lot of memory.
The reconstruction layer F M takes the output of the nonlinear mapping layer F R as input and outputs the reconstructed video. F M is composed of 3D CNN layers (the kernel size is 3 × 3 × 3 , 3 × 3 × 3 , 1 × 1 × 1 , and 3 × 3 × 3 ). Table 2 shows the network architecture of the reconstruction layer.

4. Experiments

4.1. Experimental and Training Setup

The network was trained by minimizing the errors between the training and reconstructed videos. We used the mean squared error (MSE) as the loss function because it is directly related to the peak signal-to-noise ratio (PSNR):
L MSE = 1 T k = 1 T ( x ^ k x k ) 2
where x ^ k is the k t h reconstructed sub-frame and x k is the grand truth of the k t h sub-frame. We trained our network using the DAVIS2017 [50] dataset, which is a public benchmarking dataset designed for the task of video object segmentation. We used 6864 videos with rotation and flipping for training. We used Adam as an optimizer and set the learning rate of the decoder to 0.0002 and the batch size to 8, and we trained for 20 epochs. Experiments were run on a computer with an Intel Xeon Silver 4116 processor running at 2.10 MHz, using 100 GB of RAM and an NVIDIA Quadro GV100, and running Windows 10 Education.

4.2. Simulation Experiments

We carried out color compressive video sensing simulation experiments to evaluate joint optimization. We compared the proposed method with GMM [8], Multi-Layer Perception (MLP) [17], ADMM-Net [12], and RevSCI-net [40] to evaluate its performance. GMM [8] represents video as GMM and reconstructs videos by estimating the conditional expectation value. MLP [17] is a method we proposed at ECCV2018 that jointly optimizes the reconstruction layer (MLP), the color filter, and the exposure pattern. The MLP consists of four hidden layers, and ReLU truncates each layer. Additionally, MLP reconstructs the video in 8 × 8 pixel patches, and the final output video is the concatenation of the patches. ADMM-Net [12] uses a neural network for denoising in the alternating direction method of multipliers (ADMM). These methods [8,12,17] target monochrome videos, but we modified these methods to be able to apply them to color videos. RevSCI-net [40] uses a Bayer filter [1] and a random exposure pattern and trains only the reconstruction layer. We assumed three different hardware constraints for the SBE [3] (Figure 2b), RCE [4] (Figure 2c), and unconstrained sensors (Figure 2a). The SBE sensor has a temporal constraint in the exposure pattern, and the RCE sensor has a spatial constraint in the exposure pattern. The details of the SBE and RCE sensor constraints are described in Section 2.2. The exposure pattern for an unconstrained sensor can independently control the exposure for each pixel and achieve perfect random exposure, which is ideal in compressive sensing theory. Each parameter of the handcrafted pattern for the unconstrained sensors was randomly chosen. The network size was determined based on the size of the patch volume to be reconstructed. We used a sensor with controllable exposure [4], which exposes an 8 × 8 pixel block. For color video, the volume size of W × H × C × T was set to 256 × 256 × 3 × 16 in the experiments.
We generated a captured image simulated for the random and optimized color filter using the SBE, RCE, and unconstrained sensors. We input the simulated images to the reconstruction network to recover the video, and we quantitatively evaluated the reconstruction quality using the PSNR and structural similarity (SSIM). PSNR and SSIM were calculated in RGB components. We used nine 256 × 256 pixel videos with 16 sub-frames in the evaluation. Figure 4 shows two example results from the Aerobatics and Chameleon videos. The upper half of Figure 4 (Aerobatics) shows that GMM and MLP have blurred images due to block noise, and the lips are not sharp in ADMM-Net and RevSCI-net. On the other hand, the proposed method reconstructs the lips sharply. In the bottom half of Figure 4 (Chameleon), the chameleon’s movement is too large to be reconstructed as blurred in GMM and MLP. In addition, GMM has artifacts caused by Bayer filters. In RevSCI-net, the object is less blurred than in MLP, but it is not sharp. ADMM-Net and the proposed method reconstruct the eye of the chameleon sharply. The reconstructed video by ADMM-Net is noisy in the moving parts, and the reconstructed video by the proposed method is too smooth. Table 3 shows the average PSNRs and SSIMs of GMM [8], MLP [17], ADMM-Net [12], RevSCI-net [40], and the proposed method for the SBE, RCE, and unconstrained sensors. The SSIMs of the results of the proposed method are best for all types of hardware constraints, and the PSNRs of the proposed method are better than the other methods for the SBE and RCE sensors. In the unconstrained sensor, there is little difference in reconstruction quality between the proposed method and ADMM-Net, as the randomly sensing matrices can compress enough information, and the large search space makes convergence difficult. RevSCI-net has better reconstruction quality than GMM and MLP, but the proposed method can further improve the reconstruction quality by optimizing the color filter and exposure pattern.
We analyzed the effectiveness of the proposed method by calculating p-values for the PSNRs of the proposed and previous methods. The p-values in RevSCI-net are 6.4 × 10 3 for the SBE sensor, 2.0 × 10 4 for the RCE sensor, and 6.0 × 10 5 for the unconstrained sensor, and they are less than 0.05 for all sensor constraints. This means that the joint optimization of the sensing matrix and reconstruction improves the reconstruction quality. Similarly, the p-values of GMM (the SBE sensor: 1.7 × 10 3 , the RCE sensor: 4.6 × 10 4 , and the unconstrained sensor: 3.3 × 10 4 ) and MLP (the SBE sensor: 1.9 × 10 5 , the RCE sensor: 8.1 × 10 5 , and the unconstrained sensor: 6.1 × 10 5 ) are less than 0.05, showing that the proposed method is better. On the other hand, in ADMM, the p-values (the SBE sensor: 0.40 , the RCE sensor: 0.40 , and the unconstrained sensor: 0.46 ) are larger than 0.05, and the proposed method is not better in reconstruction quality. However, by applying the proposed joint optimization method to ADMM, the reconstruction quality of ADMM can be improved.
GMM, a traditional iterative optimization method, is very time-consuming to reconstruct. On the other hand, the reconstruction time for one video is less than one second for reconstruction using a neural network. The proposed method is about 1.5 times faster than ADMM-Net, and there is little difference between the proposed method and RevSCI-net because they use the same reconstruction model.
We show a box plot of the PSNR reconstructed with GMM, MLP, ADMM-Net, RevSCI-net, and the proposed method for each sensor constraint in Figure 5. As shown in Figure 5, the proposed method has better average quality and smaller variance than the other methods in each sensor constraint.
To quantitatively verify the contributions of the optimization of the color filter and the exposure pattern, we compared the reconstruction quality with and without optimization of the color filter and the exposure pattern (shown in Table 4). We compared four networks, such as with and without color filter optimization and with and without exposure pattern optimization, for three sensor constraints. The reconstruction quality was better for all sensor constraints when both the color filter and the exposure pattern were optimized. The reconstruction quality was also better when either the color filter or the exposure pattern was optimized than when neither was optimized. Figure 6 shows the percentage of pixels that changed in the exposure state for each epoch. We also show the exposure pattern for each epoch. The figure shows that the exposure pattern converges in the first few epochs, and the exposure state of most pixels does not change after the 10th epoch. Figure 7 compares random and optimized sensing matrices. In a Bayer filter (Figure 7a), each color is arranged alternately. On the other hand, each color is continuously arranged by optimization in the optimized filter (Figure 7b). In the case of a random exposure pattern (Figure 7c), the exposure is divided into small time slices, but with optimization (Figure 7d), the exposures are more continuous. Similar results have been reported by [21], even though our study considered the hardware constraints in pattern optimization. These results support the advantage of deep sensing, which jointly optimizes the sensing matrix and the reconstruction network.
Our method optimizes the color filters and the exposure patterns for the training data; i.e., it can generate the color filters and the exposure patterns specialized for a particular context. We used the Karlsruhe Institute of Technology and Toyota Technological Institute (KITTI) dataset [51] to evaluate the effect of optimizing for a special dataset. Optimizing the color filter, exposure pattern, and decoder for a specific dataset can improve reconstruction quality compared to optimizing in general. For comparison, we used the random color filter and the random exposure pattern (same as RevSCI-net [40]) and the generally optimized color filter and exposure pattern. We used eighty scenes for training and four scenes for testing. We cropped the images in the dataset and reshaped them to 256 × 256 . Figure 8 shows the results of reconstruction, and Table 5 shows the average PSNRs and SSIMs of each exposure pattern. The centerline in Figure 8 is reproduced as optimized for KITTI, but the edge of the centerline in the generally optimized one is not clear. In the random color filter and the random exposure pattern, the separation of the centerline cannot be seen. In this way, optimization for special context provides better reconstruction quality than optimization in general.

4.3. Experiments with the Prototype Sensor

We conducted experiments using an actual compressive image captured by the camera with the sensor reported by [4,52] to demonstrate that the optimized exposure pattern can be implemented in an actual image sensor. Figure 9 shows the camera image used in the experiment. This camera was made by Hamamatsu Photonics K.K. This sensor has no color filter, so we optimized only the exposure pattern in this experiment. The pixel size of the prototype sensor is 7.4 × 7.4 μm2. The number of total pixels is 672 × 512 , and the number of effective pixels is 656 × 496 . The size of the light-receiving area is 4.8544 × 3.6704 mm, and the dynamic range is 60 dB. The exposure is controlled by a field-programmable gate array (FPGA). The compressed video was captured at 71 frames per second (fps). We employed 16 exposure patterns per frame. Thus, the reconstructed video was equivalent to 1136 fps after recovering all of the 16 sub-frames. We used the exposure pattern obtained by the sensing layer of the proposed network after the training. Moreover, we reconstructed the video from the captured image using the reconstruction layer of the proposed network. The sensor had a rolling shutter readout and temporal exposure patterns, which were temporally shifted according to the position of the image’s row. The exposure pattern was shifted every 32 rows (four blocks with an 8 × 8 patch) in the case where the resolution of the sensor was 672 × 512 pixels and the number of exposure patterns was 16 in one frame. For example, the actual sub-exposure pattern was applied to the first four blocks as sub-exposure patterns 0–15, the second four blocks were applied as patterns 1–15 and 0, and the third four blocks were applied as patterns 2–15, 0, and 1, and so on. Hence, we trained 16 different reconstruction networks to apply various shifted exposure patterns. We used these different reconstruction networks every 32 rows in an image. Since our prototype sensor is not capable of dynamically updating exposure patterns, we fixed the sensing matrix optimized by the neural network when capturing the actual scene with the prototype sensor. Figure 10 shows a single frame of the actual captured image and three of the 16 reconstructed sub-frames. The upper row shows that dropping whitewater appears at different positions in the reconstructed sub-frames, and the motion and shape have been recovered. The bottom row shows a blinking eye. Our method successfully recovered very different appearances. Because the scenes are substantially different from the videos included in the training dataset, these results also demonstrate the generalization ability of the trained network.
Since time information is encoded, the dynamic range of the sensor is restricted as a limitation of the coded exposure. In addition, if each pixel has a different exposure time, the dynamic range is more restricted because the pixel with the longest exposure time must be captured so that it is not saturated. If some pixels are saturated, the temporal information cannot be reconstructed correctly. However, image sensors often have more bits (the sensor we used had 14 bits), whereas a standard image has 8 bits, and temporal information can be convolved into this gap.

5. Conclusions

In this paper, we first argued that real sensor architectures for developing controllable exposure have various hardware constraints that make the implementation of compressive video sensing based on completely random exposure patterns impractical. To address this issue, we proposed a general framework that consists of sensing and reconstruction layers using a DNN. Additionally, we jointly optimized the encoding and decoding models under the hardware constraints. Our proposed method is applicable to various neural-network-based video reconstruction methods and to image sensors with various control constraints. This makes the proposed method applicable to other reconstruction methods and image sensors to be proposed in the future.
We presented examples of applying the proposed framework to sensors with different constraints, i.e., the SBE, RCE, and unconstrained sensors. We demonstrated that our optimal patterns and decoding network reconstructed videos with higher quality than the videos reconstructed with the handcrafted patterns in simulation and real experiments.
Experimental results show that joint optimization improves the reconstruction quality even when the same reconstruction method is used.
In future work, we will apply the proposed method to applications other than video reconstruction. An example of another application is Time-of-Fright (ToF) imaging. Since ToF requires very high temporal resolution in the image sensor, the distance resolution can be improved by applying our proposed method.

Author Contributions

Conceptualization, M.Y., A.T. and H.N.; methodology, M.Y. and H.N.; software, M.Y. and A.T.; validation, M.Y., A.T. and H.N.; formal analysis, M.Y., A.T. and H.N.; investigation, M.Y.; resources, M.O., R.-i.T., H.N. and Y.Y.; data curation, M.Y.; writing—original draft preparation, M.Y.; writing—review and editing, A.T. and H.N.; visualization, M.Y., A.T. and H.N.; supervision, H.N.; project administration, M.O., R.-i.T., H.N. and Y.Y.; funding acquisition, H.N. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by JSPS KAKENHI grant numbers 17H06102, 18K19818, 20K20628, and 23KJ1050.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Acknowledgments

We are grateful to Yukinobu Sugiyama and Kenta Endo (Hamamatsu Photonics K.K.) for the production of the prototype image sensors and for participating in valuable discussions.

Conflicts of Interest

The authors declare no conflict of interest. The funders had no role in the design of the study; in the collection, analyses, or interpretation of data; in the writing of the manuscript; or in the decision to publish the results.

References

  1. Bayer, B.E. Color Imaging Array. U.S. Patent 3,971,065, 20 July 1976. [Google Scholar]
  2. Condat, L. A new random color filter array with good spectral properties. In Proceedings of the International Conference on Image Processing (ICIP), IEEE, Cairo, Egypt, 7–10 November 2009; pp. 1613–1616. [Google Scholar]
  3. Hitomi, Y.; Gu, J.; Gupta, M.; Mitsunaga, T.; Nayar, S.K. Video from a single coded exposure photograph using a learned over-complete dictionary. In Proceedings of the International Conference on Computer Vision (ICCV), Barcelona, Spain, 6–13 November 2011; pp. 287–294. [Google Scholar]
  4. Sonoda, T.; Nagahara, H.; Endo, K.; Sugiyama, Y.; Taniguchi, R. High-speed imaging using CMOS image sensor with quasi pixel-wise exposure. In Proceedings of the International Conference on Computational Photography (ICCP), IEEE, Evanston, IL, USA, 13–15 May 2016; pp. 1–11. [Google Scholar]
  5. Liu, Y.; Li, M.; Pados, D.A. Motion-Aware Decoding of Compressed-Sensed Video. IEEE Trans. Circuits Syst. Video Technol. 2013, 23, 438–444. [Google Scholar] [CrossRef]
  6. Azghani, M.; Karimi, M.; Marvasti, F. Multihypothesis Compressed Video Sensing Technique. IEEE Trans. Circuits Syst. Video Technol. 2016, 26, 627–635. [Google Scholar] [CrossRef]
  7. Zhao, C.; Ma, S.; Zhang, J.; Xiong, R.; Gao, W. Video Compressive Sensing Reconstruction via Reweighted Residual Sparsity. IEEE Trans. Circuits Syst. Video Technol. 2017, 27, 1182–1195. [Google Scholar] [CrossRef]
  8. Yang, J.; Yuan, X.; Liao, X.; Llull, P.; Brady, D.J.; Sapiro, G.; Carin, L. Video compressive sensing using Gaussian mixture models. IEEE Trans. Image Process. 2014, 23, 4863–4878. [Google Scholar] [CrossRef]
  9. Chakrabarti, A. Learning sensor multiplexing design through back-propagation. In Proceedings of the Advances in Neural Information Processing Systems (NeurIPS), Barcelona, Spain, 5–10 December 2016; pp. 3081–3089. [Google Scholar]
  10. Nie, S.; Gu, L.; Zheng, Y.; Lam, A.; Ono, N.; Sato, I. Deeply learned filter response functions for hyperspectral reconstruction. In Proceedings of the Conference on Computer Vision and Pattern Recognition (CVPR), Salt Lake City, UT, USA, 18 June 2018; pp. 4767–4776. [Google Scholar]
  11. Iliadis, M.; Spinoulas, L.; Katsaggelos, A.K. Deep fully-connected networks for video compressive sensing. Digit. Signal Process. 2018, 72, 9–18. [Google Scholar] [CrossRef]
  12. Ma, J.; Liu, X.; Shou, Z.; Yuan, X. Deep tensor admm-net for snapshot compressive imaging. In Proceedings of the International Conference on Computer Vision (ICCV), Long Beach, CA, USA, 15–20 June 2019; pp. 10223–10232. [Google Scholar]
  13. Yuan, X.; Liu, Y.; Suo, J.; Dai, Q. Plug-and-play algorithms for large-scale snapshot compressive imaging. In Proceedings of the Proceedings of Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 14–19 June 2020; pp. 1447–1457. [Google Scholar]
  14. Han, X.; Wu, B.; Shou, Z.; Liu, X.Y.; Zhang, Y.; Kong, L. Tensor FISTA-Net for real-time snapshot compressive imaging. In Proceedings of the AAAI Conference on Artificial Intelligence, Seattle, WA, USA, 13–19 June 2020; Volume 34, pp. 10933–10940. [Google Scholar]
  15. Li, Y.; Qi, M.; Gulve, R.; Wei, M.; Genov, R.; Kutulakos, K.N.; Heidrich, W. End-to-End Video Compressive Sensing Using Anderson-Accelerated Unrolled Networks. In Proceedings of the International Conference on Computational Photography (ICCP), IEEE, Saint Louis, MO, USA, 24–26 April 2020; pp. 137–148. [Google Scholar]
  16. Hinton, G.E.; Salakhutdinov, R.R. Reducing the dimensionality of data with neural networks. Science 2006, 313, 504–507. [Google Scholar] [CrossRef]
  17. Yoshida, M.; Torii, A.; Okutomi, M.; Endo, K.; Sugiyama, Y.; Taniguchi, R.; Nagahara, H. Joint optimization for compressive video sensing and reconstruction under hardware constraints. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 634–649. [Google Scholar]
  18. Inagaki, Y.; Kobayashi, Y.; Takahashi, K.; Fujii, T.; Nagahara, H. Learning to capture light fields through a coded aperture camera. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 418–434. [Google Scholar]
  19. Wu, Y.; Boominathan, V.; Chen, H.; Sankaranarayanan, A.; Veeraraghavan, A. PhaseCam3D—Learning Phase Masks for Passive Single View Depth Estimation. In Proceedings of the International Conference on Computational Photography (ICCP), IEEE, Tokyo, Japan, 15–17 May 2019; pp. 1–12. [Google Scholar]
  20. Sun, H.; Dalca, A.V.; Bouman, K.L. Learning a Probabilistic Strategy for Computational Imaging Sensor Selection. In Proceedings of the International Conference on Computational Photography (ICCP), IEEE, Saint Louis, MO, USA, 24–26 April 2020; pp. 81–92. [Google Scholar]
  21. Iliadis, M.; Spinoulas, L.; Katsaggelos, A.K. Deepbinarymask: Learning a binary mask for video compressive sensing. Digit. Signal Process. 2020, 96, 102591. [Google Scholar] [CrossRef]
  22. Jee, S.; Song, K.S.; Kang, M.G. Sensitivity and resolution improvement in RGBW color filter array sensor. Sensors 2018, 18, 1647–1664. [Google Scholar] [CrossRef]
  23. Choi, W.; Park, H.S.; Kyung, C.M. Color reproduction pipeline for an RGBW color filter array sensor. Opt. Express 2020, 28, 15678–15690. [Google Scholar] [CrossRef] [PubMed]
  24. Li, X.; Gunturk, B.; Zhang, L. Image demosaicing: A systematic survey. In Proceedings of the Visual Communications and Image Processing 2008, International Society for Optics and Photonics, San Jose, CA, USA, 27–31 January 2008; Volume 6822, p. 68221J. [Google Scholar]
  25. Sato, S.; Wakai, N.; Nobori, K.; Azuma, T.; Miyata, T.; Nakashizuka, M. Compressive color sensing using random complementary color filter array. In Proceedings of the International Conference on Machine Vision Applications (MVA), IEEE, Nagoya, Japan, 8–12 May 2017; pp. 43–46. [Google Scholar]
  26. Hirakawa, K.; Wolfe, P.J. Spatio-Spectral Color Filter Array Design for Optimal Image Recovery. IEEE Trans. Image Process. 2008, 17, 1876–1890. [Google Scholar] [CrossRef] [PubMed]
  27. Saideni, W.; Helbert, D.; Courreges, F.; Cances, J.P. An overview on deep learning techniques for video compressive sensing. Appl. Sci. 2022, 12, 2734. [Google Scholar] [CrossRef]
  28. Xia, K.; Pan, Z.; Mao, P. Video Compressive sensing reconstruction using unfolded LSTM. Sensors 2022, 22, 7172. [Google Scholar] [CrossRef] [PubMed]
  29. Duarte, M.F.; Davenport, M.A.; Takhar, D.; Laska, J.N.; Sun, T.; Kelly, K.F.; Baraniuk, R.G. Single-pixel imaging via compressive sampling. IEEE Signal Process. Mag. 2008, 25, 83–91. [Google Scholar] [CrossRef]
  30. Bub, G.; Tecza, M.; Helmes, M.; Lee, P.; Kohl, P. Temporal pixel multiplexing for simultaneous high-speed, high-resolution imaging. Nat. Methods 2010, 7, 209–211. [Google Scholar] [CrossRef] [PubMed]
  31. Gupta, M.; Agrawal, A.; Veeraraghavan, A.; Narasimhan, S.G. Flexible voxels for motion-aware videography. In Proceedings of the European Conference on Computer Vision (ECCV), Crete, Greece, 5–11 September 2010; pp. 100–114. [Google Scholar]
  32. Ding, X.; Chen, W.; Wassell, I.J. Compressive Sensing Reconstruction for Video: An Adaptive Approach Based on Motion Estimation. IEEE Trans. Circuits Syst. Video Technol. 2017, 27, 1406–1420. [Google Scholar] [CrossRef]
  33. Wen, J.; Huang, J.; Chen, X.; Huang, K.; Sun, Y. Transformer-Based Cascading Reconstruction Network for Video Snapshot Compressive Imaging. Appl. Sci. 2023, 13, 5922. [Google Scholar] [CrossRef]
  34. Dadkhah, M.; Deen, M.J.; Shirani, S. Compressive sensing image sensors-hardware implementation. Sensors 2013, 13, 4961–4978. [Google Scholar] [CrossRef]
  35. Wei, M.; Sarhangnejad, N.; Xia, Z.; Gusev, N.; Katic, N.; Genov, R.; Kutulakos, K.N. Coded Two-Bucket Cameras for Computer Vision. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018. [Google Scholar]
  36. LeCun, Y.; Bengio, Y.; Hinton, G. Deep learning. Nature 2015, 521, 436–444. [Google Scholar] [CrossRef]
  37. Lenz, W. Beitrag zum Verständnis der magnetischen Erscheinungen in festen Körpern. Phys. Z. 1920, 21, 613–615. [Google Scholar]
  38. Brush, S.G. History of the Lenz-Ising model. Rev. Mod. Phys. 1967, 39, 883. [Google Scholar] [CrossRef]
  39. Yuan, X. Generalized alternating projection based total variation minimization for compressive sensing. In Proceedings of the International Conference on Image Processing (ICIP), IEEE, Phoenix, AZ, USA, 25–28 September 2016; pp. 2539–2543. [Google Scholar]
  40. Cheng, Z.; Chen, B.; Liu, G.; Zhang, H.; Lu, R.; Wang, Z.; Yuan, X. Memory-efficient network for large-scale video compressive sensing. In Proceedings of the Conference on Computer Vision and Pattern Recognition (CVPR), Nashville, TN, USA, 20–25 June 2021; pp. 16246–16255. [Google Scholar]
  41. Sun, Y.; Chen, X.; Kankanhalli, M.S.; Liu, Q.; Li, J. Video Snapshot Compressive Imaging Using Residual Ensemble Network. IEEE Trans. Circuits Syst. Video Technol. 2022, 32, 5931–5943. [Google Scholar] [CrossRef]
  42. Martel, J.N.; Mueller, L.K.; Carey, S.J.; Dudek, P.; Wetzstein, G. Neural sensors: Learning pixel exposures for hdr imaging and video compressive sensing with programmable sensors. IEEE Trans. Pattern Anal. Mach. Intell. 2020, 42, 1642–1653. [Google Scholar] [CrossRef] [PubMed]
  43. Saragadam, V.; Sankaranarayanan, A.C. Programmable Spectrometry: Per-pixel Material Classification using Learned Spectral Filters. In Proceedings of the International Conference on Computational Photography (ICCP), IEEE, Saint Louis, MS, USA, 24–26 April 2020; pp. 1–10. [Google Scholar]
  44. Tan, R.; Zhang, K.; Zuo, W.; Zhang, L. Color image demosaicking via deep residual learning. In Proceedings of the IEEE International Conference on Multimedia and Expo (ICME), Hong Kong, China, 10–14 July 2017; Volume 2, p. 6. [Google Scholar]
  45. Tan, D.S.; Chen, W.Y.; Hua, K.L. DeepDemosaicking: Adaptive image demosaicking via multiple deep fully convolutional networks. IEEE Trans. Image Process. 2018, 27, 2408–2419. [Google Scholar] [CrossRef] [PubMed]
  46. Gharbi, M.; Chaurasia, G.; Paris, S.; Durand, F. Deep joint demosaicking and denoising. ACM Trans. Graph. 2016, 35, 1–12. [Google Scholar] [CrossRef]
  47. Kokkinos, F.; Lefkimmiatis, S. Deep image demosaicking using a cascade of convolutional residual denoising networks. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 303–319. [Google Scholar]
  48. Park, B.; Jeong, J. Color Filter Array Demosaicking Using Densely Connected Residual Network. IEEE Access 2019, 7, 128076–128085. [Google Scholar] [CrossRef]
  49. Gomez, A.N.; Ren, M.; Urtasun, R.; Grosse, R.B. The reversible residual network: Backpropagation without storing activations. arXiv 2017, arXiv:1707.04585. [Google Scholar]
  50. Pont-Tuset, J.; Perazzi, F.; Caelles, S.; Arbeláez, P.; Sorkine-Hornung, A.; Van Gool, L. The 2017 DAVIS Challenge on Video Object Segmentation. arXiv 2017, arXiv:1704.00675. [Google Scholar]
  51. Geiger, A.; Lenz, P.; Urtasun, R. Are we ready for Autonomous Driving? The KITTI Vision Benchmark Suite. In Proceedings of the Conference on Computer Vision and Pattern Recognition (CVPR), Providence, RI, USA, 16–21 June 2012. [Google Scholar]
  52. Hamamatsu Photonics, K.K. Imaging Device. Japan Patent JP2015-216594A, 3 December 2015. [Google Scholar]
Figure 1. Capturing images with a sensing matrix. The captured image is the sum of the element-wise product between the scene and the sensing matrix. (a) Uniform color image sensing with the Bayer filter. The spatial resolution of each color is decreased to 1 / c . (b) Compressive color sensing with a non-uniform filter. Non-uniform observations make it easy to reconstruct the original signal. (c) Uniform video sensing with a uniform exposure pattern. Neighboring pixels are considered one large pixel, and each sub-pixel is exposed at a different time, reducing the spatial resolution but increasing the temporal resolution. (d) Compressive video sensing with a non-uniform exposure pattern. Spatio–temporal information is convolved into a captured image with non-uniform sampling. (e) Compressive color video sensing with a non-uniform color filter and exposure pattern. Color and spatio–temporal information are convolved into a two-dimensional image.
Figure 1. Capturing images with a sensing matrix. The captured image is the sum of the element-wise product between the scene and the sensing matrix. (a) Uniform color image sensing with the Bayer filter. The spatial resolution of each color is decreased to 1 / c . (b) Compressive color sensing with a non-uniform filter. Non-uniform observations make it easy to reconstruct the original signal. (c) Uniform video sensing with a uniform exposure pattern. Neighboring pixels are considered one large pixel, and each sub-pixel is exposed at a different time, reducing the spatial resolution but increasing the temporal resolution. (d) Compressive video sensing with a non-uniform exposure pattern. Spatio–temporal information is convolved into a captured image with non-uniform sampling. (e) Compressive color video sensing with a non-uniform color filter and exposure pattern. Color and spatio–temporal information are convolved into a two-dimensional image.
Sensors 23 07535 g001
Figure 2. Examples of exposure patterns under hardware constraints. (a) Completely controllable sensor: it can be freely sampled in spatio–temporal dimensions. (b) SBE sensor [3]: it has temporal constraints on sampling. (c) RCE sensor [4]: it has spatial constraints on sampling.
Figure 2. Examples of exposure patterns under hardware constraints. (a) Completely controllable sensor: it can be freely sampled in spatio–temporal dimensions. (b) SBE sensor [3]: it has temporal constraints on sampling. (c) RCE sensor [4]: it has spatial constraints on sampling.
Sensors 23 07535 g002
Figure 3. Network Structure. Proposed network structure to jointly optimize the color filter and exposure pattern of compressive video sensing and perform reconstruction using a DNN. The left side represents the sensing layer that compresses video into an image using the color filter and exposure pattern. The right side represents the reconstruction layer that learns the nonlinear mapping between the compressed image and full-color video for reconstruction. ( F F : a feature extraction layer. F R : a feature-level nonlinear mapping layer. F M : a reconstruction layer).
Figure 3. Network Structure. Proposed network structure to jointly optimize the color filter and exposure pattern of compressive video sensing and perform reconstruction using a DNN. The left side represents the sensing layer that compresses video into an image using the color filter and exposure pattern. The right side represents the reconstruction layer that learns the nonlinear mapping between the compressed image and full-color video for reconstruction. ( F F : a feature extraction layer. F R : a feature-level nonlinear mapping layer. F M : a reconstruction layer).
Sensors 23 07535 g003
Figure 4. Reconstruction results. We show the results reconstructed by GMM [8], MLP [17], ADMM-Net [12], RevSCI-net [40], and the proposed method with three sensor constraints: SBE, RCE, and unconstrained. Scenes from Aerobatics of the 10th sub-frame (top) and Chameleon of the 10th sub-frame (bottom). The top left or right corners of the images show an enlarged region of the images.
Figure 4. Reconstruction results. We show the results reconstructed by GMM [8], MLP [17], ADMM-Net [12], RevSCI-net [40], and the proposed method with three sensor constraints: SBE, RCE, and unconstrained. Scenes from Aerobatics of the 10th sub-frame (top) and Chameleon of the 10th sub-frame (bottom). The top left or right corners of the images show an enlarged region of the images.
Sensors 23 07535 g004
Figure 5. Box plot of PSNR reconstructed with GMM, MLP, ADMM-Net, RevSCI-net, and the proposed method. The proposed method has the best quality and the smallest variance.
Figure 5. Box plot of PSNR reconstructed with GMM, MLP, ADMM-Net, RevSCI-net, and the proposed method. The proposed method has the best quality and the smallest variance.
Sensors 23 07535 g005
Figure 6. Percentage of changed pixels at each epoch. We also show the exposure patterns at each epoch. The exposure pattern converges in the first few epochs, and only 10% or fewer of the pixels change after the 10th epoch.
Figure 6. Percentage of changed pixels at each epoch. We also show the exposure patterns at each epoch. The exposure pattern converges in the first few epochs, and only 10% or fewer of the pixels change after the 10th epoch.
Sensors 23 07535 g006
Figure 7. Comparison of random and optimized sensing matrices. We show the top left part of 16 × 16 pixels in color filters and 8 × 8 pixels in exposure patterns for the unconstrained sensor.
Figure 7. Comparison of random and optimized sensing matrices. We show the top left part of 16 × 16 pixels in color filters and 8 × 8 pixels in exposure patterns for the unconstrained sensor.
Sensors 23 07535 g007
Figure 8. Reconstruction results of the 5th sub-frame in KITTI dataset. Reconstruction quality can be improved by tuning the color filter and exposure pattern to the specific task.
Figure 8. Reconstruction results of the 5th sub-frame in KITTI dataset. Reconstruction quality can be improved by tuning the color filter and exposure pattern to the specific task.
Sensors 23 07535 g008
Figure 9. Appearance of the prototype camera made by Hamamatsu Photonics K.K. We used this prototype camera to capture the actual scene.
Figure 9. Appearance of the prototype camera made by Hamamatsu Photonics K.K. We used this prototype camera to capture the actual scene.
Sensors 23 07535 g009
Figure 10. Reconstruction results in the actual scenes. We reconstructed 16 frames of video from the captured image: (a) shows the captured image, and (b) shows the 4th, 9th, and 15th frames of the reconstructed video. Upper row: dropping white water. Bottom row: blinking eye.
Figure 10. Reconstruction results in the actual scenes. We reconstructed 16 frames of video from the captured image: (a) shows the captured image, and (b) shows the 4th, 9th, and 15th frames of the reconstructed video. Upper row: dropping white water. Bottom row: blinking eye.
Sensors 23 07535 g010
Table 1. Summarization of the related studies.
Table 1. Summarization of the related studies.
ApplicationSensing MatrixMethod
Inagaki et al. [18]Compressive light-field sensingJoint optimizationNeural network
Wu et al. [19]Passive 3D imaging
Sun et al. [20]VLBI
Hitomi et al. [3]Compressive video sensingRandomModel-based
Sonoda et al. [4]
Liu et al. [5]
Azghsni et al. [6]
Zhao et al. [7]
Yang et al. [8]
Yuan et al. [39]
Ma et al. [12]Compressive video sensingRandomNeural network
Wen et al. [33]
Cheng et al. [40]
Sun et al. [41]
Iliadis et al. [21]Compressive video sensingJoint optimizationNeural network
Li et al. [15]
Martel et al. [42]
Condat [2]RGB imagingRandomModel-based
Sato et al. [25]
Hirakawa et al. [26]RGB imagingOptimizationModel-based
Tan et al. [44]RGB imagingRandomNeural network
Tan et al. [45]
Gharbi et al. [46]
Kokkinos et al. [47]
Park et al. [48]
Chakrabarti [9]RGB imagingJoint optimizationNeural network
Nie et al. [10]Compressive hyperspectral sensingJoint optimizationNeural network
Saragadam et al. [43]
Table 2. Network architecture of reconstruction layer.
Table 2. Network architecture of reconstruction layer.
ModuleOperationKernelStrideOutput Size
F F Conv. 5 × 5 × 5 (1, 1, 1) 16 × T × H × W
Conv. 3 × 3 × 3 (1, 1, 1) 16 × T × H × W
Conv. 1 × 1 × 1 (1, 1, 1) 32 × T × H × W
Conv. 1 × 3 × 3 (1, 2, 2) 64 × T × H / 2 × W / 2
ResNet block in F R Conv. 3 × 3 × 3 (1, 1, 1) 32 × T × H / 2 × W / 2
Conv. 3 × 3 × 3 (1, 1, 1) 64 × T × H / 2 × W / 2
F M Conv. Transpose 1 × 3 × 3 (1, 2, 2) 32 × T × H × W
Conv. 3 × 3 × 3 (1, 1, 1) 16 × T × H × W
Conv. 1 × 1 × 1 (1, 1, 1) 16 × T × H × W
Conv. 3 × 3 × 3 (1, 1, 1) C × T × H × W
Table 3. Average PSNR/SSIM of nine color video reconstructions and average reconstruction times.
Table 3. Average PSNR/SSIM of nine color video reconstructions and average reconstruction times.
GMM [8]MLP [17]ADMM-Net [12]RevSCI-Net [40]Proposed
SBE sensorPSNR17.9122.1325.0423.4825.50
SSIM0.65310.71850.81090.69960.8300
RCE sensorPSNR16.5821.8424.7623.1025.23
SSIM0.56370.70860.78220.70300.8109
UnconstrainedPSNR17.1622.3225.8923.6425.69
sensorSSIM0.57560.71180.81020.73430.8253
Time162.3 s0.2050 s0.02109 s0.01378 s0.01411 s
Table 4. Average PSNR/SSIM with and without optimization of the color filter and the exposure pattern.
Table 4. Average PSNR/SSIM with and without optimization of the color filter and the exposure pattern.
HandcraftedOptimizedHandcraftedOptimizedRandomOptimized
SBESBERCERCEUnconstrainedUnconstrained
RandomPSNR23.4824.2023.1023.5023.6324.52
filterSSIM0.69960.74430.70300.69130.73430.7776
OptimizedPSNR23.7125.5025.0225.2324.3225.69
filterSSIM0.78200.83000.80460.81090.76870.8253
Table 5. Average PSNR/SSIM of reconstructed video in KITTI dataset.
Table 5. Average PSNR/SSIM of reconstructed video in KITTI dataset.
KITTI OptimizedGenerally OptimizedRevSCI-net [40]
PSNR27.5327.0125.57
SSIM0.85940.85070.8220
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Yoshida, M.; Torii, A.; Okutomi, M.; Taniguchi, R.-i.; Nagahara, H.; Yagi, Y. Deep Sensing for Compressive Video Acquisition. Sensors 2023, 23, 7535. https://doi.org/10.3390/s23177535

AMA Style

Yoshida M, Torii A, Okutomi M, Taniguchi R-i, Nagahara H, Yagi Y. Deep Sensing for Compressive Video Acquisition. Sensors. 2023; 23(17):7535. https://doi.org/10.3390/s23177535

Chicago/Turabian Style

Yoshida, Michitaka, Akihiko Torii, Masatoshi Okutomi, Rin-ichiro Taniguchi, Hajime Nagahara, and Yasushi Yagi. 2023. "Deep Sensing for Compressive Video Acquisition" Sensors 23, no. 17: 7535. https://doi.org/10.3390/s23177535

APA Style

Yoshida, M., Torii, A., Okutomi, M., Taniguchi, R. -i., Nagahara, H., & Yagi, Y. (2023). Deep Sensing for Compressive Video Acquisition. Sensors, 23(17), 7535. https://doi.org/10.3390/s23177535

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop