Next Article in Journal
Novel Multi-IMU Tight Coupling Pedestrian Localization Exploiting Biomechanical Motion Constraints
Previous Article in Journal
Assessing the Feasibility of Augmenting Fall Detection Systems by Relying on UWB-Based Position Tracking and a Home Robot
Previous Article in Special Issue
Dynamic Camera Reconfiguration with Reinforcement Learning and Stochastic Methods for Crowd Surveillance
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Bodyprint—A Meta-Feature Based LSTM Hashing Model for Person Re-Identification

1
Department of Computer Science, Sapienza University, 00198 Rome, Italy
2
Department of Communication and Social Research, Sapienza University, 00198 Rome, Italy
3
Department of Mathematics, Computer Science and Physics, University of Udine, 33100 Udine, Italy
*
Authors to whom correspondence should be addressed.
Sensors 2020, 20(18), 5365; https://doi.org/10.3390/s20185365
Submission received: 30 June 2020 / Revised: 28 August 2020 / Accepted: 14 September 2020 / Published: 18 September 2020
(This article belongs to the Special Issue Cooperative Camera Networks)

Abstract

:
Person re-identification is concerned with matching people across disjointed camera views at different places and different time instants. This task results of great interest in computer vision, especially in video surveillance applications where the re-identification and tracking of persons are required on uncontrolled crowded spaces and after long time periods. The latter aspects are responsible for most of the current unsolved problems of person re-identification, in fact, the presence of many people in a location as well as the passing of hours or days give arise to important visual appearance changes of people, for example, clothes, lighting, and occlusions; thus making person re-identification a very hard task. In this paper, for the first time in the state-of-the-art, a meta-feature based Long Short-Term Memory (LSTM) hashing model for person re-identification is presented. Starting from 2D skeletons extracted from RGB video streams, the proposed method computes a set of novel meta-features based on movement, gait, and bone proportions. These features are analysed by a network composed of a single LSTM layer and two dense layers. The first layer is used to create a pattern of the person’s identity, then, the seconds are used to generate a bodyprint hash through binary coding. The effectiveness of the proposed method is tested on three challenging datasets, that is, iLIDS-VID, PRID 2011, and MARS. In particular, the reported results show that the proposed method, which is not based on visual appearance of people, is fully competitive with respect to other methods based on visual features. In addition, thanks to its skeleton model abstraction, the method results to be a concrete contribute to address open problems, such as long-term re-identification and severe illumination changes, which tend to heavily influence the visual appearance of persons.

1. Introduction

Last years have seen the design of increasingly advanced computer vision algorithms to support a wide range of critical tasks in a plethora of application areas. These algorithms often have the responsibility of taking determining decisions in issues where failure would lead to serious consequences. In References [1,2,3], for example, vision-based systems are used for inspection of pipeline infrastructures. In particular, in the first work, the authors presented a method for subsea pipeline corrosion estimation by using colour information of corroded pipes. To manage the degraded underwater images, the authors developed ad-hoc pre-processing algorithms for image restoration and enhancement. Differently, in the second and third work, the authors focused on large infrastructures, such as sewers and waterworks. Both works proposed an anomaly detector based on unsupervised machine learning techniques, which, unlike other competitors in this specific field, do not require annotated samples for training the detection models. Correct evaluations by the algorithms reported above can provide countless advantages in terms of maintenance and hazard determination. Even in hand and body rehabilitation, as shown in References [4,5,6], last few years have seen the proliferation of vision-based systems able to provide measurements and predictions about the recovery degree of lost skills by patients affected with strokes and degenerative diseases. The first work focused on a hand skeleton model together with pose estimation and tracking algorithms both to determine palm and fingers pose estimation and to track their movements during a rehabilitation exercise. Similarly, but using a body skeleton model, the second and third work used a customized long short term memory (LSTM) model to analyze movements and limitations of patients’ body, treating a full body rehabilitation exercise like a long action over time. Hand and body skeleton models, in addition to allowing the transposition of patients’ virtual avatars into immersive environments, also allow therapists to better detect joints that require of more exercises, thus optimizing and customizing the recovery task.
Without doubt, one of the application areas in which computer vision has had more impact, in the last ten years, is the active video surveillance, that is, those automatic systems able to replace human operators in complex tasks, such as intrusion detection [7,8], event recognition [9,10], target identification [11,12], and many others. In References [13], for example, the authors proposed a reinforcement learning approach to train a deep neural network to find optimal patrolling strategies for Unmanned Aerial Vehicles (UAVs) visual coverage tasks. Unlike other works in this application area, their method explicitly considers different coverage requirements expressed as relevance maps. References [14,15,16], instead, even if use pixel-based computer vision techniques, that is, low level processing, are able to achieve remarkable results in terms of novelty recognition and change detection, thus providing a step forward in the field of aerial monitoring. Moving to stationary and Pan–Tilt–Zoom (PTZ) cameras, very recent works, as that reported in Reference [17], shown that, even in challenging application fields, robust systems can be implemented and applied, for security contexts, in everyday life. In particular, the authors proposed a continuous-learning framework for context-aware activity recognition from unlabelled video. In their innovative approach, the authors formulated an information-theoretic active learning technique that utilizes contextual information among activities and objects. Other interesting current examples are presented in References [18,19]. In the first, a novel end-to-end partially supervised deep learning approach for video anomaly detection and localization using only normal samples is presented. Instead, in the second, the authors propose an abnormal event detection hybrid modulation method via feature expectation subgraph calibrating classification in video surveillance scenes. Both works proven to provide remarkable results in terms of robustness and accuracy.
Active surveillance systems, in real contexts, are composed of different algorithms, each of which collaborates with the others for the achievement of a specific target. For example, algorithms for separating background from foreground, for example, References [20,21,22], are often used, as pre-processing stage, both to detect the objects of interest in the scene and to have a reference model of the background and its variations over time. Another example are the tracking algorithms, for example, Reference [10,23], which are used to analyze moving objects. Among these collaborative algorithms, person re-identification ones, for example, References [24,25], play a key role, especially in security, protection, and prevention areas. In fact, being able to identify a person’s identity and being able to verify the presence of that person in other locations hours or even weeks before is a fundamental step for the areas reported above. Anyway, despite the efforts of many computer vision researchers in this application area, person re-identification task still presents several problems largely unsolved. Most of the person re-identification methods, in fact, are based on visual features extracted from images to model a person’s appearance [26,27]. This leads to a first class of problems since, as known, visual features have many weaknesses, including illumination changes, shadows, direction of light, and many others. Another class of problems regards the background clutter [28,29] and occlusions [30,31], which tend, in uncontrolled environments, to lower system performances in term of accuracy. A final class of problems, very important from a practical point of view, is referred to the long-term re-identification and camouflage [32,33]. Many systems, based totally or in part on visual features, are not able to re-identify persons under the two issues reported above, thus making the use of these systems limited in real contexts.
In this paper, a meta-feature based LSTM hashing model for person re-identification is presented. The proposed method takes inspiration from some of our recent experiences in using 2D/3D skeleton based features and LSTM models to recognize hand gestures [34], body actions [35], and body affects [36] in long video sequences. Unlike these, the 2D skeleton model, in this paper, is used to generate biometric features referred to movement, gait, and bone proportions of the body. Compared to the current literature, beyond the originality of the overall pipeline and features, the proposed method presents several novelties suitably studied to address, at least in part, the classes of problems reported above, especially in uncontrolled and crowded environments. First, the method is fully based on features extracted from 2D skeleton models present in a scene. The algorithm used to extract the models, reported in Reference [37], is proven to be highly reliable in terms of accuracy, even for multi-person estimation. The algorithm uses each input image only to produce the 2D locations of anatomical keypoints for each person in the image. This means that all the aspects, for example, illumination changes, direction of light, and many others, that can influence the re-identification by systems based on visual appearance of people, can, in the proposed method, be overlooked since the images are only used to generate 2D skeletons. Second, the proposed pipeline does not use visual features, but only features derived by analysing the 2D skeleton joints. These features, through the use of the LSTM model, are designed to catch both dynamic correlations of different body parts during movements and gaits as well as bone proportions of relevant body parts. The term meta-features was born from the fact that our features are hence conceived to generalize the body in terms of movements, gaits, and proportions. Thanks to this abstraction, long-term re-identification and some kinds of camouflages can be better managed. Third, thanks to the two final layers of the proposed network, which implement the bodyprint hash through binary coding, the representative space of our method can be considered uncountable, thus providing a tool to potentially label each and every human being in the world. The proposed method was tested on three challenging datasets designed for person-re-identification in video sequences, that is, iLIDS-VID [38], Person Re-ID (PRID) 2011 [39], and Motion Analysis and Re-identification Set (MARS) [40], showing remarkable results, compared with key works of the current literature, in terms of re-identification rank. Summarizing, the contributions of the paper with respect to both the present open issues and the current state-of-the-art in terms of pipelines and models can be outlined as follows:
-
Concerning the present person re-identification field open issues, the proposed method, which is not based on visual appearance, can be considered a concrete support in uncontrolled crowded spaces, especially in long-term re-identification where people, usually, change clothes and some aspects of their visual appearance over time. In addition, the use of RGB cameras allows the method to be used in both indoor and outdoor environments; overcoming, thanks to the use of 2D skeletons, well-known problems related to scene analysis, including illumination changes, shadows, direction of light, background clutter, and occlusions.
-
Regarding the overall pipeline and model, the approach proposed in this paper presents different novelties. First, even if some features are inspired by selected works of the current literature in recognizing hand gestures, body actions, and body affects; other features are completely new as also is their joint utilize. Second, for the first time in the literature, an LSTM hashing model for person re-identification is used. This model was conceived not only to exploit the dynamic patterns of the body movements, but also to provide, via the last two layers, a mechanism by which to provide a labelling approach for millions of people thanks to binary coding properties.
The rest of the paper is structured as follows. In Section 2, a concise literature review focused on person re-identification methods based on video sequences processing is presented. In Section 3, the entire methodology is described, including the proposed LSTM hashing model and meta-features. In Section 4, three benchmark datasets and comparative results with other literature works, are reported. Finally, Section 5 concludes the paper.

2. Related Work

In this section, selected works that treat person re-identification task on the basis of video sequences by deep learning techniques are presented and discussed. The same works are then reported in the experimental tests, that is, Section 4, for a full comparison.
A first remarkable work is reported in Reference [40], where the authors, first and foremost, introduced the Motion Analysis and Re-identification Set (MARS) dataset, a very large collection of challenging video sequences that, moreover, were also used for part of the experimental tests in the present paper. Beyond the dataset, the authors also reported an extensive evaluation of the state-of-the-art methods, including a customized one based on Convolutional Neural Networks (CNNs). The latter, supported by the Cross-view Quadratic Discriminant Analysis (XQDA) metric learning scheme [41] on iLIDS-VID and PRID 2011 datasets, and by Multiple Query (MQ), only on the MARS dataset, shown to outperform several competitive approaches, thus demonstrating a good generalization ability. The work proposed in Reference [42], instead, adopts a Recurrent Neural Network (RNN) architecture in which features are extracted from each frame by using a CNN. The latter incorporates a recurrent final layer that allows information to flow between time-steps. To provide an overall appearance feature for the complete sequence, the features, from all time-steps, are then combined by using temporal pooling. The authors conduced experiments on iLIDS-VID and PRID-2011 datasets, obtaining, on both, very significant results. Other two competitors, of the work proposed in this paper, are described in References [43,44]. In the first, the authors presented an end-to-end deep neural network architecture, which integrates a temporal attention model to selectively focus on the discriminative frames and a spatial recurrent model to exploit the contextual information when measuring the similarity. In the second, the authors described a deep architecture with jointly attentive spatial-temporal pooling, enabling a joint learning of the representations of the inputs as well as their similarity measurement. Their method extends the standard RNN-CNNs by decomposing pooling into two steps—a spatial-pooling on feature map from CNN and an attentive temporal-pooling on the output of RNN. Both works were tested on iLIDS-VID, PRID 2011, and MARS, obtaining outstanding results compared with different works of the current literature. The authors of Reference [45] proposed a method for extracting a global representation of subjects through the several frames composing a video. In particular, the method attends human body part appearance and motion simultaneously and aggregates the extracted features via the vector of locally aggregated descriptors (VLAD) [46] aggregator. By considering the adversarial learning approach, in Reference [47] the authors presented a deep few-shot adversarial learning to produce effective video representations for video-based person re-identification, using few labelled training paired videos. In detail, the method is based on Variational Recurrent Neural Networks (VRNNs) [48], which can capture temporal dynamics by mapping video sequences into latent variables. Another approach is to consider the walking cycle of the subjects, such as the one presented in Reference [49], where the authors proposed a super-pixel tracking method for extracting motion information, used to select the best walking cycle through an unsupervised method. A last competitor is reported in Reference [50], where a deep attention based Siamese model to jointly learn spatio-temporal expressive video representations and similarity metrics is presented. In their approach, the authors embed visual attention into convolutional activations from local regions to dynamically encode spatial context priors and capture the relevant patterns for the propagation through the network. Also this work was compared on the three mentioned benchmark datasets, showing remarkable results.

3. Methodology: LSTM Hashing Model and Meta-Features

In this section, the proposed architecture for person re-identification, based on skeleton meta-features derived from RGB videos and bodyprint LSTM hashing, is presented. A scheme summarizing the architecture is shown in Figure 1. In detail, starting from an RGB video, the OpenPose [37] library is first used to obtain skeleton joints positions. This representation is then fed to the feature extraction module, where meta-features analysing movements both locally (i.e., in a frame-by-frame fashion) and globally (e.g., averaging over the whole motion) are generated, fully characterizing a person. Subsequently, meta-features are given as input to the bodyprint LSTM hashing module. In this last component, local meta-features are first analysed via a single LSTM layer. The LSTM representation of local meta-features is then concatenated to global meta-features and finally fed to two dense layers, so that a bodyprint hash is generated through binary coding.

3.1. Skeleton Joint Generation

The first and foremost step, for a video-based person re-identification system using skeleton information, is a skeleton joint generation phase. At this stage, the well-known OpenPose library is exploited so that the main joints of a human skeleton can be extracted from RGB video sequences. In detail, this library leverages a multi-stage CNN so that limbs Part Affinity Fields (PAFs), a set of 2D vector fields encoding location and orientation of limbs over the image domain, are generated. By exploiting these fields, OpenPose is able to produce accurate skeletons from RGB frames and track them inside RGB videos. The extensive list of skeleton joints, representing body-foot position estimations, is depicted in Figure 2. As can be seen, up to 25 joints are described for a skeleton, where a joint is defined by its ( x , y ) coordinates inside the RGB frame. The identified joints correspond to—nose (0), neck (1), right/left shoulder, elbow, wrist (2–7), hips middle point (8), right/left hip, knee, ankle, eye, ear, big toe, small toes, and heel (9–24). While joint positions alone do not provide useful information, due to their strict correlation to the video they are extracted from, they can still be used to generate a detailed description of body movements, via the feature extraction module.

3.2. Feature Extraction

In this module, skeleton joint positions identified via OpenPose are used to generate a detailed description of full-body movements. Skeleton joint positions can be exploited to produce meaningful features able to describe, for example, gait, using the lower half of the body as shown by several gait re-identification works [51,52], or body emotions, leveraging features extracted from both lower and upper body halves [36]. Motivated by the encouraging results already obtained via hand-crafted skeleton features in References [34,35,36], two meta-feature groups built from skeleton joint positions are proposed in this work to analyze the whole body: local and global meta-features defined M F l and M F g , respectively.
Concerning the M F l set, a frame-by-frame analysis of body movements is provided via the upper and lower body openness, frontal and sideways head tilt, lower, upper, head, and composite body wobble, left and right limbs relative positions, cross-limb distance and ratio, head, arms, and legs location of activity, as well as lower, upper, and full-body convex triangulation meta-features, for a total of 21 local descriptors per frame. Through these features, gait cues and relevant changes associated to the entire body structure can be captured as they happen.
Regarding the M F g group, a condensed description of body characteristics and movements is specified via the head, arms, and legs relative movement, limp, as well as feet, legs, chest, arms, and head bones extension meta-features, for a total of 19 global descriptors. Through these features, bias towards specific body parts, which might become apparent during a walk, and detailed body conformations, describing possible limb length discrepancy, for example, can be depicted.

3.2.1. Body Openness

Body Openness ( B O ) is employed as devised in a previous study [36]. This meta-feature can describe both lower and upper body openness, defined B O l and B O u , respectively. The former, is computed using the ratio between the ankles-hips distance, and left/right knees distance; indicating whether a person has an open lower body (i.e., open legs) or has bent legs. Similarly, B O u is calculated utilising the ratio between left/right elbows distance, and the neck-hips distance, thus capturing a broaden out chest. Intuitively, low B O l values depict bent legs, corresponding to a crouched position; mid-range B O l values correspond to straight yet open legs; while high B O l values denote straight but closed legs (i.e., standing position). Comparably, low B O u values indicate a broaden out chest; mid-range B O u values correspond to straight arms and torso; while high B O u values depict open elbows and bent torso. Formally, given a video sequence S, these quantities are computed for each frame f S as follows:
B O l = d ( h i p m i d d l e , a n k l e s a v g ) d ( k n e e l e f t , k n e e r i g h t ) ,
B O u = d ( n e c k , h i p m i d d l e ) d ( e l b o w l e f t , e l b o w r i g h t ) ,
where d ( · , · ) is the Euclidean distance; h i p m i d d l e , k n e e l e f t , k n e e r i g h t , n e c k , e l b o w l e f t , and e l b o w r i g h t , correspond to OpenPose joints 8, 13, 10, 1, 6, and 3, respectively; while a n k l e s a v g indicates the average ( x , y ) ankles position, that is, OpenPose joints 11 and 14. Summarizing, B O is a local meta-feature (i.e., B O M F l ) describing 2 quantities, that is, lower and upper body half openness.

3.2.2. Head Tilt

Head Tilt ( H T ) reports the degree of frontal and sideways head tilt, defined H T f r and H T s , respectively. The former, is calculated exploiting the angle between the neck-hips and nose-neck axes. The latter, is computed using the angle between the nose-neck and left-right eye axes. Intuitively, for positive H T f r values, the head is facing downward, while for negative values the head is facing upward. Similarly, for positive H T s values the head is tilted to the right side, while for negative values there is a tilt to the left side. Formally, given a video sequence S, these measures are calculated for each frame f S as follows:
H T f r = m n o s e n e c k m n e c k h i p m i d d l e 1 + m n e c k h i p m i d d l e m n o s e n e c k ,
H T s = m e y e l e f t e y e r i g h t m n o s e n e c k 1 + m n o s e n e c k m e y e l e f t e y e r i g h t ,
where n o s e , n e c k , h i p m i d d l e , e y e l e f t , and e y e r i g h t , represent the OpenPose joints 0, 1, 8, 16, 15, respectively; while the slope m of a given axis is computed using:
m = tan θ = y 2 y 1 x 2 x 1 ,
where x and y indicate the joint coordinates used to compute the various axes. Summarizing, H T is a local meta-feature (i.e., H T M F l ) indicating 2 measures, that is, frontal and sideways head tilt.

3.2.3. Body Wobble

Body Wobble ( B W ) describes whether a person has an unsteady head, upper or lower body part, defined B W h , B W u and B W l measures, respectively. Moreover, this meta-feature is employed to also indicate the composite wobbling degree of a person, designated as B W c , by accounting for the head, upper and lower halves of the body when computing this quantity. Similarly to H T , the angles between the neck-hips axis and either the left-right eye, shoulder, or hip axes, are exploited to compute these values. Moreover, the composite body wobble is computed by averaging the absolute values of head, upper and lower body wobble, to depict the general degree of body wobble. Intuitively, B W h , B W u and B W l describe toward which direction the corresponding body part is wobbling during a motion, while B W c characterizes the movement by capturing possible peculiar walks (e.g., a drunk person usually wobbles more than a sober one). Formally, given a video sequence S, B W h , B W u and B W l are computed for each frame f S as follows:
B W h = m e y e l e f t e y e r i g h t m n e c k h i p m i d d l e 1 + m n e c k h i p m i d d l e m e y e l e f t e y e r i g h t ,
B W u = m s h o u l d e r l e f t s h o u l d e r r i g h t m n e c k h i p m i d d l e 1 + m n e c k h i p m i d d l e m s h o u l d e r l e f t s h o u l d e r r i g h t ,
B W l = m h i p l e f t h i p r i g h t m n e c k h i p m i d d l e 1 + m n e c k h i p m i d d l e m h i p l e f t h i p r i g h t ,
where e y e l e f t , e y e r i g h t , n e c k , h i p m i d d l e , s h o u l d e r l e f t , s h o u l d e r r i g h t , h i p l e f t , and h i p r i g h t , indicate the OpenPose joints 16, 15, 1, 8, 5, 2, 12, and 9, respectively. Finally, the composite body wobble B W c is derived by the other body wobble meta-features as follows:
B W c = | B W h | + | B W u | + | B W l | 3 .
Summarizing, B W is a local meta-feature (i.e., B W M F l ) denoting 4 values, that is, lower, upper, head, and composite body wobble.

3.2.4. Limbs Relative Position

Limbs Relative Position ( L R P ) indicates the opposing arm and leg relative position with respect to the neck-hips axis (i.e., a vertical axis). This meta-feature is defined for left arm/right leg and right arm/left leg pairs, named L R P l r and L R P r l , respectively. Intuitively, an arm and the opposing leg tend to be synchronised when walking, and oscillate together. Thus, by computing the difference between opposing limbs and the vertical neck-hips axis (i.e., their relative position), it is possible to describe whether the synchronous oscillation is happening or not. Formally, given a video sequence S, relative positions are computed for each frame f S as follows:
L R P l r = Δ r e l a t i v e d i s t a n c e ( n e c k , h i p s , a r m l e f t a v g , l e g r i g h t a v g ) = d i s t a n c e ( n e c k , h i p s , a r m l e f t a v g ) d i s t a n c e ( n e c k , h i p s , l e g r i g h t a v g ) = | ( n e c k y h i p s y ) a r m l e f t a v g x ( n e c k x h i p s x ) a r m l e f t a v g y + n e c k x h i p s y n e c k y h i p s x | ( n e c k y h i p s y ) 2 + ( n e c k x h i p s x ) 2 | ( n e c k y h i p s y ) l e g r i g h t a v g x ( n e c k x h i p s x ) l e g r i g h t a v g y + n e c k x h i p s y n e c k y h i p s x | ( n e c k y h i p s y ) 2 + ( n e c k x h i p s x ) 2 ,
L R P r l = Δ r e l a t i v e d i s t a n c e ( n e c k , h i p s , a r m r i g h t a v g , l e g l e f t a v g ) = d i s t a n c e ( n e c k , h i p s , a r m r i g h t a v g ) d i s t a n c e ( n e c k , h i p s , l e g l e f t a v g ) = | ( n e c k y h i p s y ) a r m r i g h t a v g x ( n e c k x h i p s x ) a r m r i g h t a v g y + n e c k x h i p s y n e c k y h i p s x | ( n e c k y h i p s y ) 2 + ( n e c k x h i p s x ) 2 | ( n e c k y h i p s y ) l e g l e f t a v g x ( n e c k x h i p s x ) l e g l e f t a v g y + n e c k x h i p s y n e c k y h i p s x | ( n e c k y h i p s y ) 2 + ( n e c k x h i p s x ) 2 ,
where n e c k and h i p s correspond to OpenPose joints 1, and 8, respectively; while a r m l e f t a v g , a r m r i g h t a v g , l e g l e f t a v g , and l e g r i g h t a v g are the average ( x , y ) positions of left arm, right arm, left leg, and right leg, computed using joints (5, 6, 7), (2, 3, 4), (12, 13, 14), and (9, 10, 11), respectively. Summarizing, L R P is a local meta-feature (i.e., L R P M F l ) depicting 2 measures, that is, left/right and right/left arm-leg relative position.

3.2.5. Cross-Limb Distance and Ratio

Cross-Limb Distance Ratio ( C L D R ) denotes the cross distance between left arm/right leg and right arm/left leg pairs, as well as their ratio, defined C L D R l r , C L D R r l , and C L D R r , respectively. Similarly to the L R P meta-feature, C L D R represents the synchronised oscillation of opposite limbs, although using only the average limbs position and their ratio instead of a reference axis. Intuitively, low C L D R r l , and C L D R r l distances indicate a synchronous cross-limb oscillation; while high values denote an irregular movement. Concerning C L D R r , the closer its value is to 1, the more synchronised is the oscillation. Indeed, through these meta-features, possible peculiar movements can be grasped during a motion. For example, arms held behind the back while walking would result in low synchronous oscillation, and could be captured via low C L D R l r , C L D R r l , and C L D R r values. Formally, given a video sequence S, cross-limb distances are computed for each frame f S as follows:
C L D R l r = d ( a r m l e f t a v g , l e g r i g h t a v g ) ,
C L D R r l = d ( a r m r i g h t a v g , l e g l e f t a v g ) ,
where d ( · , · ) is the Euclidean distance; while a r m l e f t a v g , a r m l e f t a v g , l e g l e f t a v g , and l e g l e f t a v g are the average ( x , y ) positions of left arm, right arm, left leg, and right leg, computed using joints (5, 6, 7), (2, 3, 4), (12, 13, 14), and (9, 10, 11), respectively. Finally, C L D R r , that is, the cross-limb distance ratio, is calculated using the following equation:
C L D R r = 1 1 + C L D R l r C L D R r l .
Summarizing, C L D R is a local meta-feature (i.e., C L D R M F l ) describing 3 values, that is, left/right, right/left arm-leg cross limb relative position as well as their ratio.

3.2.6. Location of Activity

Location of Activity ( L O A ) quantifies movement of each body component, and is computed for the head, left/right arms and legs, defined L O A h , L O A a l , L O A a r , L O A l l , and L O A l r , respectively. Intuitively, these meta-features analyse two consecutive frames to capture position changes, and directly measure the movement of a given component, so that they can determine which part is being moved the most during the motion. For example, a person with arms flailing about, would result in high L O A a l and L O A a r values. Formally, given a video sequence S, the set of components to be analysed C = { h , a l , a r , l l , l r } , and the set containing all joints of a given body part J c , c C , the various L O C c are computed for each pair of consecutive frames f , f S as follows:
L O A c , f , f = Δ c f , f = j J c d ( j f , j f ) | J c | ,
where c C = { h , a l , a r , l l , l r } and h , a l , a r , l l , l r correspond to h e a d , a r m l e f t , a r m r i g h t , l e g l e f t , and l e g r i g h t components, composed by joints (0, 1, 15, 16, 17, 18), (5, 6, 7), (2, 3, 4), (12, 13, 14), and (9, 10, 11), respectively; while d ( · , · ) is the Euclidean distance. Summarizing, L O A is a local meta-feature (i.e., L O A M F l ) computing 5 values, that is, head, arms, and legs location of activity.

3.2.7. Body Convex Triangulation

Body Convex Triangulation ( B C T ) depicts the center of mass distribution by analysing triangle structures built between neck and wrists joints; middle hips and ankles joints; as well as head and feet joints. These meta-features, defined B C T u , B C T l , and B C T f , represent the upper, lower, and full-body mass distribution, respectively. Intuitively, by analysing the triangle inner angles ratio difference, it is possible to describe whether a person is leaning left, right, or has a balanced mass distribution (i.e., positive, negative, and 0 values, respectively) in relation to the upper, lower, and full-body triangles observed during a motion. Formally, given a triangle with vertices A, B, and C, the corresponding angles θ α , θ β , θ γ are first computed as follows:
θ α = cos 1 a · b a b ,
θ β = cos 1 c · d c d ,
θ γ = 180 θ α θ β ,
where · is the dot product; · represents the vector magnitude; while a , b , c , and d , indicate vectors A B , A C , B A , and B C , respectively. Then, given a video sequence S, the difference between the ratios of angles adjacent to the triangle base, and the non adjacent one, is used to compute the B C T l , B C T u , and B C T f values for each frame f S , via the following equations:
B C T l = θ a n k l e l e f t θ h i p s m i d d l e θ a n k l e r i g h t θ h i p s m i d d l e ,
B C T u = θ w r i s t l e f t θ n e c k w r i s t r i g h t θ n e c k ,
B C T f = θ h e e l l e f t θ n o s e h e e l r i g h t θ n o s e ,
where a n k l e l e f t , a n k l e r i g h t , and h i p s m i d d l e , that is, OpenPose joints 14, 11, 8, correspond to the lower-body triangle; w r i s t l e f t , w r i s t r i g h t , and n e c k , that is, joints 7, 4, 1, denote the upper-body triangle; while h e e l l e f t , h e e l r i g h t , and n o s e , that is, joints 21, 24, 0, indicate the full-body triangle. Summarizing, B C T is a local meta-feature (i.e., B C T M F l ) denoting 3 quantities, that is, lower, upper, and full-body convex triangulation.

3.2.8. Relative Movement

Relative Movement ( R M ) describes the amount of change a given body part has with respect to the whole body, as conceived in Reference [36]. This meta feature is computed for the head, left/right arm, and left/right leg, defined R M h , R M a l , R M a r , R M l l , and R M l r , respectively. Intuitively, R M first analyses position and velocity changes for each body component, then computes the ratio between a single body part position (or velocity) variation amount, and the sum of position (or velocity) changes for all body components. Formally, given a video sequence S, the set of components to be analysed C = { h , a l , a r , l l , l r } , and the set containing all joints of a given body part J c , c C , the average change of a component ( A C c ), over the entire recording S, is computed as follows:
A C c = f = 0 | S | 2 j J c | Δ ( j f , j f + 1 ) | | J c | | S | 2 ,
where c C = { h , a l , a r , l l , l r } and h, a l , a r , l l , l r , correspond to h e a d , a r m l e f t , a r m r i g h t , l e g l e f t , and l e g r i g h t components, composed by joints (0, 1, 15, 16, 17, 18), (5, 6, 7), (2, 3, 4), (12, 13, 14), and (9, 10, 11), respectively. Finally, the R M c , c C , is derived using the following equation:
R M c = A C c c C A C c .
Summarizing, R M is a global meta-feature (i.e., R M M F g ) computed for both position and velocity changes of head, arms, and legs; thus resulting in 10 distinct values describing the entire recording S.

3.2.9. Limp

Limp (L) denotes whether or not a leg is moved less than the other one when walking. This meta feature is calculated using the average velocity difference between left and right legs, over the entire video sequence. Intuitively, a limping person generally has much lower velocity in either leg. Thus, this aspect can be captured by exploiting L, where a limp in the right or left leg is denoted via a negative or positive L value. Formally, given a video sequence S, the Limp L is computed as follows:
L = f = 0 | S | 2 j J l e g l e f t | v ( j ) | | S | 2 f = 0 | S | 2 j J l e g r i g h t | v ( j ) | | S | 2 ,
where J l e g l e f t and J l e g r i g h t represent the joint sets of left and right leg, defined by OpenPose joints (12, 13, 14) and (9, 10, 11), respectively; while v ( · ) is the joint velocity which can be easily computed using position variations and video frame per second (FPS). Summarizing, L is a single value global meta-feature (i.e., L M F g ) indicating limp over the whole video sequence S.

3.2.10. Bones Extension

Bones Extension ( B E ) describes left/right foot, left/right leg, chest, left/right arm, and head bones extension, defined B E f l , B E f r , B E l l , B E l r , B E c t , B E a l , B E a r , and B E h respectively. Intuitively, these meta-features provide a bone size estimation by exploiting the maximum distance between the two end-points of a bone, over the entire video sequence. Formally, given a video sequence S, the set of limb bones B and the sets of joints describing the bones J b , b B ; B E b is computed as follows:
B E b = j , j J b s . t j j max f S d ( j , j ) ,
where d ( · , · ) is the Euclidean distance; ∼ identifies adjacent joints of a given bone; while B E f l , B E f r , B E l l , B E l r , B E c t , B E a l , B E a r , and B E h denote the left foot, right foot, left leg, right leg, chest, left arm, right arm, and head, via OpenPose joint sets (19, 21), (22, 24), (12, 13, 14), (9, 10, 11), (1, 8), (5, 6, 7), (2, 3, 4), (0, 1), respectively. Summarizing, B E is global meta-feature (i.e., B E M F g ) defining bone length over feet, legs, chest, arms, and head, for a total of 8 distinct values over the entire recording S.

3.3. Bodyprint LSTM Hashing

The last step of the proposed framework, is the construction of bodyprints through the binary coding technique. In this last module, the set M F l is analysed through a LSTM so that time variations of local meta-features can be fully captured. The LSTM output is then concatenated to the set M F g , so that time-invariant body characteristics are merged together with time-varying ones. Finally, two dense layers are employed to implement the binary coding technique, so that a unique representation for a given person is built, ultimately allowing re-identification of that person.

3.3.1. LSTM

All local meta-features M F l are generated for each frame of the input video sequence containing skeleton joint positions. While the proposed features provide a detailed local description of the motion (i.e., for each specific frame), they do to not account for possible time correlation between two different frames of the input sequence. To fully exploit this information, a single layer LSTM network was chosen due to its inherent ability to handle input sequences [53]. This network leverages forget gates and peep-hole connections so that non relevant information is gradually ignored, thus improving its ability to correlate both close and distant information in a given sequence. Formally, a generic LSTM cell at time t is described by the following equations:
i t = σ ( W x i x t + W h i h t 1 + W c i c t 1 + b i ) ,
f t = σ ( W x f x t + W h f h t 1 + W c f c t 1 + b f ) ,
c t = f t c t 1 + i t tanh ( W x c x t + W h c h t 1 + W b c ) ,
o t = σ ( W x o x t + W h o h t 1 + W c o c t + b o ) ,
W t = σ t tanh ( c t ) ,
where i , f , o , c , and h represent input gate, forget gate, output gate, cell activation, and hidden vectors, respectively. Moreover, W x i , W x f , W x o , and W x c are weights connecting the various components to the input, while W c i , W c f , and W c o correspond to diagonal weights for peep-hole connections. Additionally b i , b f , b o , and b c denote the bias associated to input gate, forget gate, output gate, and cell. Finally, the Hadamard product is used for vector multiplication. To conclude, the LSTM output h T , that is, the last hidden state summarizing motion characteristics of a T-length video sequence, is concatenated to M F g , consequently producing a z vector of size | h T | + | M F g | , representing a body motion.

3.3.2. Bodyprint Hashing

The body motion characterization z can be transformed into a unique identifier via binary coding hashing, so that a bodyprint is ultimately built and used for person re-identification. Following the approach in Reference [54], Supervised Deep Hashing (SDH) with a relaxed loss function is exploited to obtain a bodyprint binary code, produced by feeding the z -vector to two dense layers. The first layer is used to merge the concatenation of global and local meta-features, while the second one is used to obtain a binary code of dimension k. Intuitively, through SDH, similar body descriptions should result in similar codes and vice-versa, ultimately enabling a bodyprint to be used in re-identification since it is associated to a specific person.
The key aspect of this SDH approach is the relaxed loss function, where the Euclidean distance between the hash of two samples is used in conjunction with a regularizer to relax the binary constraint. This relaxation is necessary due to the sign function, usually used to obtain binary codes, leading to a discontinuous, non-differentiable and non-treatable problem via back-propagation. Formally, given two input sequences S 1 and S 2 and their corresponding bodyprints b 1 , b 2 { 1 , + 1 } k , it is possible to define y = 0 in case the sequences are similar (i.e., they are derived from the same person), and y = 1 otherwise. Consequently, the relaxed loss L r with respect to the two sequences is computed as follows:
L r ( b 1 , b 2 , y ) = 1 2 ( 1 y ) b 1 b 2 2 2 + 1 2 y max ( m b 1 b 2 2 2 , 0 ) + α ( | b 1 | 1 1 + | b 2 | 1 1 ) ,
where · 1 and · 2 represent L1 and L2-norm, respectively; 1 denotes an all-ones vector; | · | indicates the element-wise absolute value operation; m > 0 is a similarity threshold; while α is a parameter modulating the regularizer weight. In detail, the first term penalises sequences generated from the same person and mapped to different bodyprints; the second term punishes sequences of different persons encoded to close binary codes, according to the threshold m; while the third term is the regularizer exploited to relax the binary constraint. Supposing there are N pairs randomly selected from the training sequences { S i , 1 , S i , 2 , y i | i = 1 , , N } , the resulting loss function to be minimised is:
L r = i = 1 N { 1 2 ( 1 y i ) b i , 1 b i , 2 2 2 + 1 2 y i max ( m b i , 1 b i , 2 2 2 , 0 ) + α ( | b i , 1 | 1 1 + | b i , 2 | 1 1 } , s . t . b i , j { 1 , + 1 } k , i { 1 , , N } , j { 1 , 2 } .
While this function can be applied via the back-propagation algorithm with mini-batch gradient descent method, the subgradients of both max and absolute value operations are non-differentiable at certain points. Thus, as described in Reference [54], the partial derivatives at those points are defined to be 1 and are computed, for the three terms of Equation (32) T 1 , T 2 , T 3 , as follows:
T 1 b i , j = ( 1 ) j + 1 ( 1 y i ) ( b i , 1 b 1 , 2 ) . T 2 b i , j = ( 1 ) j y i ( b i , 1 b i , 2 ) , b i , 1 b i , 2 2 2 < m ; 0 , o t h e r w i s e . T 3 b i , j = α δ ( b i , j ) , δ ( x ) = 1 , 1 x 0 o r x 1 ; 1 , o t h e r w i s e .
To conclude, bodyprints can be computed by applying s i g n ( b ) , and this final representation is then exploited for person re-identification as extensively shown in the experimental section.

4. Results and Discussion

In this section, the results obtained with the proposed method are discussed and compared with other state-of-the-art approaches. The comparisons are performed on three public video re-identification datasets, which are discussed below.

4.1. Datasets and Settings

Concerning the datasets, the experiments were performed on iLIDS-VID [38], PRID 2011 [39], and MARS [40]. The iLIDS-VID dataset is comprised of 600 image sequences belonging to 300 people acquired by two non-overlapping cameras. Each sequence has a number of frame ranging from 23 to 192, with an average of 73. The PRID 2011 dataset consists of 400 image sequences belonging to 200 people acquired by two adjacent cameras. Each sequence has a number of frame ranging from 5 to 675, with an average of 100. Finally, the MARS dataset consists of 1261 videos acquired by 2 to 6 cameras. The entire dataset contains 20,175 tracklets, of which 3248 are distractors. The chosen datasets are challenging for the following reasons. iLIDS-VID presents clothing similarities, occlusions, cluttered background and variations across camera views. PRID 2011 main challenges are the lighting and, as for iLIDS-VID, the variations across camera views. MARS, instead, has the above mentioned challenges besides the distractors.
Concerning the model, it was implemented in Pytorch and the following parameters were used. For the LSTM input, a tensor having a shape of [ 256 , 50 , 21 ] was employed, where 256 is the batch size, 50 are the frames used for each subject and 21 are the local meta-features M F l . For the LSTM output a tensor having a shape of [ 256 , s ] is instead utilized, where s [ 100 , 200 , 500 ] . The value of s was chosen empirically during the training of the model. As depicted in Figure 3, for iLIDS-VID and PRID 2011, that is, the dataset with the smaller number of identities, we used 200 as value for s to have better results over s = 100 and to avoid the overfitting, obtained with s = 500 . Instead, for MARS dataset s was set to 500 due to the high number of identities. A value of s higher than 500 led to a high training time with a negligible accuracy improvement. In relation to the dense layers, the first one had an input size of s + | M F g | , namely 119, 219, and 519, while the second dense layer used a dimension k [ 16 , 32 , 64 ] to create the bodyprint representation. Regarding the tensor precision, since the used meta-features comprise ratios, the tensor data type was set as a 32-bit float number. For the relaxation parameter α and similarity threshold m, 0.01 and 2 were chosen, respectively. The model was trained on a NVidia RTX 2080 GPU for 1000 epochs with a learning rate of 0.001.

4.2. Experiment Analysis

In Table 1, the results obtained with the proposed method are compared with current key works of the state-of-the-art. In particular, the comparison was performed with deep network based methods, namely, RNN [42], CNN + XQDA + MQ [40], Spatial and Temporal RNN (SPRNN) [43], Attentive Spatial-Temporal Pooling Network (ASTPN) [44], Deep Siamese Network (DSAN) [50], PersonVLAD + XQDA [45], VRNN + KissME [47], and Superpixel-Based Temporally Aligned Representation (STAR) [49]. Regarding the proposed method, 5 different versions were used for comparison. The Bodyprintlocal uses only local features, Bodyprintglobal uses only global features, while Bodyprintk, k [ 16 , 32 , 64 ] , uses both local and global features but with different size of the bodyprint hashing. By first analysing the local-only and global-only version of bodyprint, it is possible to observe that for the iLIDS-VID dataset performances are consistent with the state-of-the-art. For the PRID 2011 and MARS datasets, instead, there is a noticeable drop in the performance. This can be associated with the higher number of identities and to the fact that the proposed model was designed to use synergistically local and global features. In general, we have that the local-only version of the model performs better with respect to the global-only. This is amenable to the fact that due their granularity, the local features have a better description power, while the global features can result similar for different subjects. By considering, instead, the full bodyprint model, we have that starting from the 16-bits size hashing vector, the obtained ranking can overcome many state-of-the-art works.
Concerning the iLIDS-VID and the state-of-the-art algorithm that has the best rank-1 performance on it, that is, PersonVLAD, we have that Bodyprint16 performs 2.8 % worst with respect to it. However, with higher bodyprint vector sizes, that is, 32 and 64 bits, there are an in line performance between Bodyprint32 and PersonVLAD, and a 2.7 % improvement by using the 64 bits bodyprint version. Moreover, by considering the second best algorithm, that is, ASTPN, the gain obtained with 32 and 64 bits bodyprint vectors is 8.3% and 11.4%, respectively, which is an impressive result. For the PRID 2011 dataset, we have that Bodyprint16 and Bodyprint32 rank-1 results are in line with the other methods. In detail, Bodyprint16 is 4 % below SPRNN (i.e., the second best performing rank-1 method on this dataset) while Bodyprint32 is 0.1 % over it. For rank-5 and rank-20, we have that both Bodyprint16 and Bodyprint32 are slightly below SPRNN results. Regarding Bodyprint64, we have that it is the best algorithm at rank-5, while it is the third best result for rank-1 and rank-20, in which PersonVLAD has the best results. Finally, considering the MARS dataset, we have that at rank-1 Bodyprint16 and Bodyprint32 are in line with the state-of-the-art, while Bodyprint64 substantially outperforms other literature works. Conversely, for the rank-5, we have that Bodyprint64 is the second best method after PersonVLAD, which has obtained a score of 94.9 % . For the rank-20, instead, Bodyprint64 is in line with the other works, by obtaining a value of 95.3 % . Despite the method performing generally well, there are some limitations that can influence the accuracy. These limitations are discussed in the next section.

4.3. Limitations

The proposed model presents two major limitations: occlusions and static frames (i.e., only one frame per subject). These two situations strongly influence the feature and meta-feature extraction and computation, leading to a worse performance of the model. Regarding the occlusions, we have that for the global features the average value is lowered with respect to the number of the occluded frames. For the local features, instead, we have two cases. The first case occurs when the lower part of a subject is occluded, hence only the upper local features are available. In this case, 9 local features are available. On the contrary, for the second case we have that the upper part of the subject is occluded, allowing the extraction of the lower local features only. In this case, 5 local features are available. Despite in the first case there are 4 more local features, the performance of both cases are almost identical. Since the proposed model has been designed to consider the whole body of a subject, we have that some global features cannot be computed in case of occlusions, contributing to the lowering of the performance. Concerning the static frames, we have that all the meta-features that need more than one frame to be computer are set to 0. This means that those features lose their descriptive potential and, as for the occlusions, there is a drop in the performance. In detail, when used with static frames or in sequences with a lot of occlusions the rank-1 value is around 55 % .
A final remark must be made on the quality of the analysed images. Since the proposed method strongly relies on OpenPose framework, an important requirement is that the frames containing the subjects must not be too small in terms of spatial resolution. Otherwise, OpenPose will not extract all the skeleton parts, leading to the same problems encountered with occlusions. The same does not hold for data acquired with depth cameras, since the skeleton is directly provided and not estimated from RGB images. Anyway, considering the very good results of the proposed system and considering also that it is not based on visual features, thus overcoming a wide range of drawbacks of these systems, we can conclude that the system proposed in this paper can be considered a real contribute to the scientific community about this topic.

5. Conclusions

In this paper, a novel meta-feature based LSTM hashing model for person re-identification in RGB video sequences is presented. The proposed method is not based on visual features, but on meta-features extracted from the 2D skeleton models present in the scene. The meta-features are designed to catch movements, gaits, and bone proportions of a human body, thus providing an abstraction useful to overcome a wide range of drawbacks of the common competitors, including long-term re-identification and camouflage. The usefulness of the proposed method was tested on three benchmark dataset, that is, iLIDS-VID, PRID 2011, and MARS; thus demonstrating a step forward in the current literature.

Author Contributions

Conceptualization, D.A., A.F., D.P. and C.P.; methodology, D.A., A.F., D.P. and C.P.; software, A.F. and D.P.; validation, D.A., C.P.; writing—original draft preparation, D.A., A.F., D.P. and C.P.; writing—review and editing, D.A., L.C., A.F., G.L.F., D.P. and C.P.; supervision, L.C. and G.L.F. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Acknowledgments

This work was supported in part by the MIUR under grant “Departments of Excellence 2018-2022” of the Department of Computer Science of Sapienza University.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Khan, A.; Ali, S.S.A.; Anwer, A.; Adil, S.H.; Mériaudeau, F. Subsea Pipeline Corrosion Estimation by Restoring and Enhancing Degraded Underwater Images. IEEE Access 2018, 6, 40585–40601. [Google Scholar] [CrossRef]
  2. Piciarelli, C.; Avola, D.; Pannone, D.; Foresti, G.L. A Vision-Based System for Internal Pipeline Inspection. IEEE Trans. Ind. Inform. 2019, 15, 3289–3299. [Google Scholar] [CrossRef]
  3. Fang, X.; Guo, W.; Li, Q.; Zhu, J.; Chen, Z.; Yu, J.; Zhou, B.; Yang, H. Sewer Pipeline Fault Identification Using Anomaly Detection Algorithms on Video Sequences. IEEE Access 2020, 8, 39574–39586. [Google Scholar] [CrossRef]
  4. Placidi, G.; Avola, D.; Iacoviello, D.; Cinque, L. Overall design and implementation of the virtual glove. Comput. Biol. Med. 2013, 43, 1927–1940. [Google Scholar] [CrossRef] [PubMed]
  5. Avola, D.; Cinque, L.; Foresti, G.L.; Marini, M.R.; Pannone, D. VRheab: A fully immersive motor rehabilitation system based on recurrent neural network. Multimed. Tools Appl. 2018, 77, 24955–24982. [Google Scholar] [CrossRef]
  6. Avola, D.; Cinque, L.; Foresti, G.L.; Marini, M.R. An interactive and low-cost full body rehabilitation framework based on 3D immersive serious games. J. Biomed. Inform. 2019, 89, 81–100. [Google Scholar] [CrossRef]
  7. Cermeño, E.; Pérez, A.; Sigüenza, J.A. Intelligent video surveillance beyond robust background modeling. Expert Syst. Appl. 2018, 91, 138–149. [Google Scholar] [CrossRef]
  8. Wang, Y.; Zhu, L.; Yu, Z.; Guo, B. An Adaptive Track Segmentation Algorithm for a Railway Intrusion Detection System. Sensors 2019, 19, 2594. [Google Scholar] [CrossRef] [Green Version]
  9. Ahmad, K.; Conci, N.; De Natale, F.G.B. A saliency-based approach to event recognition. Signal Process. Image Commun. 2018, 60, 42–51. [Google Scholar] [CrossRef]
  10. Zhang, J.; Wu, C.; Wang, Y. Human Fall Detection Based on Body Posture Spatio-Temporal Evolution. Sensors 2020, 20, 946. [Google Scholar] [CrossRef] [Green Version]
  11. Avola, D.; Foresti, G.L.; Cinque, L.; Massaroni, C.; Vitale, G.; Lombardi, L. A multipurpose autonomous robot for target recognition in unknown environments. In Proceedings of the IEEE International Conference on Industrial Informatics (INDIN) 2016, Poitiers, France, 19–21 July 2016; pp. 766–771. [Google Scholar]
  12. Zhang, W.; Zhong, S.; Xu, W.; Wu, Y. Motion Correlation Discovery for Visual Tracking. IEEE Signal Process. Lett. 2018, 25, 1720–1724. [Google Scholar] [CrossRef]
  13. Piciarelli, C.; Foresti, G.L. Drone patrolling with reinforcement learning. In Proceedings of the International Conference on Distributed Smart Cameras (ICDSC) 2019, Trento, Italy, 9–11 September 2019; pp. 1–6. [Google Scholar]
  14. Avola, D.; Foresti, G.L.; Martinel, N.; Micheloni, C.; Pannone, D.; Piciarelli, C. Aerial video surveillance system for small-scale UAV environment monitoring. In Proceedings of the IEEE International Conference on Advanced Video and Signal Based Surveillance (AVSS) 2017, Lecce, Italy, 29 August–1 September 2017; pp. 1–6. [Google Scholar]
  15. Avola, D.; Cinque, L.; Fagioli, A.; Foresti, G.L.; Massaroni, C.; Pannone, D. Feature-based SLAM algorithm for small scale UAV with nadir view. In Proceedings of the International Conference on Image Analysis and Processing (ICIAP) 2019, Trento, Italy, 9–13 September 2019; pp. 457–467. [Google Scholar]
  16. Avola, D.; Cinque, L.; Foresti, G.L.; Martinel, N.; Pannone, D.; Piciarelli, C. A UAV Video Dataset for Mosaicking and Change Detection From Low-Altitude Flights. IEEE Trans. Syst. Man Cybern. Syst. 2020, 50, 2139–2149. [Google Scholar] [CrossRef] [Green Version]
  17. Hasan, M.; Paul, S.; Mourikis, A.I.; Roy-Chowdhury, A.K. Context-Aware Query Selection for Active Learning in Event Recognition. IEEE Trans. Pattern Anal. Mach. Intell. 2020, 42, 554–567. [Google Scholar] [CrossRef] [Green Version]
  18. Fan, Y.; Wen, G.; Li, D.; Qiu, S.; Levine, M.D.; Xiao, F. Video anomaly detection and localization via Gaussian Mixture Fully Convolutional Variational Autoencoder. Comput. Vis. Image Underst. 2020, 195, 1–12. [Google Scholar] [CrossRef] [Green Version]
  19. Ye, O.; Deng, J.; Yu, Z.; Liu, T.; Dong, L. Abnormal Event Detection via Feature Expectation Subgraph Calibrating Classification in Video Surveillance Scenes. IEEE Access 2020, 8, 97564–97575. [Google Scholar] [CrossRef]
  20. Avola, D.; Cinque, L.; Foresti, G.L.; Massaroni, C.; Pannone, D. A keypoint-based method for background modeling and foreground detection using a PTZ camera. Pattern Recognit. Lett. 2017, 96, 96–105. [Google Scholar] [CrossRef]
  21. Avola, D.; Bernardi, M.; Cinque, L.; Foresti, G.L.; Massaroni, C. Adaptive bootstrapping management by keypoint clustering for background initialization. Pattern Recognit. Lett. 2017, 100, 110–116. [Google Scholar] [CrossRef]
  22. Liang, D.; Pan, J.; Sun, H.; Zhou, H. Spatio-Temporal Attention Model for Foreground Detection in Cross-Scene Surveillance Videos. Sensors 2019, 19, 5142. [Google Scholar] [CrossRef] [Green Version]
  23. Ammar, S.; Bouwmans, T.; Zaghden, N.; Neji, M. Deep detector classifier (DeepDC) for moving objects segmentation and classification in video surveillance. IET Image Process. 2020, 14, 1490–1501. [Google Scholar] [CrossRef]
  24. Avola, D.; Cascio, M.; Cinque, L.; Fagioli, A.; Foresti, G.L.; Massaroni, C. Master and rookie networks for person re-identification. In Proceedings of the International Conference on Computer Analysis of Images and Patterns (CAIP) 2019, Salerno, Italy, 3–5 September 2019; pp. 470–479. [Google Scholar]
  25. Gohar, I.; Riaz, Q.; Shahzad, M.; Zeeshan Ul Hasnain Hashmi, M.; Tahir, H.; Ehsan Ul Haq, M. Person Re-Identification Using Deep Modeling of Temporally Correlated Inertial Motion Patterns. Sensors 2020, 20, 949. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  26. Almasawa, M.O.; Elrefaei, L.A.; Moria, K. A Survey on Deep Learning-Based Person Re-Identification Systems. IEEE Access 2019, 7, 175228–175247. [Google Scholar] [CrossRef]
  27. Leng, Q.; Ye, M.; Tian, Q. A Survey of Open-World Person Re-Identification. IEEE Trans. Circuits Syst. Video Technol. 2020, 30, 1092–1108. [Google Scholar] [CrossRef]
  28. Song, C.; Huang, Y.; Ouyang, W.; Wang, L. Mask-guided contrastive attention model for person re-identification. In Proceedings of the International IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) 2018, Salt Lake City, UT, USA, 18–23 June 2018; pp. 1179–1188. [Google Scholar]
  29. Zhou, S.; Wang, F.; Huang, Z.; Wang, J. Discriminative feature learning with consistent attention regularization for person re-identification. In Proceedings of the International IEEE/CVF International Conference on Computer Vision (ICCV) 2019, Seoul, Korea, 27–28 October 2019; pp. 8039–8048. [Google Scholar]
  30. Miao, J.; Wu, Y.; Liu, P.; Ding, Y.; Yang, Y. Pose-guided feature alignment for occluded person re-identification. In Proceedings of the International IEEE/CVF International Conference on Computer Vision (ICCV) 2019, Seoul, Korea, 27–28 October 2019; pp. 542–551. [Google Scholar]
  31. Hou, R.; Ma, B.; Chang, H.; Gu, X.; Shan, S.; Chen, X. VRSTC: Occlusion-free video person re-identification. In Proceedings of the International IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) 2019, Long Beach, CA, USA, 15–21 June 2019; pp. 7176–7185. [Google Scholar]
  32. Li, J.; Zhang, S.; Wang, J.; Gao, W.; Tian, Q. Global-local temporal representations for video person re-identification. In Proceedings of the International IEEE/CVF International Conference on Computer Vision (ICCV) 2019, Seoul, Korea, 27–28 October 2019; pp. 3957–3966. [Google Scholar]
  33. Huang, Y.; Xu, J.; Wu, Q.; Zhong, Y.; Zhang, P.; Zhang, Z. Beyond Scalar Neuron: Adopting Vector-Neuron Capsules for Long-Term Person Re-Identification. IEEE Trans. Circuits Syst. Video Technol. 2019. [Google Scholar] [CrossRef]
  34. Avola, D.; Bernardi, M.; Cinque, L.; Foresti, G.L.; Massaroni, C. Exploiting Recurrent Neural Networks and Leap Motion Controller for the Recognition of Sign Language and Semaphoric Hand Gestures. IEEE Trans. Multimed. 2019, 21, 234–245. [Google Scholar] [CrossRef] [Green Version]
  35. Avola, D.; Cascio, M.; Cinque, L.; Foresti, G.L.; Massaroni, C.; Rodolà, E. 2D Skeleton-Based Action Recognition via Two-Branch Stacked LSTM-RNNs. IEEE Trans. Multimed. 2019. [Google Scholar] [CrossRef]
  36. Avola, D.; Cinque, L.; Fagioli, A.; Foresti, G.L.; Massaroni, C. Deep Temporal Analysis for Non-Acted Body Affect Recognition. IEEE Trans. Affect. Comput. 2020. [Google Scholar] [CrossRef]
  37. Cao, Z.; Hidalgo Martinez, G.; Simon, T.; Wei, S.; Sheikh, Y.A. OpenPose: Realtime Multi-Person 2D Pose Estimation using Part Affinity Fields. IEEE Trans. Pattern Anal. Mach. Intell. arXiv 2018, arXiv:1812.08008. [Google Scholar] [CrossRef] [Green Version]
  38. Wang, T.; Gong, S.; Zhu, X.; Wang, S. Person re-identification by video ranking. In Proceedings of the European Conference on Computer Vision (ECCV) 2014, Zurich, Switzerland, 6–12 September 2014; pp. 688–703. [Google Scholar]
  39. Hirzer, M.; Beleznai, C.; Roth, P.M.; Bischof, H. Person re-identification by descriptive and discriminative classification. In Proceedings of the Scandinavian Conference on Image Analysis (SCIA) 2011, Ystad, Sweden, 23–27 May 2011; pp. 91–102. [Google Scholar]
  40. Zheng, L.; Bie, Z.; Sun, Y.; Wang, J.; Su, C.; Wang, S.; Tian, Q. MARS: A video benchmark for large-scale person re-identification. In Proceedings of the European Conference on Computer Vision (ECCV) 2016, Amsterdam, The Netherlands, 11–14 October 2016; pp. 868–884. [Google Scholar]
  41. Liao, S.; Hu, Y.; Zhu, X.; Li, S.Z. Person re-identification by Local Maximal Occurrence representation and metric learning. In Proceedings of the International IEEE Conference on Computer Vision and Pattern Recognition (CVPR) 2015, Boston, MA, USA, 7–12 June 2015; pp. 2197–2206. [Google Scholar]
  42. McLaughlin, N.; Martinez del Rincon, J.; Miller, P. Recurrent convolutional network for video-based person re-identification. In Proceedings of the International IEEE Conference on Computer Vision and Pattern Recognition (CVPR) 2016, Las Vegas, NV, USA, 27–30 June 2016; pp. 1325–1334. [Google Scholar]
  43. Zhou, Z.; Huang, Y.; Wang, W.; Wang, L.; Tan, T. See the forest for the trees: Joint spatial and temporal recurrent neural networks for video-based person re-identification. In Proceedings of the International IEEE Conference on Computer Vision and Pattern Recognition (CVPR) 2017, Honolulu, HI, USA, 21–26 July 2017; pp. 6776–6785. [Google Scholar]
  44. Xu, S.; Cheng, Y.; Gu, K.; Yang, Y.; Chang, S.; Zhou, P. Jointly attentive spatial-temporal pooling networks for video-based person re-identification. In Proceedings of the International IEEE/CVF International Conference on Computer Vision (ICCV) 2017, Venice, Italy, 22–29 October 2017; pp. 1–10. [Google Scholar]
  45. Wu, L.; Wang, Y.; Shao, L.; Wang, M. 3-D PersonVLAD: Learning Deep Global Representations for Video-Based Person Reidentification. IEEE Trans. Neural Netw. Learn. Syst. 2019, 30, 3347–3359. [Google Scholar] [CrossRef] [Green Version]
  46. Girdhar, R.; Ramanan, D.; Gupta, A.; Sivic, J.; Russell, B. ActionVLAD: Learning spatio-temporal aggregation for action classification. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR) 2017, Honolulu, HI, USA, 21–26 July 2017; pp. 3165–3174. [Google Scholar]
  47. Wu, L.; Wang, Y.; Yin, H.; Wang, M.; Shao, L. Few-Shot Deep Adversarial Learning for Video-Based Person Re-Identification. IEEE Trans. Image Process. 2020, 29, 1233–1245. [Google Scholar] [CrossRef] [Green Version]
  48. Chung, J.; Kastner, K.; Dinh, L.; Goel, K.; Courville, A.; Bengio, Y. A recurrent latent variable model for sequential data. In Proceedings of the International Conference on Neural Information Processing Systems (NIPS) 2015, Montreal, QC, Canada, 7–12 December 2015; pp. 2980–2988. [Google Scholar]
  49. Gao, C.; Wang, J.; Liu, L.; Yu, J.G.; Sang, N. Superpixel-Based Temporally Aligned Representation for Video-Based Person Re-Identification. Sensors 2019, 19, 3861. [Google Scholar] [CrossRef] [Green Version]
  50. Wu, L.; Wang, Y.; Gao, J.; Li, X. Where-and-When to Look: Deep Siamese Attention Networks for Video-Based Person Re-Identification. IEEE Trans. Multimed. 2019, 21, 1412–1424. [Google Scholar] [CrossRef] [Green Version]
  51. Nguyen, T.N.; Huynh, H.H.; Meunier, J. Skeleton-based abnormal gait detection. Sensors 2016, 16, 1792. [Google Scholar] [CrossRef] [Green Version]
  52. Nambiar, A.; Bernardino, A.; Nascimento, J.C. Gait-based Person Re-identification: A Survey. ACM Comput. Surv. (CSUR) 2019, 52, 1–34. [Google Scholar] [CrossRef]
  53. Hochreiter, S.; Schmidhuber, J. Long short-term memory. Neural Comput. 1997, 9, 1735–1780. [Google Scholar] [CrossRef] [PubMed]
  54. Liu, H.; Wang, R.; Shan, S.; Chen, X. Deep supervised hashing for fast image retrieval. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR) 2016, Las Vegas, NV, USA, 27–30 June 2016; pp. 2064–2072. [Google Scholar]
Figure 1. Long short term memory (LSTM) bodyprint hashing architecture scheme.
Figure 1. Long short term memory (LSTM) bodyprint hashing architecture scheme.
Sensors 20 05365 g001
Figure 2. OpenPose body-foot keypoint schematic.
Figure 2. OpenPose body-foot keypoint schematic.
Sensors 20 05365 g002
Figure 3. Train (left column) and test (right column) losses of the proposed method obtained on: iLIDS-VID, PRID 2011, and MARS datasets. Performances for these three datasets are summarized in (a,b), (c,d), and (e,f), respectively. By increasing the LSTM representational power (i.e., s size) convergence on the training set is reached much faster (i.e., (a,c,e) figures). However, due to the low number of identities in iLIDS-VID and PRID 2011 datasets, the highest s amount (i.e., 500) might result in overfitting scenarios (i.e., (b,d) figures).
Figure 3. Train (left column) and test (right column) losses of the proposed method obtained on: iLIDS-VID, PRID 2011, and MARS datasets. Performances for these three datasets are summarized in (a,b), (c,d), and (e,f), respectively. By increasing the LSTM representational power (i.e., s size) convergence on the training set is reached much faster (i.e., (a,c,e) figures). However, due to the low number of identities in iLIDS-VID and PRID 2011 datasets, the highest s amount (i.e., 500) might result in overfitting scenarios (i.e., (b,d) figures).
Sensors 20 05365 g003
Table 1. Quantitative comparison between the proposed method and the current state-of-the-art methods on the chosen datasets. The best results are highlighted in bold.
Table 1. Quantitative comparison between the proposed method and the current state-of-the-art methods on the chosen datasets. The best results are highlighted in bold.
MethodiLIDS-VIDPRID 2011MARS
Rank-1Rank-5Rank-20Rank-1Rank-5Rank-20Rank-1Rank-5Rank-20
RNN [42]588496709097---
CNN + XQDA + MQ [40]5381.495.177.393.599.368.382.689.4
SPRNN [43]55.286.59779.494.499.370.69097.6
ASTPN [44]628694779599447081
DSAN [50]61.986.898.67796.499.473.58597.5
PersonVLAD + XQDA [45]70.788.299.28896.299.782.894.999
VRNN + KissME [47]64.690.297.984.296.998.961.279.596.9
STAR [49]67.591.798.869.294.999.18089.395.1
Bodyprintlocal58.780.192.167.485.396.555.76772.3
Bodyprintglobal5678.490.665.983.295.555.464.972.2
Bodyprint1667.988.594.375.490.89876.584.490.2
Bodyprint3270.390.195.679.592.398.77787.493.9
Bodyprint6473.494.299.182.79799.286.592.695.3

Share and Cite

MDPI and ACS Style

Avola, D.; Cinque, L.; Fagioli, A.; Foresti, G.L.; Pannone, D.; Piciarelli, C. Bodyprint—A Meta-Feature Based LSTM Hashing Model for Person Re-Identification. Sensors 2020, 20, 5365. https://doi.org/10.3390/s20185365

AMA Style

Avola D, Cinque L, Fagioli A, Foresti GL, Pannone D, Piciarelli C. Bodyprint—A Meta-Feature Based LSTM Hashing Model for Person Re-Identification. Sensors. 2020; 20(18):5365. https://doi.org/10.3390/s20185365

Chicago/Turabian Style

Avola, Danilo, Luigi Cinque, Alessio Fagioli, Gian Luca Foresti, Daniele Pannone, and Claudio Piciarelli. 2020. "Bodyprint—A Meta-Feature Based LSTM Hashing Model for Person Re-Identification" Sensors 20, no. 18: 5365. https://doi.org/10.3390/s20185365

APA Style

Avola, D., Cinque, L., Fagioli, A., Foresti, G. L., Pannone, D., & Piciarelli, C. (2020). Bodyprint—A Meta-Feature Based LSTM Hashing Model for Person Re-Identification. Sensors, 20(18), 5365. https://doi.org/10.3390/s20185365

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop