Next Article in Journal
Demonstrating a Novel, Hidden Source of Spectral Distortion in X-ray Photon Counting Detectors and Assessing Novel Trigger Schemes Proposed to Avoid It
Previous Article in Journal
Multi-Harmonic Modulation in a Fiber-Optic Gyroscope
Previous Article in Special Issue
Validity of a Smartphone Application in Calculating Measures of Heart Rate Variability
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Discovering Homogeneous Groups from Geo-Tagged Videos

School of Computer Science and Engineering, Kunsan National University, 558 Daehak-ro, Gunsan 54150, Republic of Korea
*
Author to whom correspondence should be addressed.
These authors contributed equally to this work.
Sensors 2023, 23(9), 4443; https://doi.org/10.3390/s23094443
Submission received: 22 November 2022 / Revised: 24 April 2023 / Accepted: 28 April 2023 / Published: 1 May 2023
(This article belongs to the Special Issue Wearable and Mobile Sensors and Data Processing)

Abstract

:
The popularity of intelligent devices with GPS and digital compasses has generated plentiful videos and images with text tags, timestamps, and geo-references. These digital footprints of travelers record their time and spatial movements and have become indispensable information resources, vital in applications such as how groups of videographers behave and in future-movement prediction. In this paper, first we propose algorithms to discover homogeneous groups from geo-tagged videos with view directions. Second, we extend the density clustering algorithm to support fields-of-view (FoVs) in the geo-tagged videos and propose an optimization model based on a two-level grid-based index. We show the efficiency and effectiveness of the proposed homogeneous-pattern-discovery approach through experimental evaluation on real and synthetic datasets.

1. Introduction

With the development and widespread use of localization technologies, the scale and scope of trajectory collection has increased to an unprecedented level [1]. For example, Facebook, Twitter, animal trackers, and vehicles with embedded GPS generate a large amount of geo-tagged information, and large-scale geo-tagged-data analysis is beneficial to applications and services. These technologies enable collection of geo-tagged images/videos from various sources such as eye-tracking data [2,3]. In a previous study, the authors collected eye-tracking data from 1003 images of 15 viewers, recorded each observer’s look tracking path and look position, and used that as ground truth data for training a saliency model using machine learning. The term geo-tagged video refers to a video obtained using a camera device with an embedded GPS sensor to capture the video’s shooting time, location, and direction [4,5,6]. GeoUGV [4] is an extensive geo-tagged video dataset. Each video file in the dataset is accompanied by a geo-tagged metadata sequence, collected by 289 users in more than 20 cities worldwide over a period of 10 years, totaling 2397 videos. Users enable smartphones, portable cameras, or dash cams to capture videos with spatial and time information. Facebook and Twitter can obtain the photographer’s time, location, and scene. A driving recorder can record the scene of a traffic accident. This generated many geo-tagged videos and pictures to help us analyze the time, place, and scene, help us analyze the relationship between the photographers, and predict the things and routes that the photographer is interested in. For example, The Lifelog [7] is a personal record of individual daily life in varying detail for various purposes. The record contains a comprehensive dataset of a human’s activities, and the data could be used to increase knowledge about how people live their lives. In addition, the information recorded in these logs provides favorable conditions for analyzing a user’s interest in scenes and things, and even their gender and preferences.
This paper studies the problem of discovering homogeneous groups in geo-tagged videos. For example, when we get a video, we do not know the relationship between the cameramen. Maybe they are family, friends, or strangers. However, we can infer by analyzing these data. Again, this information helps to analyze famous places and who has been to the same places or taken similar routes. For example, Figure 1 shows three different field-of-view (FoV) sequences generated when a family walked on the street. When the family walks on the street, there are many shops around. The father likes to drink beer, so the attention of father focuses on the beer store. The mother likes cosmetics, so the mother focuses her attention on the cosmetics store when passing by a cosmetics store. Finally, a child is more interested in toys or ice cream, so when passing DQ (Dairy Queen), the vision of this child is easily placed in the ice cream shop. Suppose that their eyes stay on the things they are interested in for 5 s. Then, these fields of view form the FoV sequence. In this example, although the distance of family members in space meets a certain threshold, it is assumed that they also meet certain constraints in time. Nevertheless, different FoV directions produced three FoV scenes in different directions. In Figure 1, the father and child look at the same store at t 1 and t 2 . The child looks at the same store as the mother at t 3 . Furthermore, at the final timestamp direction of the FoV is the same for all three people. In this instance, we assume that they are all less than two meters apart in spatial distance and satisfy certain time constraints. These three geo-tagged videos enable us to regard the people as a group. In summary, our contributions are as follows:
  • We propose an algorithm to discover homogeneous groups from geo-tagged videos.
  • We propose a density clustering method for geo-tagged FoV clustering based on the DBSCAN method (FDBC).
  • We propose an efficient filtering algorithm to reduce the candidates using a two-level grid index (FCBG).
  • We prove the efficiency of the proposed algorithm constructs using a real-world shopping dataset [8] and a synthetic dataset which was generated from BDD100K [9].
The rest of our paper is organized as follows. Section 2 summarizes the work related to geo-tagged videos. Section 3 describes preliminary definitions for geo-tagged videos. Section 4 introduces our novel algorithms. Section 5 describes the experimental design and presents experimental observations and results. Finally, we conclude the paper.

2. Related Work

Related work can be classified as (i) travel-pattern mining, (ii) geo-tagged videos, (iii) research on clustering of moving objects.

2.1. Travel-Pattern Mining

The purpose of travel patterns is to find groups of objects that move together over a period of time, such as flocks [10,11,12], convoys [13,14], swarms [15,16], moving clustering [17], and platoons [18]. Flock and Convoy algorithms differ from other clustering methods because Flock and Convoy have stringent requirements for continuity in time. However, the Convoy algorithm based on density clustering (DBSCAN) breaks the limitation of disks in the Flock algorithm. Unlike Flock and Convoy, the Swarm algorithm is more relaxed about the continuity of time. The Platoon [18] algorithm requires consecutive partial times. In other words, two objects only move together at the beginning and end, and there is no shared moving route for a long period in the middle. For example, two cars only meet while staying at the same gas station, leading to invalid results. The constraints of the moving cluster algorithm [17] require a minimum percentage for co-moving objects between two consecutive timestamps. In order to solve these problems, a unified definition of the common movement patterns has been proposed [19,20,21,22]. Although these methods solve the problem of time and space, these co-movement patterns are all aimed at trajectory analyses. In other words, their purpose is to process only points in each partition, whereas the objects of this paper are geo-tagged videos. In the partition, the processing object is the FoVs polygon. We need to consider the constraints of the co-movement and the viewing area, viewing angle, and viewing direction of FoVs.

2.2. Geo-Tagged Videos

Some studies have tried to research effective indexing and search algorithms for FOV scenes. Grid-based indexing based on FOV scenes considered a limited radius and direction to support a k-NVS query [23]. The query was proposed to retrieve a recent video segment. MBTR (minimum-boundary-tilt rectangles) in leaf nodes effectively represent the motion scene in the index [24]. A novel geospatial image and video filtering tool (GIFT) was proposed in [25]. Ay et al. [26] proposed a new method of video query. In spatial databases, some studies [27] focus on modeling and representing the visual space of the scene. Kim et al. [28] proposed a representation model for geo-tagged videos with FoV as a vector model. However, none of the above methods enable to solve the problem of homogeneous groups.

2.3. Clustering Algorithms

Many clustering algorithms have been proposed for spatial data. K-means clustering first randomly selects K objects as the initial clustering centers, then calculates the distance between each object and each seed-cluster center and assigns each object to the cluster center closest to it [29]. The BIRCH algorithm uses a tree structure to help us cluster quickly [30]. A density-based algorithm contains two parameters (eps and minPts) to identify dense regions based on density reachability [31,32]. Density clustering algorithms are suitable for clusters of arbitrary shapes in geo-tagged photos and videos with FoVs [4,33,34,35]. In order to better avoid the lossy problem, the FoVs clustering method in our paper is based on DBSCAN.
Many density clustering algorithms use grid indexes to reduce the candidates [36,37]. The geographic space is divided into cells of the same size, each cell corresponds to a piece of storage space, and the index item registers the spatial objects that fall into the grid. For example, to query which spatial objects are in a particular spatial range, the grid of the cells is very quickly determined according to the spatial range search for the spatial object in the corresponding cell, and, finally, the spatial object obtained. Ma et al. [23] proposed a novel three-level grid-based index structure and introduced many related query types, including typical spatial queries and some based on a bounded radius and viewing-direction restriction.

3. Preliminaries

In this paper, a geo-tagged video is recorded using a camera with a GPS sensor and a magnetic compass. Let VDB = { v 1 , v 2 , , v e } be a set of the geo-tagged videos and each video can be described by an id, a set of time, and an FoV sequence v = ( i d , T , F ) . Let T = ( t 1 , t 2 , , t e ) be a sequence of all timestamps in a geo-video. Additionally, we defined d as the intervals of the partition for the timeline. T =  ( t 1 , t 2 , , t n ) is a sequence of timestamps after partition, where ( T T , t 1 t n t e ), i.e.,
T = { t | t T } .
Definition 1 (FoV Scene).
Let f denote an FoV scene at time t. The FoV scene is formally defined as:
f = ( i d , p , γ , θ , R , t ) .
Here, p describes the location of camera which consists of longitude and latitude coordinates read from the GPS sensor. γ represents the viewing angle of the camera, θ represents the view-direction value concerning the North, the visible distance is R as the maximum distance, and t is the timestamp in T [27]. The FOV scene and parameters in 2D are shown in Figure 2.
Definition 2 (FoV Sequence).
Let F  =  { ( f 1 ) , ( f 2 ) ,…, ( f n ) } be an FoV sequence in a geo-video, where f is an FoV scence ( t n T ). f n is the n-th FoV in a video. The FoV sequence is formally defined as:
F = { f v e | f n v e } .
An example of an FoV sequence is shown in Figure 3. In Figure 3, an FoV sequence represents a geo-video v 1 . It consists of each FoV scene and its corresponding timestamp, i.e., F =  { f 1 , f 2 , f 3 , f 4 , f 5 } .
Definition 3 (Overlap Coefficient).
The Overlap Coefficient ( δ ) is the threshold of the common region for two FoVs. Given two FoVs of two videos (f and f ), the common area coefficient between f and f is defined as follows:
A r e a i n t e r ( f , f ) = f f f + f f f
where δ∈ (0, 1], | f | and | f | are the areas of two FoVs, and | f f | .
Formula (4) is defined based on the Jaccard index [38]. An example is shown in Figure 2; the black region represents the intersection of the area between two FoVs (f and f in the same snapshot). If the proportion of overlapping area ( A r e a i n t e r ( f , f ) is more significant than or equal to the threshold δ , the two FoVs satisfy the intersection-of-area constraint.
Definition 4 (Closeness).
Given two FoVs f and f in a snapshot s t . Closeness is if the distance between two FoVs is less than the maximum distance threshold ϵ; moreover, it is satisfied if the intersection of the viewable area of the two FoVs is greater than the minimum area intersection threshold A r e a i n t e r , as shown in Formula (4). The closeness of FoVs is defined as
{ f | f s t , A r e a i n t e r ( f , f ) δ d i s t ( f p , f p ) ε } ,
where d i s t ( f p , f p ) denotes the minimal distance between the positions of f and f .
Definition 5 (Clusters of FoVs).
Let SC  = { S C 1 , S C 2 ,…, S C t } be the collection of FoVs from snapshots at the timestamp in { t 1 , t 2 , , t n } , where S C t = { C 1 , C 2 , , C n } denotes a set of clusters in a snapshot S t , where C n is an FoV cluster. If the number of FoVs in S C t that are satisfying Formula (5) is at least m i n o , then these FoVs will form a cluster, where m i n o is the minimal numbers of FoVs.
For example, the FoVs satisfy the distance threshold as Figure 4 shows. If m i n o  = 2, we can obtain five clusters { ( f v 1 , f v 2 ) , ( ( f v 3 , f v 4 ) , ( f v 5 , f v 6 ) , ( f v 7 , f v 8 , f v 9 ) , ( f v 10 , f v 11 ) } in the snapshot at t 1 . However, we only can capture one cluster if we change to m i n o  = 3, i.e.,  ( f v 7 , f v 8 f v 9 ) .
Let the time segment T i be the i-th time segment of T, i.e., ( T i T ). If  T [ n ] that satisfies T n + 1 = T n + 1 , ( 1 i n ), then T n is the FoV sequential segment, where n is the timestamp of the n-th snapshot. When T is composed of s sequential time segments, T is s-succession. Further, if every adjacent FoV’s time, such as ( T n , T [ n + 1 ] T ), to satisfy the difference value is no bigger than g, i.e., ( 1 n < T , T n + 1 T [ n ] g ), then T is g-connection. Additionally, the groups also need another constraint k; it is the minimal number of FoVs belonging to the same cluster.
For example, given two times segments T 1  = (5, 6, 7), T 2  = (5, 6, 8), and T =  T 1 , T 2 , T 1 is an FoV sequential segment, but T 2 is not because the 7 is lost in T 2 . In the above demonstrate, given a segment T = (5, 6, 8, 9, 10), Segment T is two-sequential and two-connection. The minimal length segment (5, 6) of T is 2, therefore, the two-sequential FoV segments are T 1  = (5, 6) and T 2  = (8, 9, 10) in T.
Definition 6 (Homogeneous Groups).
Let P  =  { P 1 , P 2 ,…, P n } be the homogeneous groups of geo-videos required to satisfy five constraints as follows:
 (1)
FC (FoVs Cluster): the FoVs in a set of geo-tagged videos belong to the same cluster at each timestamp of T.
 (2)
Occurrence: k is the minimal number of occurrence for FoVs belong to the same cluster.
 (3)
Significance:  | C |  ≥  m i n o .
 (4)
Succession:  T is s-sequential.
 (5)
Connection:  T is g-connected.
Consider the example in Figure 4; a dotted circle denotes a cluster. Given m i n o  = 2, k = 3, s = 2, g = 2 and a r e a i n t e r  = 0.5, then we are able to capture three homogeneous groups P  =  { { v 3 , v 4 } , { v 5 , v 6 } , { v 7 , v 8 } } for T = < t 1 , t 2 , t 4 , t 5 >. In other words, { v 3 , v 4 } , { v 5 , v 6 } , and { v 7 , v 8 } belong to the same cluster at timestamps 1, 2, 4, and 5.
Definition 7 (Grid).
Let key be the ID of the global grid cells. Let G r  =  { g 1 , g 2 ,…, g n } be an integer grid. Then ( m i n x , m i n y , m a x x , m a x y ) are the boundaries in a snapshot S t (x represents longitude, y represents latitude). For each empty cell, g = ( k e y , f a l s e , f). Then, let e be an edge of one FoV, S e g be the set of segments of an edge, S I D be a set of cells’ ID of segments in an edge, and  K e y s be the cells’ ID covered by f. G r is the index set of the finally obtained FoVs. For each k e y in K e y s , if FoV f be inserted into the cell g, g = ( k e y , t r u e , f) is the grid cell of G r .
Definition 8 (Grid-clustering).
Let SC  =  { S C 1 , S C 2 ,…, S C t } be the clusters of FoVs in snapshots S t based on the grid index. Let f M B R be the minimum-bounding rectangle (MBR) of the FoV f. We use V r to represent a collection of all grid cell IDs covered by MBR, such that
R a n g e C ( f M B R , V r ) = { f | f i d V r , }
where the R a n g e C are candidates of Grid-clustering.
The Table 1 summarizes the notations used in the paper.

4. Proposed Approach

4.1. Naive Algorithm Design

The skeleton of the homogeneous-group mining in the geographic video database VDB contains three stages. Firstly, we transform the geo-tagged video into snapshots, i.e., a snapshot at timestamp 1, a snapshot at timestamp 2,…, a snapshot at timestamp n. Secondly, we cluster the FoVs for each snapshot at each timestamp based on density clustering (FDBC). Due to limited space, we omit the algorithm in this paper.
In Algorithm 1, FoVs are arbitrarily selected from each snapshot until all FoVs have been visited. If there are at least m i n o FoVs within a radius of ϵ , i.e., ( F = { f | d i s t ( f , f S t ) ϵ } | F | m i n o ), these FoVs are considered as part of the candidate cluster, and  E x p a n d F o V C l u s t e r generates a more refined cluster that satisfies the constraint. Next, the overlap between FoVs in the cluster is determined by the o v e r l a p A r e a function. Then, if the number of FoVs that satisfy the intersecting constraint is greater than the minimum FoV number m i n o , they are inserted into S C t e m p as candidates. Otherwise, they are marked as noise.
Algorithm 1: FDBC( S t , ϵ , m i n o , δ ).
Sensors 23 04443 i001
The detail for o v e r l a p A r e a is described as Definition 3, detecting whether a common area exists between two FoVs. We calculate the overlap area coefficient ( δ ) after judging the distance between two FoVs. However, it is not obligatory that as long as two FoVs intersect they view the same thing. In consequence of this, we follow to the intersection area A r e a i n t e r . A result of common-area weight of no less than A r e a i n t e r ( δ A r e a i n t e r ) is stored in the list C that is used for E x p a n d F o V C l u s t e r . The  E x p a n d F o V C l u s t e r is depicted in Algorithm 2.
Algorithm 2: ExpandFoVCluster( S t , ϵ , m i n o , F , V i s i t e d [ ] ).
Sensors 23 04443 i002
Algorithm 2 describes the expansion of the FoV cluster. For each FoV in C that has not been visited, mark it as visited and look for its neighbor C n e w . If the number in C n e w is not less than the minimum FoV number, use it as a candidate set for overlapping area calculation, and calculate the FoV afterward. If the number meets m i n o , it will be merged with the previously generated cluster, and the operation will be repeated until reaching the boundary of the cluster. ExpandFoVCluster is the working core of the algorithm and an essential operation is used for spatial query, retrieving all the FoVs belonging to a given region of the space and having a common viewable region. It can be shown that clusters are invariant with respect to the order of FoV selection in ExpandFoVCluster, and that the complexity of FDBC is O ( N 2 ) the worst, where N is the number of FoVs.
Finally, we discovered groups from videos according to clusters of snapshot’s timestamps. It is described in Algorithm 3. Before that, we introduced that the conditions for grouping geo-tagged videos are view field, location, and the time the people move together. In this phase, we use constraints to judge these geo-tagged videos; whether they are a group. If geo-tagged video objects in the same cluster meet the time constraints, we regard them as a group.
In more detail, the time cost of the FDBC algorithm can be expressed as follows:
C o s t F D B C = Θ ( S t , ϵ ) + Π ( S t , ϵ , m i n o ) + Γ ( C , δ ) + Λ ( S C ) ,
where Θ ( S t , ϵ ) is the time cost of FovNeighbors, Π ( S t , ϵ , m i n o ) is the time cost of ExpandFoVCluster, Γ ( C , δ ) is the time cost of densityoverlapArea, and  Λ ( S C ) is the time cost of BoundaryFOVBelong.
The Θ ( S t , ϵ ) representation is the time it takes to calculate the distance between each field of view (fov) of S t and determine whether the distance is part of the adjacent fov by making sure it is within the radius ϵ . The time required to construct the neighboring fov depends on the size of the set | F | in S t and the radius, as shown in Equation (8). As the radius ϵ increases, the number of candidate fovs increases, resulting in a longer overall algorithm execution time.
Θ ( S t , ϵ ) | F | θ F , ϵ
where θ F , ϵ represents the time required to calculate the neighbors associated with fov. Π ( S t , ϵ , m i n o ) means the time cost to ExpandFovCluster satisfiying m i n o and ϵ from a dataset F by FovNeighbors. The time cost can be estimated approximately using the following equation:
Π ( S t , ϵ , m i n o ) | F | π | F |     θ S t , ϵ π ,
where F is a set of candidates in F, and  π is the average time taken to perform ExpandFoVCluster. The number of m i n o affects the overall running time of the algorithm. θ S t , ϵ is a selectivity rate of Θ ( S t , ϵ ) .
Γ ( C , δ ) | F | π S t , ϵ , m i n o
Γ ( C , δ ) means the time cost to do densityOverlapArea for C with the parameter δ using Equation (4), where π S t , ϵ , m i n o is a selectivity rate of Π ( S t , ϵ , m i n o ) .
Λ ( S C ) | F | π S t , ϵ , m i n o γ c , δ
Λ ( S C ) means the time cost to do BoundaryFOVBelong. Each operation is performed for each element in the cluster, resulting in a time complexity of n 2 , and the algorithm execution time is influenced by the number of fov in each cluster where γ c , δ is a selectivity rate of Γ ( C , δ ) .
Algorithm 3: BoundaryFoVBelong( S C t e m p , m i n o ).
Sensors 23 04443 i003
Scan clusters in the order of each snapshot with the snapshots after it, and detect the FoV belonging to the same geo-video in the cluster between snapshots. Add these FoVs into the A . C . When the number of FoVs | ( A . C ) | in A . C is not greater than m i n o , put the timestamps of FoVs into A . T . It is initialized using the common FoVs in clusters with a size no smaller than m i n o . After that, generate new candidates A by extending. Next, for every group candidate A in A , we execute the C a n d i d a t e s F i l t e r to obtain the final homogeneous groups, which is described in detail in Algorithm 4.
Definition 9 (Groups Candidates).
Considering Definitions 5 and 6, the homogeneous groups need to satisfy the Closeness and certain time constraints. Therefore, let A = { a 1 ,…, a m } be a set of the candidate set, and let a i  = <C, T> be a candidate where a i is a sub-candidate of A .
Algorithm 4: MiningHomogenousGroups( SC , m i n o , k, s, g).
Sensors 23 04443 i004
In Algorithm 4, we judge the succession and connection. For succession, the most significant gap between the two time-segments is whether it is lower than g. If not, remove it. Furthermore, we remove it if the sequential time is bigger or equal to s.
Then, we judge the occurrence that the minimal number of occurrences for FoVs belongs to the same cluster in each segment; if it is less than k, we also remove a from A . After that, all the homogeneous groups are stored in P . As an example, a glance at Table 2 illustrates an instance. If we set the parameters m i n o  = 2, s = 1, g = 2, k = 2, and δ  = 0.5, then all of the groups that we would capture are shown in Table 2.
For the groups part, we assume the number of clusters is N and the computational complexity is N 2 + | T 2 | .

4.2. Performance Enhancement Using a Grid Index Approach

This paper proposes constructing a grid index based on area-density clustering for each snapshot. Because FoV clustering is different from point clustering, an essential condition needs to calculate the overlap area. The purpose of using the grid is to use MBR to narrow the scope of the calculation further. The details are described in the following.
For example, if we want to find all the FoVs associated with a specific FoV, we need to compute the f M B R . It is represented by the area in the red box in Figure 5a. Furthermore, we can get the id of the cell that the FoV to be queried belongs to. The process is depicted in Algorithm 5. After that, we need to judge whether the number of FoVs in V f is greater than the threshold m i n o . If the quantity condition is met, we execute the second level. The second level further divides each grid in the first level into c × c sub-cells; it in depicted in Figure 5b. The width τ of a sub cell is as follows:
τ = ( ϵ / 2 ) / c
Algorithm 5: FCBG( S t , ϵ , m i n o , c, δ ).
Sensors 23 04443 i005
These sub-cells are used to calculate the intersection of the area between two FoVs. These FoVs will be marked as visited if they satisfy the intersection condition. Next, we need to judge the distance again. If the distance between f and f is greater than the distance threshold ϵ , remove f from candidate cluster C. Finally, if the number of FoVs in C is no less than m i n o , we obtain one of the clusters in a snapshot. The computational cost of FCBG is O ( N l o g N ) with a grid index, where N is the number FoVs in the snapshot.
In more detail, the time cost of the FDBC algorithm can be expressed as follows:
C o s t F C B G = Φ ( S t , ϵ , c ) + Γ ( C , δ ) + Λ ( S C ) ,
where Φ ( S t , ϵ ) is the time cost of BuildFovGridIndex and calculates MBR, Π ( S t , ϵ , m i n o ) is the time cost of ExpandFoVCluster, and Γ ( C , δ ) is the time cost of densityoverlapArea, and Λ ( S C ) is the time cost of BoundaryFOVBelong.

5. Experimental Evaluation

We describe the selection of the parameters for clustering and evaluate the FoV clustering’s effectiveness based on the DBSCAN clustering (FDBC) algorithm and the FoV clustering based on Grid (FCBG) by experimenting with two datasets. Moreover, the investigation of the analogous interest groups captured by utilizing two clustering methods is discussed in this chapter. The description of the experimental setting and the geo-tagged-videos datasets are described in Section 5.1. The performance evaluation of clustering under different datasets by attributing various values of the parameters is described in Section 5.2.

5.1. Experiment Setting

All experiments were performed using an Intel (R) Core (TM) i7-3770 CPU machine with 16 GB of RAM and 20 TB of hard disk space. The algorithms were implemented based on JDK1.8.
Datasets: This experiment was based on two types of datasets, the BDD dataset [9] and the geo-tagged-videos data constructed based on a real-world shopping dataset for trajectories. The data sample was recorded per second. The experiment set the different intervals of the duration depending on the datasets used in our experiment. The statistics of the datasets are listed in Table 3.

5.2. Performance of Clustering

Selection of m i n o Value: The rule of thumb [39] is Formula (14), where dim represents the dimension of the data to be clustered.
m i n o d i m + 1
If m i n o = 1, each independent point is a cluster. On the other hand, if m i n o 2 , the result is the same as the hierarchical distance to the nearest neighbor. Therefore, m i n o must be a value greater than or equal to 3. If the selected value is too small, the result in the sparse cluster will be considered as a boundary point because the density is less than m i n o , and will not be used for further expansion of the class; if the value is too large, two adjacent clusters with higher density may be merged into the same cluster. At the same time, the value of m i n o should also be selected according to the size of the dataset. In this paper, due to our research it is based on the two-dimensional FoV model, hence, we set m i n o to 4, 8, 12, 16, and 20.
Selection of ϵ value: If the value of ϵ is too small, most of the data are missed. On the contrary, if the parameter setting is too large, multiple clusters or most objects are merged into the same cluster. In our work, we used the Euclidean method (as following Formula (15) [40] to calculate the k-distance graph. That is, given the parameter k, calculate the k-th nearest neighbor distance corresponding to each position and and sort in descending order; call this picture a sorted k-distance graph. Here, k = m i n o .
d i s t f p , f p = ( f p . x f p . x ) 2 + ( f p . y f p . y ) 2
One of the conditions for two FoVs to have a common area is that the distance between the two positions is not greater than ϵ . Therefore, we only discuss distances within 100. The number of distances is in different ranges. This paper utilized a histogram to display the frequency of the distance range from 2.3 to 100 and choose the average in adjacent partitions with significant differences in frequency of distance as the parameter, as per Formula (16), where N is the frequency of distance.
ε = 1 N j = 1 N ( i = 1 k d i s t i ( f p , f p ) k ) j
After that, we compare the running time of FDBC and FCBG with different parameters, i.e., the minimal number of FoV ( m i n o ), minimal eps between FoVs ( ϵ ), minimal overlap area ( δ ). As shown in Figure 6, Figure 7 and Figure 8, FCBG is faster than FDBC under the same parameters.
The effect of m i n o : To test the influence of the m i n o threshold, (Shopping data: fixed ϵ = 3, δ = 0.5; BDD100K: ϵ = 50, δ = 0.5) we changed the minimal number of FoV threshold from 4 to 20. Figure 6 shows the clustering cost of the algorithm FDBC and the algorithm FCBG based on the m i n o threshold. According to Figure 6, we observe that the running time of FDBC becomes longer as the minimal FoV threshold increases because the first step of FDBC is to find the nearest neighbor of ϵ . It calculates FoV one by one and then judges the minimal number which increases the computation. However, FCBG, according to the MBR, reduces the region of calculation. After m i n o = 12, the time cost rises smoothly in Shopping data because after the FDBC calculates the distance, if m i n o is not satisfied, the following mark calculation is required. Therefore, the time cost has a slow rise depending on the different distribution densities of the data. It can be seen that there are not many clusters that meet the 12 FoVs as a class. In addition, the BDD100K shows that FCBG also has a significant increase with the increase in m i n o ; but overall, the time of FCBG is lower than that of FDBC.
The effect of ϵ : To test the influence of the distance threshold, we fixed Shopping data and BDD100K to m i n o = 4 and δ = 0.5, and changed the distance threshold from 3 m to 27 m. Figure 7 shows the clustering cost of the algorithm FDBC and the algorithm FCBG based on the distance threshold. According to Figure 7, we observe that the running time of FDBC becomes longer as the distance threshold increases because the first step of FDBC is to find the nearest neighbor of m i n o within the distance threshold. A more extensive calculation is needed as the distance increases because this means the number of FoVs that need to be calculated will increase. On the contrary, distance is not the primary influencing factor of FCBG. As FCBG is based on the MBR of FoV and based on the grid index, it calculates the overlap coefficient δ = 0.5 and obtains m i n o FoVs whose distance threshold is not greater than eps. Therefore, the time cost of FCBG does not fluctuate significantly with an increase in ϵ , as shown in the figure. In addition, in terms of running time, the efficiency of our optimized FCBG is better than that of FDBC.
The effect of δ : To evaluate the effect of varying δ , we fixed the threshold (Shopping: ϵ = 3 m, m i n o = 4; BDD100K: ϵ = 50 m, m i n o = 4), and set δ from 0.2 to 0.4. Figure 8 illustrates the clustering time of the two algorithms. In the experiment, the running time depends on the distribution of the data. When the density that satisfies the overlap threshold is large, the amount of calculation is significant, and vice versa. Nevertheless, the overall efficiency of FCBG is still higher than that of FDBC.
In order to further observe the clustering performance, this experiment studies the cluster size to better compare the clustering effects of different methods on different datasets. That is, the different datasets are used to cluster FoVs for each snapshot in S = { S 1 , S 2 ,…, S t } at timestamp t. We average the number of clusters generated in S t to compare the clustering effects of FBDC and FCBG, as Figure 9 shows.
In Figure 9, since the number of different shopping videos is less than the number of videos in BDD100K, the obtained shopping cluster size is less than that of BDD100K. According to the description in Figure 5, the number of clusters that the algorithm FCBG can capture is higher than the number captured by the algorithm FDBC in both data sets.

6. Conclusions

In this paper, we firstly proposed novel algorithms to discover homogeneous groups from geo-tagged videos, using view directions. Secondly, we extended the density clustering algorithm to support FoV clustering and proposed an optimization model based on a two-level grid-based index structure. For experimental validation, we used a real shopping dataset and an FoV-extended synthetic video dataset. The synthetic video set was extended from the real BDD dataset. The results show that our algorithms are effective to discover homogeneous groups from geo-tagged videos using view directions. It is expected that cameras with directional sensors will become much more popular in the future, and we believe that the new algorithms proposed in this paper can be heavily utilized for social media analysis.

Author Contributions

Conceptualization, X.D. and K.W.N.; Methodology, X.D. and K.W.N.; Software, X.D.; Validation, D.J.L. and K.W.N.; Formal analysis, X.D.; Investigation, X.D.; Writing—original draft, X.D.; Writing—review & editing, D.J.L. and K.W.N.; Project administration, K.W.N. All authors have read and agreed to the published version of the manuscript.

Funding

This work is supported by the Korea Agency for Infrastructure Technology Advancement (KAIA) grant funded by the Ministry of Land, Infrastructure and Transport (Grant RS-2022-00143336).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Syafrudin, M.; Alfian, G.; Fitriyani, N.L.; Rhee, J. Performance analysis of IoT-based sensor, big data processing, and machine learning model for real-time monitoring system in automotive manufacturing. Sensors 2018, 18, 2946. [Google Scholar] [CrossRef] [PubMed]
  2. Judd, T.; Ehinger, K.; Durand, F.; Torralba, A. Learning to predict where humans look. In Proceedings of the 2009 IEEE 12th International Conference on Computer Vision, Kyoto, Japan, 29 September–2 October 2009; pp. 2106–2113. [Google Scholar]
  3. Kyriakou, K.; Resch, B.; Sagl, G.; Petutschnig, A.; Werner, C.; Niederseer, D.; Liedlgruber, M.; Wilhelm, F.; Osborne, T.; Pykett, J. Detecting moments of stress from measurements of wearable physiological sensors. Sensors 2019, 19, 3805. [Google Scholar] [CrossRef]
  4. Lu, Y.; To, H.; Alfarrarjeh, A.; Kim, S.H.; Yin, Y.; Zimmermann, R.; Shahabi, C. GeoUGV: User-generated mobile video dataset with fine granularity spatial metadata. In Proceedings of the 7th International Conference on Multimedia Systems, Klagenfurt, Austria, 10–13 May 2016; pp. 1–6. [Google Scholar]
  5. Ding, W.; Tian, J.; Lee, Y.; Yang, K.; Nam, K.W. VVS: Fast Similarity Measuring of FoV-Tagged Videos. IEEE Access 2020, 8, 190734–190745. [Google Scholar] [CrossRef]
  6. Jalal, A.; Kamal, S.; Kim, D. A depth video sensor-based life-logging human activity recognition system for elderly care in smart indoor environments. Sensors 2014, 14, 11735–11759. [Google Scholar] [CrossRef]
  7. Gurrin, C.; Smeaton, A.F.; Doherty, A.R. LifeLogging: Personal Big Data. Found. Trends Inf. Retr. 2014, 8, 1–125. [Google Scholar] [CrossRef]
  8. Brščić, D.; Kanda, T.; Ikeda, T.; Miyashita, T. Person tracking in large public spaces using 3-D range sensors. IEEE Trans. Hum.-Mach. Syst. 2013, 43, 522–534. [Google Scholar] [CrossRef]
  9. Yu, F.; Chen, H.; Wang, X.; Xian, W.; Chen, Y.; Liu, F.; Madhavan, V.; Darrell, T. BDD100K: A Diverse Driving Dataset for Heterogeneous Multitask Learning. In Proceedings of the 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition, CVPR 2020, Seattle, WA, USA, 13–19 June 2020; Computer Vision Foundation/IEEE: Piscataway, NJ, USA, 2020; pp. 2633–2642. [Google Scholar] [CrossRef]
  10. Vieira, M.R.; Bakalov, P.; Tsotras, V.J. On-line discovery of flock patterns in spatio-temporal data. In Proceedings of the 17th ACM SIGSPATIAL International Conference on Advances in Geographic Information Systems, Seattle, WA, USA, 4–6 November 2009; pp. 286–295. [Google Scholar]
  11. Tanaka, P.S.; Vieira, M.R.; Kaster, D.S. Efficient Algorithms to Discover Flock Patterns in Trajectories. In Proceedings of the GeoInfo, Kuala Lumpur, Malaysia, 28–30 October 2015; pp. 56–67. [Google Scholar]
  12. Turdukulov, U.; Calderon Romero, A.O.; Huisman, O.; Retsios, V. Visual mining of moving flock patterns in large spatio-temporal data sets using a frequent pattern approach. Int. J. Geogr. Inf. Sci. 2014, 28, 2013–2029. [Google Scholar] [CrossRef]
  13. Jeung, H.; Yiu, M.L.; Zhou, X.; Jensen, C.S.; Shen, H.T. Discovery of convoys in trajectory databases. Proc. VLDB Endow. 2008, 1, 1068–1080. [Google Scholar] [CrossRef]
  14. Jeung, H.; Shen, H.T.; Zhou, X. Convoy queries in spatio-temporal databases. In Proceedings of the 2008 IEEE 24th International Conference on Data Engineering, Cancun, Mexico, 7–12 April 2008; pp. 1457–1459. [Google Scholar]
  15. Li, Z.; Ding, B.; Han, J.; Kays, R. Swarm: Mining Relaxed Temporal Moving Object Clusters. Proc. VLDB Endow. 2010, 3, 723–734. [Google Scholar] [CrossRef]
  16. Van den Bergh, F.; Engelbrecht, A.P. A study of particle swarm optimization particle trajectories. Inf. Sci. 2006, 176, 937–971. [Google Scholar] [CrossRef]
  17. Kalnis, P.; Mamoulis, N.; Bakiras, S. On discovering moving clusters in spatio-temporal data. In Proceedings of the International Symposium on Spatial and Temporal Databases, Angra dos Reis, Brazil, 22–24 August 2005; pp. 364–381. [Google Scholar]
  18. Li, Y.; Bailey, J.; Kulik, L. Efficient mining of platoon patterns in trajectory databases. Data Knowl. Eng. 2015, 100, 167–187. [Google Scholar] [CrossRef]
  19. Fan, Q.; Zhang, D.; Wu, H.; Tan, K.L. A general and parallel platform for mining co-movement patterns over large-scale trajectories. Proc. VLDB Endow. 2016, 10, 313–324. [Google Scholar] [CrossRef]
  20. Chen, L.; Gao, Y.; Fang, Z.; Miao, X.; Jensen, C.S.; Guo, C. Real-time distributed co-movement pattern detection on streaming trajectories. Proc. VLDB Endow. 2019, 12, 1208–1220. [Google Scholar] [CrossRef]
  21. Cho, N.; Kang, Y. Space-time density of field trip trajectory: Exploring spatio-temporal patterns in movement data. Spat. Inf. Res. 2017, 25, 141–150. [Google Scholar] [CrossRef]
  22. Lee, H.; Kang, Y. Mining tourists’ destinations and preferences through LSTM-based text classification and spatial clustering using Flickr data. Spat. Inf. Res. 2021, 29, 825–839. [Google Scholar] [CrossRef]
  23. Ma, H.; Ay, S.A.; Zimmermann, R.; Kim, S.H. Large-scale geo-tagged video indexing and queries. GeoInformatica 2014, 18, 671–697. [Google Scholar] [CrossRef]
  24. Kim, Y.; Kim, J.; Yu, H. GeoTree: Using spatial information for georeferenced video search. Knowl.-Based Syst. 2014, 61, 1–12. [Google Scholar] [CrossRef]
  25. Cai, Y.; Lu, Y.; Kim, S.H.; Nocera, L.; Shahabi, C. Querying geo-tagged videos for vision applications using spatial metadata. EURASIP J. Image Video Process. 2017, 2017, 19. [Google Scholar] [CrossRef]
  26. Ay, S.A.; Zimmermann, R.; Kim, S.H. Viewable scene modeling for geospatial video search. In Proceedings of the 16th ACM international conference on Multimedia, Vancouver, BC, Canada, 27–31 October 2008; pp. 309–318. [Google Scholar]
  27. Constantinou, G.; Shahabi, C.; Kim, S.H. Spatial Keyframe Extraction Of Mobile Videos For Efficient Object Detection At The Edge. In Proceedings of the 2020 IEEE International Conference on Image Processing (ICIP), Abu Dhabi, United Arab Emirates, 25–28 October 2020; pp. 1466–1470. [Google Scholar]
  28. Kim, S.H.; Ay, S.A.; Yu, B.; Zimmermann, R. Vector model in support of versatile georeferenced video search. In Proceedings of the First Annual ACM SIGMM Conference on Multimedia Systems, Phoenix, AZ, USA, 22–23 February 2010; pp. 235–246. [Google Scholar]
  29. Huang, Z. Extensions to the k-means algorithm for clustering large data sets with categorical values. Data Min. Knowl. Discov. 1998, 2, 283–304. [Google Scholar] [CrossRef]
  30. Peng, K.; Zheng, L.; Xu, X.; Lin, T.; Leung, V.C. Balanced iterative reducing and clustering using hierarchies with principal component analysis (PBIRCH) for intrusion detection over big data in mobile cloud environment. In Proceedings of the International Conference on Security, Privacy and Anonymity in Computation, Communication and Storage, Zhangjiajie, China, 16–18 November 2018; pp. 166–177. [Google Scholar]
  31. Ester, M.; Kriegel, H.; Sander, J.; Xu, X. A Density-Based Algorithm for Discovering Clusters in Large Spatial Databases with Noise. In Proceedings of the Second International Conference on Knowledge Discovery and Data Mining (KDD-96), Portland, OR, USA, 2–4 August 1996; Simoudis, E., Han, J., Fayyad, U.M., Eds.; AAAI Press: Washington, DC, USA, 1996; pp. 226–231. [Google Scholar]
  32. Lee, Y.; Nam, K.W.; Ryu, K.H. Fast mining of spatial frequent wordset from social database. Spat. Inf. Res. 2017, 25, 271–280. [Google Scholar] [CrossRef]
  33. Mantiuk, R.K.; Denes, G.; Chapiro, A.; Kaplanyan, A.; Rufo, G.; Bachy, R.; Lian, T.; Patney, A. Fovvideovdp: A visible difference predictor for wide field-of-view video. ACM Trans. Graph. (TOG) 2021, 40, 1–19. [Google Scholar] [CrossRef]
  34. Park, J.Y.; Ryu, D.J.; Nam, K.W.; Jang, I.; Jang, M.; Lee, Y. DeepDBSCAN: Deep density-based clustering for geo-tagged photos. ISPRS Int. J. Geo-Inf. 2021, 10, 548. [Google Scholar] [CrossRef]
  35. Nam, K.W.; Yang, K. RealROI: Discovering Real Regions of Interest From Geotagged Photos. IEEE Access 2022, 10, 83489–83497. [Google Scholar] [CrossRef]
  36. Mao, Y.; Zhong, H.; Qi, H.; Ping, P.; Li, X. An adaptive trajectory clustering method based on grid and density in mobile pattern analysis. Sensors 2017, 17, 2013. [Google Scholar] [CrossRef] [PubMed]
  37. Suo, Y.; Chen, W.; Claramunt, C.; Yang, S. A ship trajectory prediction framework based on a recurrent neural network. Sensors 2020, 20, 5133. [Google Scholar] [CrossRef] [PubMed]
  38. Moulton, R.; Jiang, Y. Maximally Consistent Sampling and the Jaccard Index of Probability Distributions. In Proceedings of the 2018 IEEE International Conference on Data Mining (ICDM), Singapore, 17–20 November 2018; pp. 347–356. [Google Scholar]
  39. Tang, J.; Bi, W.; Liu, F.; Zhang, W. Exploring urban travel patterns using density-based clustering with multi-attributes from large-scaled vehicle trajectories. Phys. A Stat. Mech. Appl. 2021, 561, 125301. [Google Scholar] [CrossRef]
  40. Danielsson, P.E. Euclidean distance mapping. Comput. Graph. Image Process. 1980, 14, 227–248. [Google Scholar] [CrossRef]
Figure 1. An example of geo-tagged videos with FoVs.
Figure 1. An example of geo-tagged videos with FoVs.
Sensors 23 04443 g001
Figure 2. Representative diagram of an FoVScene model in 2 dimensions.
Figure 2. Representative diagram of an FoVScene model in 2 dimensions.
Sensors 23 04443 g002
Figure 3. An example of an FoV sequence.
Figure 3. An example of an FoV sequence.
Sensors 23 04443 g003
Figure 4. Examples of snapshots in various FoV videos.
Figure 4. Examples of snapshots in various FoV videos.
Sensors 23 04443 g004
Figure 5. An example of FoV clustering using a grid index.
Figure 5. An example of FoV clustering using a grid index.
Sensors 23 04443 g005
Figure 6. Effect of various m i n o .
Figure 6. Effect of various m i n o .
Sensors 23 04443 g006
Figure 7. Effect of various ϵ .
Figure 7. Effect of various ϵ .
Sensors 23 04443 g007
Figure 8. Effect of various δ .
Figure 8. Effect of various δ .
Sensors 23 04443 g008
Figure 9. The clustering performance with FDBC and FCBG.
Figure 9. The clustering performance with FDBC and FCBG.
Sensors 23 04443 g009
Table 1. Frequently used notations.
Table 1. Frequently used notations.
NotationDescription
F FoV sequence in a geo-video
T Sequence of all timestamps after partition in a geo-video
δ Overlap coefficient of common region for two FoVS
SC FoV collection of snapshots at the timestamp
C Clusters in a snapshot S t
kMinimal number of occurrence for FoVs belonging to the same cluster
m i n o Minimal number of FoVs
Table 2. The homogeneous groups corresponding to Figure 4.
Table 2. The homogeneous groups corresponding to Figure 4.
Groups
GroupsID123456
VideoID1, 23, 45, 67, 87, 8, 97, 9
SnapshotID1, 2, 41, 2, 4, 51, 2, 4, 51, 2, 3, 4, 51, 3, 51, 3, 5
Table 3. The Detail of Datasets.
Table 3. The Detail of Datasets.
StatisticsShoppingBDD100K
Number of geo-videos450029,490
The longest time-domain length of video42,00030,000
The total of FoVs135,0001,238,580
The number of snapshots1040
Cluster size44/70412/965
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Di, X.; Lew, D.J.; Nam, K.W. Discovering Homogeneous Groups from Geo-Tagged Videos. Sensors 2023, 23, 4443. https://doi.org/10.3390/s23094443

AMA Style

Di X, Lew DJ, Nam KW. Discovering Homogeneous Groups from Geo-Tagged Videos. Sensors. 2023; 23(9):4443. https://doi.org/10.3390/s23094443

Chicago/Turabian Style

Di, Xuejing, Dong June Lew, and Kwang Woo Nam. 2023. "Discovering Homogeneous Groups from Geo-Tagged Videos" Sensors 23, no. 9: 4443. https://doi.org/10.3390/s23094443

APA Style

Di, X., Lew, D. J., & Nam, K. W. (2023). Discovering Homogeneous Groups from Geo-Tagged Videos. Sensors, 23(9), 4443. https://doi.org/10.3390/s23094443

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop