Two–Stage Detection and Localization of Inter–Frame Tampering in Surveillance Videos Using Texture and Optical Flow
Abstract
:1. Introduction
- We propose an inter–frame video tampering detection technique (stage 1) to identify suspicious tampered videos by tracing out alterations in texture patterns. To detect these alterations, we propose the extraction of texture features by employing the MR–LBP. The proposed method simultaneously detects both frame duplication and insertion tampering, unlike current state–of–the–art techniques [12,15,37,38,39], which are limited to the detection of only one type of tampering. Additionally, the proposed method imposes no constraints on video formats, the type of capturing device, frame rate, or the minimum number of duplicated or inserted frames required to detect tampering; it can detect the duplication and insertion of as few as ten frames. In contrast, the deep learning–based method described in Ref. [28] only detects and locates tampering regions when tampered frames occur in multiples of 10 and is unable to detect tampering involving fewer than 25 frames. It also assumes that tampered frames are only inserted in the static section of the video when frame duplication is performed. In the proposed method, the video has gray–level frames encoded with local binary pattern (LBP) results in a feature vector of dimension 256, making it computationally efficient compared to the deep learning–based method with high dimensional features.
- For the localization of video inter–frame tampering (stage 2), we suggest independently employing the OF aggregation and standard deviation of MR–LBP features of consecutive frames; this removes false positives. The inconsistency in OF aggregation pinpoints the exact start and end of the tampered region in the surveillance video.
- Due to a lack of benchmark datasets, the method is trained and tested on our developed large dataset, the COMSATS Structured Video Tampering Evaluation Dataset (CSVTED), which comprises challenging videos with different complexity levels and tampering quality and is cross–validated on publicly available datasets. The benchmark public domain datasets contain a variety of videos, including Event–Object–Person (EOP)–based tampering. High detection accuracy using the CSVTED and other public domain datasets strongly validates the method’s generalization (in previous studies, no cross–dataset validation has been carried out). It has a high generalization capability.
- The performance of the proposed method is compared with state–of–the–art methods in terms of accuracy. Comparison results show that MR–LBP features make a major contribution to detecting frame duplication and insertion tampering, with accuracies of 99.71% and 99.87%, respectively.
- The method is computationally efficient with a processing time of microseconds.
2. Literature Review
Sr. No. | Author, Year | Method/Features | Tampering Identified | Dataset | Results | Merits/Demerits | |||
---|---|---|---|---|---|---|---|---|---|
P (%) | R (%) | DA (%) | Other | ||||||
1 | Ulutas et al. (2017) [52] | Binary feature extraction and PSNR | Duplication, Mirroring | 10 videos | 99.98 100 | 99.30 97.34 | 99.35 98.20 | - |
|
2 | Kingra Aggarwal et al. (2017) [21] | Prediction residual (PR) and optical flow (OF) | Insertion, Deletion, Duplication | Developed tampered dataset personally | - | - | 80, 83, 75 92, 83, 88 100, 96, 100 | LA: 80% |
|
3 | Ulutas et al. (2018) [38] | Bag-of-words (BOW) model with SIFT | Duplication (moving and static views) | 31 videos | 97.94 98.57 | 97.65 99.13 | 96.73 98.17 | - |
|
4 | Zhao, Wang et al. (2018) [43] | Similarity between H-S-V histograms, SURF feature extraction along with FLANN matching | Insertion, Deletion, Duplication, and Localization | 10 test shots | 98.07 | 100 | 99.01 | - |
|
5 | Huang, Zhang et al. (2018) [49] | Wavelet packet de-composition, quaternion DCT features | Deletion, Insertion | 115 videos from OV and SULFA, 124 personally recorded | 0.9876 | 0.9847 | - | - |
|
6 | Jia et al. (2018) [15] | Optical flow (OF) sum consistency and correlation between frames. | Duplication, Computation | VTL: 55 SULFA: 36 DERF: 24 videos time: 1.623 µs/pixel | 0.985 | 0.985 | - | - |
|
7 | Fadl et al. (2018) [53] | Energy difference b/w frames, SNR, and spatiotemporal energy | Duplication, Insertion, Deletion | 120 videos from SULFA, 28 from and 3 from IVY Lab | 0.97 0.99 0.97 | 0.99 0.99 0.95 | - | F1: 0.98 0.99 0.96 |
|
8 | Bakas et al. (2018) [27] | 3D-CNN | Insertion, Deletion, Duplication | UCF101: 9000 videos | - | - | 97% (average) | - |
|
9 | Long, Basharat et al. (2019) [5] | I3D along with Resnet152 | Duplication | MFC-18, VIRAT: 12, IPhone-4: 17 videos | - | - | - | AUC: 84.05 81.46 |
|
10 | Fadl, Sondos, et al. (2020) [11] | Temporal average (TP), Edge change ratio (ECR), and GLCM | Duplication, Duplication with Shuffling | 51 from SULFA, LASIESTA, and IVY lab | 0.99 0.95 | 0.98 0.98 | - | - |
|
11 | Kharat et al. (2020) [12] | Motion vectors and SIFT feature are used with random sample consensus algorithm to locate tampering | Duplication | 20 videos from YouTube Movies | 99.9 | 99.7 | 99.8 | - |
|
12 | Fadl et al. (2021) [28] | 2D-CNN with multi-class support vector machine (MSVM) | Insertion, Deletion, Duplication | 13135 videos from SULFA, VIRAT, LASIESTA, and IVY. | - | - | 99.9 98.7 98.5 | - |
|
13 | Alsakar et al. (2021) [25] | Correlation with arbitrary number of core tensors | Insertion, Deletion | 18 videos taken from TRACE library | 96 92 | 94 90 | - | F1: 95, 91 |
|
14 | Panchal et al. (2023) [54] | Sets of video quality assessment attributes are selected and multiple linear regression is applied | Deletion | Developed dataset using 80 videos of TDTVD, SULFA, UCF-101, and VTD | - | - | 96.25% | - |
|
15 | Shehnaz and Kaur (2024) [13] | HoG with LBP | Duplication, Deletion, Insertion | Developed tampered dataset using VTD and SULFA | 99.4 | 99.2 | 99.6 | F1: 99.5 |
|
16 | Akhtar et al. (2024) [19] | 2D-CNN with autoencoder and LSTM/GRU | Insertion, Deletion | Developed CSVTED dataset of 2555 videos | 98.77 84.59 | 98.99 94.20 | 98.98 94.18 | F1: 98.87 89.05 |
|
3. Proposed Method
3.1. Problem Formulation
3.2. Stage 1: Proposed Method for Detection
3.2.1. Preprocessing
3.2.2. Feature Extraction
3.2.3. Classification
3.3. Stage 2: Proposed Method for Localization
3.3.1. Optical Flow Aggregation Calculation
3.3.2. Standard Deviation
4. Evaluation Protocols
4.1. Experimental Setup
4.2. Dataset Description and Preparation
4.3. Evaluation Procedure
5. Experimental Results
5.1. Frame Duplication Detection
5.1.1. Impact of Using of Aggregation Only
5.1.2. Impact of Using MR–LBP Only
5.1.3. Impact of Using MR–LBP and Aggregation
5.1.4. Cross–Dataset Evaluation
5.2. Frame Insertion Detection
Cross–Dataset Evaluation
6. Discussion
7. Conclusions
Author Contributions
Funding
Data Availability Statement
Conflicts of Interest
References
- Akhtar, N.; Saddique, M.; Asghar, K.; Bajwa, U.I.; Hussain, M.; Habib, Z. Digital Video Tampering Detection and Localization: Review, Representations, Challenges and Algorithm. Mathematics 2022, 10, 168. [Google Scholar] [CrossRef]
- Nabi, S.T.; Kumar, M.; Singh, P.; Aggarwal, N.; Kumar, K. A comprehensive survey of image and video forgery techniques: Variants, challenges, and future directions. Multimed. Syst. 2022, 28, 939–992. [Google Scholar] [CrossRef]
- Mohiuddin, S.; Malakar, S.; Kumar, M.; Sarkar, R. A comprehensive survey on state-of-the-art video forgery detection techniques. Multimed. Tools Appl. 2023, 82, 33499–33539. [Google Scholar] [CrossRef]
- Huang, C.C.; Zhang, Y.; Thing, V.L.L. Inter-frame video forgery detection based on multi-level subtraction approach for realistic video forensic applications. In Proceedings of the IEEE 2nd International Conference on Signal and Image Processing (ICSIP), Singapore, 4–6 August 2017; IEEE: Piscataway, NJ, USA, 2017. [Google Scholar]
- Long, C.; Basharat, A.; Hoogs, A.; Singh, P.; Farid, H. A Coarse-to-fine Deep Convolutional Neural Network Framework for Frame Duplication Detection and Localization in Forged Videos. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR) Workshops, Long Beach, CA, USA, 16–20 June 2019; IEEE: Piscataway, NJ, USA, 2019. [Google Scholar]
- Johnston, P.; Elyan, E. A review of digital video tampering: From simple editing to full synthesis. Digit. Investig. 2019, 29, 67–81. [Google Scholar] [CrossRef]
- Wang, W.; Farid, H. Exposing digital forgeries in video by detecting duplication. In Proceedings of the 9th Workshop on Multimedia & Security, Dallas, TX, USA, 20–21 September 2007. [Google Scholar]
- Wang, Q.; Li, Z.; Zhang, Z.; Ma, Q. Video Inter-Frame Forgery Identification Based on Consistency of Correlation Coefficients of Gray Values. J. Comput. Commun. 2014, 2, 51–57. [Google Scholar] [CrossRef]
- Singh, G.; Singh, K. Video frame and region duplication forgery detection based on correlation coefficient and coefficient of variation. Multimed. Tools Appl. 2019, 78, 11527–11562. [Google Scholar] [CrossRef]
- Zhang, Z.; Hou, J.; Ma, Q.; Li, Z. Efficient video frame insertion and deletion detection based on inconsistency of correlations between local binary pattern coded frames. Secur. Commun. Networks 2015, 8, 311–320. [Google Scholar] [CrossRef]
- Fadl, S.; Megahed, A.; Han, Q.; Qiong, L. Frame duplication and shuffling forgery detection technique in surveillance videos based on temporal average and gray level co-occurrence matrix. Multimed. Tools Appl. 2020, 79, 17619–17643. [Google Scholar] [CrossRef]
- Kharat, J.; Chougule, S. A passive blind forgery detection technique to identify frame duplication attack. Multimed. Tools Appl. 2020, 79, 8107–8123. [Google Scholar] [CrossRef]
- Shehnaz; Kaur, M. Detection and localization of multiple inter-frame forgeries in digital videos. Multimed. Tools Appl. 2024, 83, 71973–72005. [Google Scholar] [CrossRef]
- Feng, C.; Xu, Z.; Jia, S.; Zhang, W.; Xu, Y. Motion-adaptive frame deletion detection for digital video forensics. IEEE Trans. Circuits Syst. Video Technol. 2016, 27, 2543–2554. [Google Scholar] [CrossRef]
- Jia, S.; Xu, Z.; Wang, H.; Feng, C.; Wang, T. Coarse-to-Fine Copy-Move Forgery Detection for Video Forensics. IEEE Access 2018, 6, 25323–25335. [Google Scholar] [CrossRef]
- Zampoglou, M.; Markatopoulou, F.; Mercier, G.; Touska, D.; Apostolidis, E.; Papadopoulos, S.; Cozien, R.; Patras, I.; Mezaris, V.; Kompatsiaris, I. Detecting Tampered Videos with Multimedia Forensics and Deep Learning. In Proceedings of the International Conference on Multimedia Modeling, Thessaloniki, Greece, 8–11 January 2019; Springer: Cham, Switzerland, 2019. [Google Scholar]
- Johnston, P.; Elyan, E.; Jayne, C. Video tampering localisation using features learned from authentic content. Neural Comput. Appl. 2020, 32, 12243–12257. [Google Scholar] [CrossRef]
- Shelke, N.A.; Kasana, S.S. Multiple forgery detection in digital video with VGG-16-based deep neural network and KPCA. Multimed. Tools Appl. 2024, 83, 5415–5435. [Google Scholar] [CrossRef]
- Akhtar, N.; Hussain, M.; Habib, Z. DEEP-STA: Deep Learning-Based Detection and Localization of Various Types of Inter-Frame Video Tampering Using Spatiotemporal Analysis. Mathematics 2024, 12, 1778. [Google Scholar] [CrossRef]
- Wang, Q.; Li, Z.; Zhang, Z.; Ma, Q. Video inter-frame forgery identification based on optical flow consistency. Sens. Transducers 2014, 166, 229. [Google Scholar]
- Kingra, S.; Aggarwal, N.; Singh, R.D. Inter-frame forgery detection in H.264 videos using motion and brightness gradients. Multimed. Tools Appl. 2017, 76, 25767–25786. [Google Scholar] [CrossRef]
- Singh, R.D.; Aggarwal, N. Optical flow and prediction residual based hybrid forensic system for inter-frame tampering detection. J. Circuits, Syst. Comput. 2017, 26, 1750107. [Google Scholar] [CrossRef]
- Yu, L.; Wang, H.; Han, Q.; Niu, X.; Yiu, S.; Fang, J.; Wang, Z. Exposing frame deletion by detecting abrupt changes in video streams. Neurocomputing 2016, 205, 84–91. [Google Scholar] [CrossRef]
- Stamm, M.C.; Lin, W.S.; Liu, K.J.R. Temporal forensics and anti-forensics for motion compensated video. IEEE Trans. Inf. Forensics Secur. 2012, 7, 1315–1329. [Google Scholar] [CrossRef]
- Alsakar, Y.M.; Mekky, N.E.; Hikal, N.A. Detecting and Locating Passive Video Forgery Based on Low Computational Complexity Third-Order Tensor Representation. J. Imaging 2021, 7, 47. [Google Scholar] [CrossRef] [PubMed]
- Sitara, K.; Mehtre, B. A comprehensive approach for exposing inter-frame video forgeries. In Proceedings of the 2017 IEEE 13th International Colloquium on Signal Processing & its Applications (CSPA), Penang, Malaysia, 10–12 March 2017; IEEE: Piscataway, NJ, USA, 2017. [Google Scholar]
- Bakas, J.; Naskar, R. A Digital Forensic Technique for Inter-Frame Video Forgery Detection Based on 3D CNN. In Proceedings of the International Conference on Information Systems Security, Funchal, Purtugal, 22–24 January 2018; Springer: Cham, Switzerland, 2018. [Google Scholar]
- Fadl, S.; Han, Q.; Li, Q. CNN spatiotemporal features and fusion for surveillance video forgery detection. Signal Process. Image Commun. 2021, 90, 116066. [Google Scholar] [CrossRef]
- Tyagi, S.; Yadav, D. A detailed analysis of image and video forgery detection techniques. Vis. Comput. 2023, 39, 813–833. [Google Scholar] [CrossRef]
- Long, C.; Smith, E.; Basharat, A.; Hoogs, A. A c3d-based convolutional neural network for frame dropping detection in a single video shot. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), Honolulu, HI, USA, 21–26 July 2017; IEEE: Piscataway, NJ, USA, 2017. [Google Scholar]
- Subramanyam, A.V.; Emmanuel, S. Pixel estimation based video forgery detection. In Proceedings of the 2013 IEEE International Conference on Acoustics, Speech and Signal Processing, Vancouver, BC, Canada, 26–31 May 2013; IEEE: Piscataway, NJ, USA, 2013. [Google Scholar]
- Qureshi, M.A.; Deriche, M. A bibliography of pixel-based blind image forgery detection techniques. Signal Process. Image Commun. 2015, 39, 46–74. [Google Scholar] [CrossRef]
- Fayyaz, M.A.; Anjum, A.; Ziauddin, S.; Khan, A.; Sarfaraz, A. An improved surveillance video forgery detection technique using sensor pattern noise and correlation of noise residues. Multimed. Tools Appl. 2020, 79, 5767–5788. [Google Scholar] [CrossRef]
- Kaur, H.; Jindal, N. Deep convolutional neural network for graphics forgery detection in video. Wirel. Pers. Commun. 2020, 112, 1763–1781. [Google Scholar] [CrossRef]
- Lin, G.-S.; Chang, J.-F.; Chuang, C.-H. Detecting frame duplication based on spatial and temporal analyses. In Proceedings of the 2011 6th International Conference on Computer Science & Education (ICCSE), Singapore, 3–5 August 2011; IEEE: Piscataway, NJ, USA, 2011. [Google Scholar]
- El-Shafai, W.; Fouda, M.A.; El-Rabaie, E.-S.M.; El-Salam, N.A. A comprehensive taxonomy on multimedia video forgery detection techniques: Challenges and novel trends. Multimed. Tools Appl. 2023, 83, 4241–4307. [Google Scholar] [CrossRef] [PubMed]
- Fadl, S.M.; Han, Q.; Li, Q. Authentication of surveillance videos: Detecting frame duplication based on residual frame. J. Forensic Sci. 2018, 63, 1099–1109. [Google Scholar] [CrossRef]
- Ulutas, G.; Ustubioglu, B.; Ulutas, M.; Nabiyev, V.V. Frame duplication detection based on BoW model. Multimed. Syst. 2017, 24, 549–567. [Google Scholar] [CrossRef]
- Bozkurt, I.; Ulutaş, G. Detection and localization of frame duplication using binary image template. Multimed. Tools Appl. 2023, 82, 31001–31034. [Google Scholar] [CrossRef]
- Singh, R.D.; Aggarwal, N. Video content authentication techniques: A comprehensive survey. Multimed. Syst. 2018, 24, 211–240. [Google Scholar] [CrossRef]
- Al-Sanjary, O.I.; Sulong, G. Detection of video forgery: A review of literature. J. Theor. Appl. Inf. Technol. 2015, 74, 208–220. [Google Scholar]
- Liao, S.-Y.; Huang, T.-Q. Video copy-move forgery detection and localization based on Tamura texture features. In Proceedings of the 6th International Congress on Image and Signal Processing (CISP), Hangzhou, China, 16–18 December 2013; IEEE: Piscataway, NJ, USA, 2013. [Google Scholar]
- Zhao, D.-N.; Wang, R.-K.; Lu, Z.-M. Inter-frame passive-blind forgery detection for video shot based on similarity analysis. Multimed. Tools Appl. 2018, 77, 25389–25408. [Google Scholar] [CrossRef]
- Bakas, J.; Naskar, R.; Dixit, R. Detection and localization of inter-frame video forgeries based on inconsistency in correlation distribution between Haralick coded frames. Multimed. Tools Appl. 2019, 78, 4905–4935. [Google Scholar] [CrossRef]
- Shelke, N.A.; Kasana, S.S. Multiple forgeries identification in digital video based on correlation consistency between entropy coded frames. Multimed. Syst. 2022, 28, 267–280. [Google Scholar] [CrossRef]
- Chen, S.; Tan, S.; Li, B.; Huang, J. Automatic detection of object-based forgery in advanced video. IEEE Trans. Circuits Syst. Video Technol. 2015, 26, 2138–2151. [Google Scholar] [CrossRef]
- Huang, C.C.; Lee, C.E.; Thing, V.L.L. A Novel Video Forgery Detection Model Based on Triangular Polarity Feature Classification. Int. J. Digit. Crime Forensics 2020, 12, 14–34. [Google Scholar] [CrossRef]
- Kumar, V.; Gaur, M. Multiple forgery detection in video using inter-frame correlation distance with dual-threshold. Multimed. Tools Appl. 2022, 81, 43979–43998. [Google Scholar] [CrossRef]
- Huang, T.; Zhang, X.; Huang, W.; Lin, L.; Su, W. A multi-channel approach through fusion of audio for detecting video inter-frame forgery. Comput. Secur. 2018, 77, 412–426. [Google Scholar] [CrossRef]
- Raskar, P.S.; Shah, S.K. VFDHSOG: Copy-move video forgery detection using histogram of second order gradients. Wirel. Pers. Commun. 2022, 122, 1617–1654. [Google Scholar] [CrossRef]
- Voronin, V.; Sizyakin, R.; Zelensky, A.; Nadykto, A.; Svirin, I. Detection of deleted frames on videos using a 3D convolutional neural network. In Proceedings of the Counterterrorism, Crime Fighting, Forensics, and Surveillance Technologies II, Berlin, Germany, 10–11 September 2018; SPIE: Bellingham, WA, USA, 2018. [Google Scholar]
- Ulutas, G.; Ustubioglu, B.; Ulutas, M.; Nabiyev, V. Frame duplication/mirroring detection method with binary features. IET Image Process. 2017, 11, 333–342. [Google Scholar] [CrossRef]
- Fadl, S.M.; Han, Q.; Li, Q. Inter-frame forgery detection based on differential energy of residue. IET Image Process. 2019, 13, 522–528. [Google Scholar] [CrossRef]
- Panchal, H.D.; Shah, H.B. Multiple forgery detection in digital video based on inconsistency in video quality assessment attributes. Multimed. Syst. 2023, 29, 2439–2454. [Google Scholar] [CrossRef]
- Hsiao, P.-Y.; Chou, S.-S.; Huang, F.-C. Generic 2-D gaussian smoothing filter for noisy image processing. In Proceedings of the TENCON 2007—2007 IEEE Region 10 Conference, Taipei, Taiwan, 30 October–2 November 2007; IEEE: Piscataway, NJ, USA, 2007. [Google Scholar]
- Hsiao, P.-Y.; Chen, C.-H.; Chou, S.-S.; Li, L.-T.; Chen, S.-J. A parameterizable digital-approximated 2D Gaussian smoothing filter for edge detection in noisy image. In Proceedings of the 2006 IEEE International Symposium on Circuits and Systems, Kos, Greece, 21–24 May 2006; IEEE: Piscataway, NJ, USA, 2006. [Google Scholar]
- Deng, G.; Cahill, L. An adaptive Gaussian filter for noise reduction and edge detection. In Proceedings of the 1993 IEEE Conference Record Nuclear Science Symposium and Medical Imaging Conference, San Francisco, CA, USA, 31 October–6 November 1993; IEEE: Piscataway, NJ, USA, 1993. [Google Scholar]
- Yang, J.; Hauptmann, A.G. Exploring temporal consistency for video analysis and retrieval. In Proceedings of the 8th ACM International Workshop on Multimedia Information Retrieval, Santa Barbara, CA, USA, 26–27 October 2006. [Google Scholar]
- Song, K.-C.; Yan, Y.-H.; Chen, W.-H.; Zhang, X. Research and perspective on local binary pattern. Acta Autom. Sin. 2013, 39, 730–744. [Google Scholar] [CrossRef]
- Mahale, V.H.; Ali, M.M.; Yannawar, P.L.; Gaikwad, A.T. Image inconsistency detection using local binary pattern (LBP). Procedia Comput. Sci. 2017, 115, 501–508. [Google Scholar] [CrossRef]
- Gaikwad, A.; Mahale, V.; Ali, M.M.; Yannawar, P.L. Detection and Analysis of Video Inconsistency Based on Local Binary Pattern (LBP). In Proceedings of the International Conference on Recent Trends in Image Processing and Pattern Recognition, Solapur, India, 21–22 December 2018; Springer: Cham, Switzerland, 2018. [Google Scholar]
- Bourouis, S.; Alroobaea, R.; Alharbi, A.M.; Andejany, M.; Rubaiee, S. Recent advances in digital multimedia tampering detection for forensics analysis. Symmetry 2020, 12, 1811. [Google Scholar] [CrossRef]
- Joy, S.; Kurian, L. Video Forgery Detection Using Invariance of Color Correlation. Int. J. Comput. Sci. Mob. Comput. 2014, 3, 99–105. [Google Scholar]
- Fadl, S.; Han, Q.; Qiong, L. Exposing video inter-frame forgery via histogram of oriented gradients and motion energy image. Multidimens. Syst. Signal Process. 2020, 31, 1365–1384. [Google Scholar] [CrossRef]
- Lucas, B.D.; Kanade, T. An iterative image registration technique with an application to stereo vision. In Proceedings of the IJCAI’81: 7th International Joint Conference on Artificial Intelligence, Vancouver, BC, Canada, 24–28 August 1981. [Google Scholar]
- Bruhn, A.; Weickert, J.; Schnörr, C. Lucas/Kanade meets Horn/Schunck: Combining local and global optic flow methods. Int. J. Comput. Vis. 2005, 61, 211–231. [Google Scholar] [CrossRef]
- Panchal, H.D.; Shah, H.B. Video tampering dataset development in temporal domain for video forgery authentication. Multimed. Tools Appl. 2020, 79, 24553–24577. [Google Scholar] [CrossRef]
- Shelke, N.A.; Kasana, S.S. Multiple forgery detection and localization technique for digital video using PCT and NBAP. Multimed. Tools Appl. 2022, 81, 22731–22759. [Google Scholar] [CrossRef]
- Sitara, K.; Mehtre, B. Detection of inter-frame forgeries in digital videos. Forensic Sci. Int. 2018, 289, 186–206. [Google Scholar] [CrossRef]
Reference | # Videos | Test Videos | # Frames Duplicated | Resolution | Frame Rate | Scenario |
---|---|---|---|---|---|---|
Ulutas et al. [38] | 31 | 5 | 20, 30, 40, 50, 55, 60, 70, 80 | 320 × 240 | 29.97, 30 | - |
Jia et al. [15] | 115 | - | 10, 20, 40 | 320 × 240 | 29.97, 30 | - |
Ulutas et al. [52] | 10 | 10 | 20, 30, 40, 50, 55, 60, 70, 80 | 320 × 240 | 29.97, 30 | - |
Fadl et al. [28] | FD: 62 + FI: 287 | FD: 12 + FI: 57 | 10 to 600 | 720 ×1280, 240 × 320, 288 × 352, 576 × 704 | 23.98 to 30 | - |
CSVTED | FD: 225 + FI: 225 | FD: 45 + FI: 45 | 10, 15, 20, 25, 30, 35, 40, 45, 50 | 640 × 360, 640 × 480, 1920 × 1080, 1280 × 720, | 12.50, 15, 25, 29.97, 30 | Morning, Evening, Night, and Fog |
Method | PR (%) | RR (%) | DA (%) |
---|---|---|---|
Proposed method with MR–LBP features only | 95.92 | 99.90 | 95.93 |
Proposed method with OF–based features only | 99.84 | 99.84 | 99.69 |
Proposed Model 1 (MR–LBP with standard deviation) | 99.23 | 99.89 | 99.10 |
Proposed Model 2 (MR–LBP with OF aggregation) | 99.81 | 99.90 | 99.71 |
Method | Dataset | PR | RR | DA |
---|---|---|---|---|
Proposed Model 1 (MR–LBP with standard deviation) | Ulutas Dataset [38] | 99.67 | 99.92 | 99.59 |
Panchal Dataset [67] | 99.61 | 99.95 | 99.57 | |
Proposed Model 2 (MR–LBP with OF aggregation) | Ulutas Dataset [38] | 99.83 | 99.92 | 99.75 |
Panchal Dataset [67] | 99.75 | 99.89 | 99.64 |
Method | PR (%) | RR (%) | F1 (%) | DA (%) |
---|---|---|---|---|
Proposed method with MR–LBP features only | 25.35 | 100 | 40.45 | 97.29 |
Proposed method with OF–based features only | 82.22 | 82.22 | 82.22 | 99.67 |
Proposed model 1 (MR–LBP with standard deviation) | 73.78 | 100 | 84.91 | 99.67 |
Proposed model 2 (MR–LBP with OF aggregation) | 91.40 | 94.4 | 92.88 | 99.87 |
Method | Dataset | PR (%) | RR (%) | F1 (%) | DA (%) |
---|---|---|---|---|---|
Proposed model 1 (MR–LBP with standard deviation) | Panchal Dataset [67] | 93.75 | 100 | 96.77 | 99.97 |
Proposed model 2 (MR–LBP with OF aggregation) | Panchal Dataset [67] | 100 | 94.64 | 97.25 | 99.98 |
Frame Duplication | Frame Insertion | |||
---|---|---|---|---|
Methods | Time per Frame (s) | Time per Pixel (µs) | Time per Frame (s) | Time per Pixel (µs) |
Ulutas et al. [38] | 0.2 | 2.6 | × | × |
Ulutas et al. [52] | 0.01 | × | × | × |
Alsakar et al. [25] | × | × | 10.75 | × |
Jia et al. [15] | × | 1.623 | × | × |
Akhtar et al. [19] | × | × | 0.227 | × |
Proposed model 1 (MR–LBP with standard deviation) | 3.6 | 3.3 | 3.39 | 3.1 |
Proposed model 2 (MR–LBP with OF aggregation) | 3.8 | 3.6 | 3.7 | 3.4 |
Method | Evaluation | Cross–Validation | |||||
---|---|---|---|---|---|---|---|
Size of Dataset | FD | FI | FD | FI | |||
DA | F1 | DA | F1 | DA | DA | ||
Ulutas et al. [52] | 10 | 99.35 | 99.64 | × | × | × | × |
Ulutas et al. [38] | 31 | 96.73 | 97.79 | × | × | × | × |
Sitara et al. [69] | 90 | 94.5 | × | × | × | × | × |
Jia et al. [15] | 115 | 98 | 98.5 | × | × | × | × |
Kharat et al. [12] | 20 | 99.7 | 99.82 | × | × | × | × |
Bozkurt et al. [39] | 13 | 98.59 | × | × | × | × | × |
Alsakar et al. [25] (HARRIS features) | 18 | × | × | × | 63 | × | × |
Alsakar et al. [25] (GLCM features) | 18 | × | × | × | 67 | × | × |
Alsakar et al. [25] (SVD features) | 18 | × | × | × | 95 | × | × |
Fadl et al. [28] | 349 | 98.5 | × | 99.9 | × | × | |
Shelke and Kasana [68] | 100 | 98.56 | 96 | 98.28 | 97.05 | × | × |
Akhtar et al. [19] | 2555 | × | × | 98.98 | 98.87 | × | × |
Proposed | 450 | 99.71 | 99.85 | 99.87 | 92.88 | 99.75 | 99.98 |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2024 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Akhtar, N.; Hussain, M.; Habib, Z. Two–Stage Detection and Localization of Inter–Frame Tampering in Surveillance Videos Using Texture and Optical Flow. Mathematics 2024, 12, 3482. https://doi.org/10.3390/math12223482
Akhtar N, Hussain M, Habib Z. Two–Stage Detection and Localization of Inter–Frame Tampering in Surveillance Videos Using Texture and Optical Flow. Mathematics. 2024; 12(22):3482. https://doi.org/10.3390/math12223482
Chicago/Turabian StyleAkhtar, Naheed, Muhammad Hussain, and Zulfiqar Habib. 2024. "Two–Stage Detection and Localization of Inter–Frame Tampering in Surveillance Videos Using Texture and Optical Flow" Mathematics 12, no. 22: 3482. https://doi.org/10.3390/math12223482
APA StyleAkhtar, N., Hussain, M., & Habib, Z. (2024). Two–Stage Detection and Localization of Inter–Frame Tampering in Surveillance Videos Using Texture and Optical Flow. Mathematics, 12(22), 3482. https://doi.org/10.3390/math12223482