Next Article in Journal
Remote Blood Pressure Estimation via the Spatiotemporal Mapping of Facial Videos
Next Article in Special Issue
FSVM: A Few-Shot Threat Detection Method for X-ray Security Images
Previous Article in Journal
Design Consideration, Numerical and Experimental Analyses of Mode-Division-Multiplexed (MDM) Silicon Photonics Integrated Circuit with Sharp Bends
Previous Article in Special Issue
Behavior-Based Video Summarization System for Dog Health and Welfare Monitoring
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

S-BIRD: A Novel Critical Multi-Class Imagery Dataset for Sewer Monitoring and Maintenance Systems

by
Ravindra R. Patil
1,*,
Mohamad Y. Mustafa
1,
Rajnish Kaur Calay
1 and
Saniya M. Ansari
2
1
Faculty of Engineering Science and Technology, UiT The Arctic University of Norway, 8514 Narvik, Norway
2
Department of E & TC Engineering, Ajeenkya D Y Patil School of Engineering, Pune 411047, India
*
Author to whom correspondence should be addressed.
Sensors 2023, 23(6), 2966; https://doi.org/10.3390/s23062966
Submission received: 30 January 2023 / Revised: 25 February 2023 / Accepted: 7 March 2023 / Published: 9 March 2023
(This article belongs to the Special Issue Artificial Intelligence in Computer Vision: Methods and Applications)

Abstract

:
Computer vision in consideration of automated and robotic systems has come up as a steady and robust platform in sewer maintenance and cleaning tasks. The AI revolution has enhanced the ability of computer vision and is being used to detect problems with underground sewer pipes, such as blockages and damages. A large amount of appropriate, validated, and labeled imagery data is always a key requirement for learning AI-based detection models to generate the desired outcomes. In this paper, a new imagery dataset S-BIRD (Sewer-Blockages Imagery Recognition Dataset) is presented to draw attention to the predominant sewers’ blockages issue caused by grease, plastic and tree roots. The need for the S-BIRD dataset and various parameters such as its strength, performance, consistency and feasibility have been considered and analyzed for real-time detection tasks. The YOLOX object detection model has been trained to prove the consistency and viability of the S-BIRD dataset. It also specified how the presented dataset will be used in an embedded vision-based robotic system to detect and remove sewer blockages in real-time. The outcomes of an individual survey conducted at a typical mid-size city in a developing country, Pune, India, give ground for the necessity of the presented work.

1. Introduction

An underground sewerage system is an essential feature of town planning as it transports the wastewater away from its source for safe disposal in the environment with minimum impact on the surroundings. However, underground pipe systems have maintenance problems. Sewer blockages and various damages such as cracks, fractures, joint displacement, etc. all can cause overflow, leaching of sewage into soil and interference with drinking water supply lines. Poor maintenance also leads sewer pipes to deteriorate early.
Therefore, it is important for any responsible authority to ensure that sewers are in good condition and run properly. The Ministry of Housing and Urban Affairs conferred Standard Operating Procedure (SOP) for cleaning sewers and septic tanks in November 2018 [1]. Regular inspections are necessary to identify any event of crack or blockage so that corrective measures are taken in time to avoid a crisis. In the past, manual inspection was often used followed by circuit television (CCTV) which has been one of the most used methods in the US and European municipalities in recent decades. However, these methods are labor-intensive and error-prone.
Artificial Intelligence (AI) is used in computer vision technology that consists of intelligent algorithms to interpret meaningful digital information from images and videos, which, when combined with automated robotic systems, provide powerful vision and intelligence to detect various sewer problems and to plan corrective actions. However, training AI-based Deep Neural Object Detection Models and achieving sewer inspection objectives based on them requires large amounts of appropriate and labeled data. A dataset is a collection of featured and significant information in any field that is used to learn AI models for purposes such as detection, classification, regression, clustering, segmentation, etc. Data is usually in the form of images, text, numbers, time series, graphs, etc. The performance of the best detection model trained using a poor dataset is always inferior to the performance of a poor detection model trained using a highly featured and quality dataset. At the center of every object detector, whether single-stage or two-stage, is a classifier that secures the identities of all desired object classes. Clearly, the accuracy rate and performance of any detection model are highly dependent on the quality of the input imagery dataset.
Therefore, relevant dataset collection is a very important prerequisite for any AI model to predict outcomes with the desired accuracy and also has emerged as a prominent research theme in respective research communities. This involves data acquisition or collection, appropriately labeling the data and finally enhancement of obtainable data or models [2]. Due to the open-access research policy of many funding agencies, a large amount of data pertaining to many fields is available on various platforms. In many instances data may be available from data-sharing platforms like DataHub [3], Kaggle datasets [4], Mendeley Data [5], etc. and data searching platforms like Google Dataset Search [6], IEEE DataPort [7], etc. After tackling several challenges in data search, a researcher can succeed in obtaining the required dataset [8]. However, the European Commission recognized the difficulties in obtaining and tracing open data in 2011 and started to regulate data publishing activities in Europe [9]. Six snags in obtaining and tracing open data were identified: deficient details about the existence and accessibility of data, ambiguity about data ownership by public authorities, ambiguity about reuse terms, critical nature and cost of data, complex licensing processes and restrictive fees, specific reuse agreements with commercial members and reuse restrictions for state-owned companies.
Specifically, data acquisition includes tasks such as searching, augmenting and generating as needed, and in our case, the dataset is not only created due to unavailability but also prepossessed, augmented and labeled individually for classification and detection tasks. Manual or automated techniques are used for dataset generation, while synthetic data is generated to fill the lacking portion of the dataset. A standardized or benchmark dataset is always a central aspect to obtain the best-fit learning models and the application of transfer learning techniques with the developed dataset plays an important role in the advancement of AI-based models [10]. In computer vision, a dataset of digital images containing object class information is grouped as needed into a training set, validation set, and test set to serve as input to a detection model for learning, evaluation, and testing purposes, respectively. A workflow with decision-making for the S-BIRD dataset presented in this paper is shown in Figure 1, which displays the process from generation requirements to the training results.
In this paper, a new critical multi-class imagery dataset S-BIRD (Sewer-Blockages Imagery Recognition Dataset) is presented to identify sewer blockages caused by grease, plastics and tree roots. The lack of a standardized matrix for algorithms applied in the real-world development of sewer monitoring and maintenance systems is a critical issue, and the submitted dataset addresses this. So, the S-BIRD sets the standard for detection outcomes in real-time scenarios. Validation results of the S-BIRD dataset are given and development on an embedded vision platform to overcome actual sewer blockages problem is considered. In the conferred work, all computer vision and model training operations are implemented using Python programming, OpenCV, PyTorch framework, and some other machine learning libraries on the DGX workstation system including the Linux platform. Both the presented dataset and the corresponding results highlight the importance and necessity of such research work for the treatment of wastewater sewer blockages.

2. Needs of the S-BIRD Dataset

In earlier work, a survey on sewer robotic systems and computer vision practices in sewer inspection works was carried out and that gave information about practical issues concerning sewerage systems under the Pune Municipal Corporation (PMC), India [11]. It was concluded that sewer blockage is the main issue of sewers in Pune and to date, there is no robust algorithm and robotic system available for both real-time detection and removal of sewer pipe blockages.
Unlike many Western countries, India has single sewer lines for both sewage and stormwater. Thus, this combined drainage system is a big problem, particularly for cleaning and removing blockages.
In order to develop the function of detecting and identifying sewer blockages in real time, authenticated datasets are a prerequisite. Thus, all available means were used to search for datasets. Several municipalities and various authorities were also contacted for relevant data information, but no concrete work and datasets that may be used for real-time detection of sewer blockages were available. Furthermore, it was not possible to acquire a specific dataset for Indian conditions focusing on the issue of sewer blockages. The harmful, unhygienic and foul smell of a sewer environment is always a major concern when capturing frames of sewer problems for dataset generation. It is appropriate to imply that independent binding, copyright or confidentiality issues relating to earlier works are also responsible for the unavailability of the datasets.
Sewer blockages are mainly caused by grease, plastic and tree roots. Other elements inside the sewer mix up with the black water and become difficult to identify. So, other elements are usually treated as a blackish sewer blockage, which is identified as black grease in the dataset. We also considered imagery data of grease, plastic and tree roots as mentioned above in the dataset S-BIRD, which is used for training of object detection model to locate and recognize the sewer blockages in real-time.
Obviously, blind systems cannot be as efficient as vision-based sewer robotic systems. Figure 2 shows the concept of constructing the S-BIRD dataset that takes grease, plastic and tree roots into account.

3. Tools in S-BIRD Dataset Creation

In this section, the tools involved in creating the S-BIRD dataset are provided for detailed viewing.

3.1. Sewer Pipeline

In an unhygienic, muddy and smelly sewer pipe environment due to sewage, toiletry, sanitation, and stormwater from combined drainage systems, capturing real-time frames of sewer issues was a very difficult task for an individual. For simulating a sewer network, PVC pipelines of 200 mm diameter, which are widely used in residential sewers, were used to construct a typical sewer network. The constructed sewer pipeline is shown in Figure 3.
In this case, there is no big difference between a real sewer environment and a laboratory setup or simulated sewer network. Exactly the same blockage types with inherent nature have been created inside the sewer network consisting of all featured information. The only difference was that the simulated sewer network did not have the stench and noxious atmosphere. The detection model trained using the developed S-BIRD dataset in the respective sewer network is capable to work in practical situations.

3.2. Sewer Inspection Camera

Real-time frames of sewer barriers that include grease, plastics, and tree roots are captured by the watertight sewer camera shown in Figure 4, and its characteristics are given in Table 1.
This camera sensor is capable of capturing real-time frames at different angles not only for the intended aspect ratio but also for varying brightness due to attached modifiable white LEDs.

4. A Novel S-BIRD and Corresponding Results

This section discusses compiled imagery data (Section 4.1), its arithmetic details (Section 4.2), preprocessing and augmentation techniques applied to captured frames (Section 4.3), and annotated heatmap and object count histograms (Section 4.4).

4.1. Imagery Data Collection

All images of sewer blockages are captured under different lighting conditions and from different angles to gather the necessary perceptions and features. Figure 5 reveals some blockage frames of tree roots in the newly created dataset.
Dissimilar colored plastic is captured in the picture and key information for the detection and recognition task is achieved as shown in Figure 6.
There could be other elements within the black sewage mass such as plastic bags or other debris, but they look completely blackish as they are often mixed with black water and grease.
Figure 7 exhibits grease blockage frames capturing diverse and significant colored information. There are a number of sources for grease-type sewer blockages which mainly include wastage from domestic and high- or low-density production plants that produce huge chemical and processed waste.

4.2. Arithmetic Details of Captured Frames

The arithmetic details of the captured frames are listed in Table 2 for further implementation. Certainly, annotating the objects in each captured frame was time-consuming but the task was still performed individually with high skill and accuracy without labeling errors. The annotations contain information about the location, i.e., center x, center y, width, height and class of objects present in each frame of the S-BIRD dataset.
Figure 8 stipulates the total number of annotations for class balance, i.e., annotations for each sewer block type and these are 4131 for grease, 3471 for tree roots and 2631 for plastic.
The location of annotations, i.e., bounding boxes for considered blockage types in all captured frames is shown by heatmap in Figure 9. A heatmap represents informative data in a graphical or two-dimensional form where a color-coding system is used to represent values, and in the above heatmap, values are annotation details. It confers a quick visible summary to perceive the intricate nature of the dataset. Here, the correlation between annotated values is made easier to understand using colors in a heatmap compared to numerical tables. The yellow color denotes a highly positioned region of annotations whereas the light green color indicates lower positioning. All depicted heatmaps show that the locations of annotations are mostly in the center of the frames of object classes.
The imagery data is balanced into three groups such as training data with 4928 frames (70%), validation data with 1408 frames (20%) and testing data with 704 frames (10%) as shown in Figure 10.
Table 3 provides annotation details for the classes in the training data.

4.3. Preprocessing and Augmentation Techniques

Here, two preprocessing techniques have been implemented on captured frames such as auto-orientation of pixel data, i.e., discarding the EXIF rotation and validating the pixel sort as well as resizing to 416 × 416 (px) by stretching the frame without losing source frame information. An image preprocessing benefits to reduce model training time and speed up inference of detection models.
Here, two preprocessing techniques have been implemented on captured frames such as auto-orientation of pixel data, i.e., discarding the EXIF rotation and validating the pixel sort as well as resizing to 416 × 416 (px) by stretching the frame without losing source frame information. Image preprocessing benefits from reduced model training time and sped-up inference of detection models.
Figure 11 shows the aspect ratio distribution graph for the S-BIRD dataset and makes clear that all frames are 416 × 416 (px), i.e., square in size.
Further, image-level augmentation techniques have also been implemented to generate new training instances from existing training data.
Figure 12a shows the output frame of the gray scaling applied 25 percent to the input training frame which helps to increase the training variation but does not remove the color information when making inferences. Salt and pepper noise, also known as impulse noise, is applied to 5 percent of the pixels of the input frames as shown in Figure 12b which helps the detection model to turn out to be more flexible for camera artifacts through training. This noise involves adding some bright pixels to dark regions and some dark pixels to bright regions of the frames. It also helps to prevent adverse effects and avoid overfitting.
To strengthen the detection model against light and camera setting changes, random exposure adaptations were instigated between −25 and +25 percent for the input frame as shown in Figure 12c.
Two advanced augmentation techniques, namely cutout and mosaic, were exploited as shown in Figure 13a and Figure 13b, respectively. Adding cutouts to training frames is extremely useful for the detection model to be strong against the object occlusion state. For this, three cutouts were inserted in 10 percent of each of the total sizes of the input frames. Next, the mosaic technique helps the detection model to work well on small objects by joining several images from the training set in collage [12]. In this, four different sewer block frames were added in a single frame.
Augmentation techniques facilitate enhancing the efficiency of the object detection model by increasing the number and variegation of learning instances and related annotations. These techniques also reduce training time and costs for search detection models. So, discrete output versions have been generated for source frames.
In Table 4, the numerical details of training frames in S-BIRD are demonstrated after applying preprocessing and augmentation techniques.
The graph in Figure 14 shows the escalated annotations for each sewer block type in S-BIRD’s training data, after using annotation techniques. Now there are 26,847 annotations for grease, 21,553 for tree roots and 20,661 for plastics making a total of 69,061 augmented annotations, i.e., bounding boxes. Total annotations have increased by 61,865, i.e., 859.714%. Both preprocessing and augmentation techniques have been implemented using OpenCV, a computer vision and machine learning library, along with Python programming on the Linux platform from scratch to achieve the desired results.

4.4. Annotated Heatmap and Object Count Histogram

Two important parameters, namely the annotated heatmap and the object count histogram have been examined to assess the efficacy of the training data. The location of the entire annotations for grease, plastic and tree roots in S-BIRD’s training data is illustrated by heatmaps in Figure 15. The specified heatmap informs us of the utmost generic position and weightage of all the annotations for revealed classes. From the color information of the heatmaps, it can be seen that most of the annotation locations are at the far left and right of both the top and bottom sides of the frames of object classes.
A histogram is a chart that plots numeric data into bins represented by individual columns. Figure 16 details the number of frames on the y-axis and bins, i.e., the number of corresponding objects for all classes on x-axis, with the help of the object count histogram.
The number of objects, i.e., annotations for both grease and tree roots blocks are up to nine shreds as shown in Figure 16a,b. There is obviously one grease object for 1730 frames andfour to five grease objects for 1400 to 1600 frames as given in Figure 16a. In total, 1926 frames contain a single tree root object and about 1500 frames contained three to four tree root objects as shown in Figure 16b. The number of plastic objects varies up to seven shreds as shown in Figure 16c in which four plastic objects are in 2494 frames and perceptibly one plastic object in about 2200 frames.
Figure 16d represents the object count histogram of all classes where 11,339 frames contain four to five objects. It also shows details for a much lower aggregate overall for a single object in frames as compared to the ratio for 69,061 annotations. The findings obtained for both parameters such as the annotated heatmap and the object count histogram prove the high veracity and standard for each imagery data class in S-BIRD.

5. Training of Object Detection Model

5.1. Insight on Conformation of Object Detector Models

Ordinarily, object detectors have two important segments, the backbone with pretraining to extract the features of input frames and the head which utilizes feature maps to predict classes and bounding boxes. Some layers are placed between the backbone and the head of recent object detectors to collect feature maps from distinct phases known as the neck. Object detectors with a backbone and densely predicted head are known as single-stage detectors, such as YOLO and SSD, while two-stage detectors have a backbone and head with dense and sparse predictions such as R-FCN, Faster R-CNN as shown in Figure 17. However, since single-stage detectors are faster than two stage detectors, they are used for multifarious real-time embedded applications. These object detectors embedded in robotic artifices are utilized to detect various faults in the sewerage system [13,14].
Table 5 lists some instances of the conformation parts in the object detector models.
The popular one-stage YOLO detection model is constantly being improved for better performance. An advanced version of the YOLO detection model is the recently introduced YOLOX which comprises three different basic embarkations, such as (a) anchor-free design which uses a center-based approach with each pixel detection mechanism for the selection of just one positive instance which then estimate four distances such as left, top, right, and bottom from positives to the border, i.e., prediction consists of a single 4D vector to encode the location of the bounding box at every foreground pixel, (b) decoupled head for classification and regression, and (c) advanced label allocation tactics namely SimOTA which lessen the training time and evade other clarifier hyperparameters in the SinkhornKnopp algorithm, making it faster and more efficient than its equivalents [19]. The performance of YOLOX has been improved with addition of mosaic and mixup augmentation. YOLOv3 and Spatial Pyramid Pooling (SPP) layers with Darknet53 are employed as baseline by YOLOX. This detection model of different sizes has attained consistent improvements against all compatible counterparts when tested on modified CSPNet backbone in addition to the Darknet53 backbone.

5.2. Training of YOLOX Using S-BIRD

So, the small YOLOX detection model in PyTorch framework allowing mobile deployment has been trained to detect the main types of sewer blockages such as grease, plastic and tree roots using the newly developed S-BIRD. Annotations for sewer block types in S-BIRD were implemented in Pascal VOC format as per the requirement to advance the training process. The Tesla V100-DGXS-32GB GPU workstation was used as a training platform via Docker Container with a defined image.
Table 6 makes available particulars on crucial traits in the YOLOX-s training process.
The results obtained for the timing and precision of the YOLOX-s trained model for S-BIRD are given in Table 7 and Table 8, respectively.
From Table 7 and Figure 18, YOLOX-s has achieved 90.04% AP for grease blocks, 90.81% AP for plastic blocks, 89.30% AP for tree root blocks, and 90.05% mean-AP computed at IoU (Intersection over Union) threshold 0.5. Another m-AP calculated over different IoU thresholds, from 0.5 to 0.95 with a step of 0.05 is 78.85%. The best-fit model is selected using cross-validation or rotation estimation technique [20]. The visual upshots of precisely detected sewer blocks such as tree roots, plastic and grease, are delineated in Figure 19. Of course, multiple sewer blockages in the same frame have also been considered for real-time detection purposes. Overall, the obtained results of the YOLOX-trained model prove the consistency and viability of the new S-BIRD dataset presented.

5.3. Embedded Vision with S-BIRD

The embedded vision is a pioneering and comprehensive platform for real-world visual implementations in the areas of home life equipment, health, daily services, security through detection and tracking, etc. [21,22]. So, the object detection model trained using S-BIRD will be a significant addition to existing or newly developed embedded vision-based sewer robotic systems.
PIRAT [23], KARO [24], KURT [25], MAKRO [26], KANTARO [27], SIAR [28], etc. are some of the popular developments in the field of sewer robotics that serve the purpose of sewer inspection. Figure 20 shows the block diagram of an automated system that has a power-driven cutting tool to remove sewer blocks located by a detector trained using S-BIRD.
Here, Jetson nano has been selected as the embedded platform having a 4 GB GPU card of 128-Cuda cores and is suitable for running deep neural-network-based object detector models and for processing contiguous frames in real-time. Cameras such as a webcam, arducam, or raspicam are used to capture the surrounding frames for the purpose of navigation and processing, and then the output frames of detected sewer blockages are displayed on the screen to a remote location as shown in Figure 21.
In order to solve the recurring problem of underground sewer barriers in the practical world, a smart and comprehensive vision-based automation system with an AI detector trained using S-BIRD is certainly capable of meeting the needs of responsible authorities of any country.

6. Conclusions

In conclusion, a new critical multi-class imagery dataset S-BIRD which includes frames of main sewer blocks such as grease, plastic and tree roots is introduced to fulfill the need for implementing computer vision to automated robotic systems for identifying blockages in the sewerage pipes.
Arithmetic details for both compiled, as well as preprocessed and augmented data are discussed. The obtained results for preprocessing and augmentation demonstrate the increased number and variegation of learning instances and related annotations for the efficient performance of the object detection model. The procured details of heatmaps and object count histograms prove the high strength, veracity and standard for each imagery data class in S-BIRD.
The trained small YOLOX model achieved 90.04% AP for grease blocks, 90.81% AP for plastic blocks, 89.30% AP for tree root blocks, 90.05% Mean-AP at 0.5 IoU threshold, and 78.85% Mean-AP at 0.5 to 0.95 IoU thresholds for 300 epochs using S-BIRD. The relevant outcomes prove the consistency and viability of the new S-BIRD dataset presented. The object detectors trained using the presented S-BIRD will be a valuable addition to the existing or newly developed embedded vision-based sewer monitoring and maintenance systems for detecting sewer blockages in real-time scenarios.

Author Contributions

Conceptualization, R.R.P. and M.Y.M.; methodology, R.R.P.; software, R.R.P.; Dataset Creation, R.R.P.; validation, R.R.P. and M.Y.M.; formal analysis, R.R.P. and M.Y.M.; investigation, R.R.P., M.Y.M., R.K.C. and S.M.A.; writing—original draft preparation, R.R.P.; writing—review and editing, M.Y.M., R.K.C. and R.R.P.; visualization, M.Y.M. and R.R.P.; project administration, R.K.C., M.Y.M. and S.M.A.; funding acquisition, R.K.C. and M.Y.M. All authors have read and agreed to the published version of the manuscript.

Funding

The publication charges for this article have been funded by a grant from the publication fund of UiT the Arctic University of Norway.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The research data will be made available on the request.

Acknowledgments

Authors acknowledge the support from SPRING Eu-India Project and UiT the Arctic University of Norway, Narvik, Norway for PhD studies of Ravindra R. Patil (No. 821423 and GOI No. BT/IN/EU-WR/60/SP/2018). Thanks are due to the Department of Computer Engineering and IT, COEP Technological University (COEP Tech) for providing the high computing GPU server facility procured under TEQIP-III (A world bank project) for our research work.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Information Manual—Standard Operating Procedure (SOP) for Cleaning of Sewers and Septic Tanks by Central Public Health & Environmental Engineering Organization (CPHEEO), Ministry of Housing and Urban Affairs, Government of India. Available online: http://cpheeo.gov.in/upload/5c0a062b23e94SOPforcleaningofSewersSepticTanks.pdf (accessed on 28 January 2023).
  2. Roh, Y.; Heo, G.; Whang, S.E. A survey on data collection for machine learning: A big data-ai integration perspective. IEEE Trans. Knowl. Data Eng. 2019, 33, 1328–1347. [Google Scholar] [CrossRef] [Green Version]
  3. Bhardwaj, A.; Karger, D.; Subramanyam, H.; Deshpande, A.; Madden, S.; Wu, E.; Elmore, A.; Parameswaran, A.; Zhang, R. Collaborative data analytics with DataHub. Proc. VLDB Endow. 2015, 8, 1916. [Google Scholar] [CrossRef] [PubMed]
  4. Kaggle. Available online: https://www.kaggle.com/ (accessed on 22 January 2023).
  5. Mendeley Data. Available online: https://data.mendeley.com/ (accessed on 18 January 2023).
  6. Google Dataset Search. Available online: https://datasetsearch.research.google.com/ (accessed on 12 January 2023).
  7. IEEE DataPort. Available online: https://ieee-dataport.org/dataset (accessed on 9 January 2023).
  8. Chapman, A.; Simperl, E.; Koesten, L.; Konstantinidis, G.; Ibáñez, L.D.; Kacprzak, E.; Groth, P. Dataset search: A survey. VLDB J. 2020, 29, 251–272. [Google Scholar] [CrossRef] [Green Version]
  9. European Commission. Digital Agenda: Commission’s Open Data Strategy, Questions and Answers. Memo/11/891. 12 December 2011. Available online: https://ec.europa.eu/commission/presscorner/detail/en/MEMO_11_891 (accessed on 4 January 2023).
  10. Weiss, K.; Khoshgoftaar, T.M.; Wang, D. A survey of transfer learning. J. Big Data 2016, 3, 1–40. [Google Scholar] [CrossRef] [Green Version]
  11. Patil, R.R.; Ansari, S.M.; Calay, R.K.; Mustafa, M.Y. Review of the State-of-the-art Sewer Monitoring and Maintenance Systems Pune Municipal Corporation-A Case Study. TEM J. 2021, 10, 1500–1508. [Google Scholar] [CrossRef]
  12. Bochkovskiy, A.; Wang, C.Y.; Liao, H.Y.M. Yolov4: Optimal speed and accuracy of object detection. arXiv 2020, arXiv:2004.10934. [Google Scholar]
  13. Cheng, J.C.; Wang, M. Automated detection of sewer pipe defects in closed-circuit television images using deep learning techniques. Autom. Constr. 2018, 95, 155–171. [Google Scholar] [CrossRef]
  14. Kumar, S.S.; Wang, M.; Abraham, D.M.; Jahanshahi, M.R.; Iseley, T.; Cheng, J.C. Deep learning–Based automated detection of sewer defects in CCTV videos. J. Comput. Civ. Eng. 2020, 34, 04019047. [Google Scholar] [CrossRef]
  15. Wang, C.Y.; Liao, H.Y.M.; Wu, Y.H.; Chen, P.Y.; Hsieh, J.W.; Yeh, I.H. CSPNet: A new backbone that can enhance learning capability of CNN. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), Seattle, WA, USA, 14–19 June 2020; pp. 1571–1580. [Google Scholar]
  16. Redmon, J.; Farhadi, A. Yolov3: An incremental improvement. arXiv 2018, arXiv:1804.02767. [Google Scholar]
  17. Redmon, J.; Divvala, S.; Girshick, R.; Farhadi, A. You only look once: Unified, real-time object detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 779–788. [Google Scholar]
  18. Ren, S.; He, K.; Girshick, R.; Sun, J. Faster r-cnn: Towards real-time object detection with region proposal networks. In Proceedings of the 28th International Conference on Neural Information Processing Systems—Volume 1 (NIPS’15), Montreal, QC, Canada, 7–12 December 2015; MIT Press: Cambridge, MA, USA, 2015; pp. 91–99. [Google Scholar]
  19. Ge, Z.; Liu, S.; Wang, F.; Li, Z.; Sun, J. Yolox: Exceeding yolo series in 2021. arXiv 2021, arXiv:2107.08430. [Google Scholar]
  20. Berrar, D. Cross-Validation. In Encyclopedia of Bioinformatics and Computational Biology; Ranganathan, S., Gribskov, M., Nakai, K., Schönbach, C., Eds.; Academic Press: Oxford, UK, 2019; pp. 542–545. [Google Scholar] [CrossRef]
  21. Vaidya, O.S.; Patil, R.; Phade, G.M.; Gandhe, S.T. Embedded Vision Based Cost Effective Tele-operating Smart Robot. Int. J. Innov. Technol. Explor. Eng. (IJITEE) 2019, 8, 1544–1550. [Google Scholar]
  22. Patil, R.R.; Vaidya, O.S.; Phade, G.M.; Gandhe, S.T. Qualified Scrutiny for Real-Time Object Tracking Framework. Int. J. Emerg. Technol. 2020, 11, 313–319. [Google Scholar]
  23. Kirkham, R.; Kearney, P.D.; Rogers, K.J.; Mashford, J. PIRAT—A system for quantitative sewer pipe assessment. Int. J. Robot. Res. 2000, 19, 1033–1053. [Google Scholar] [CrossRef]
  24. Kuntze, H.B.; Schmidt, D.; Haffner, H.; Loh, M. KARO-A flexible robot for smart sensor-based sewer inspection. In Proceedings of the International Conference No Dig, Dresden, Germany, 22 September 1995; Volume 95, pp. 367–374. [Google Scholar]
  25. Kirchner, F.; Hertzberg, J. A prototype study of an autonomous robot platform for sewerage system maintenance. Auton. Robot. 1997, 4, 319–331. [Google Scholar] [CrossRef]
  26. Rome, E.; Hertzberg, J.; Kirchner, F.; Licht, U.; Christaller, T. Towards autonomous sewer robots: The MAKRO project. Urban Water 1999, 1, 57–70. [Google Scholar] [CrossRef]
  27. Nassiraei, A.A.; Kawamura, Y.; Ahrary, A.; Mikuriya, Y.; Ishii, K. Concept and design of a fully autonomous sewer pipe inspection mobile robot “kantaro”. In Proceedings of the 2007 IEEE International Conference on Robotics and Automation, Rome, Italy, 10–14 April 2007; pp. 136–143. [Google Scholar]
  28. Alejo, D.; Mier, G.; Marques, C.; Caballero, F.; Merino, L.; Alvito, P. SIAR: A ground robot solution for semi-autonomous inspection of visitable sewers. In Advances in Robotics Research: From Lab to Market; Springer: Cham, Switzerland; Berlin/Heidelberg, Germany, 2020; pp. 275–296. [Google Scholar]
Figure 1. Workflow diagram of the presented S-BIRD dataset.
Figure 1. Workflow diagram of the presented S-BIRD dataset.
Sensors 23 02966 g001
Figure 2. S-BIRD dataset for main sewer blockages.
Figure 2. S-BIRD dataset for main sewer blockages.
Sensors 23 02966 g002
Figure 3. Constructed sewer pipeline.
Figure 3. Constructed sewer pipeline.
Sensors 23 02966 g003
Figure 4. Watertight sewer camera.
Figure 4. Watertight sewer camera.
Sensors 23 02966 g004
Figure 5. Tree root blockage frames in the S-BIRD dataset.
Figure 5. Tree root blockage frames in the S-BIRD dataset.
Sensors 23 02966 g005
Figure 6. Plastic blockage frames in the S-BIRD dataset.
Figure 6. Plastic blockage frames in the S-BIRD dataset.
Sensors 23 02966 g006
Figure 7. Grease blockage frames in the S-BIRD dataset.
Figure 7. Grease blockage frames in the S-BIRD dataset.
Sensors 23 02966 g007
Figure 8. Annotation figures for class (sewer blockage type) balance.
Figure 8. Annotation figures for class (sewer blockage type) balance.
Sensors 23 02966 g008
Figure 9. Annotation heatmap details for captured frames.
Figure 9. Annotation heatmap details for captured frames.
Sensors 23 02966 g009
Figure 10. Imagery data balancing of particular sewer blockage type.
Figure 10. Imagery data balancing of particular sewer blockage type.
Sensors 23 02966 g010
Figure 11. Aspect ratio distribution graph.
Figure 11. Aspect ratio distribution graph.
Sensors 23 02966 g011
Figure 12. Visual upshots of standard augmentation techniques: (a) greyscaling, (b) salt and pepper noise, (c) random exposure.
Figure 12. Visual upshots of standard augmentation techniques: (a) greyscaling, (b) salt and pepper noise, (c) random exposure.
Sensors 23 02966 g012
Figure 13. Visual upshots of advanced augmentation techniques: (a) cutout and (b) mosaic.
Figure 13. Visual upshots of advanced augmentation techniques: (a) cutout and (b) mosaic.
Sensors 23 02966 g013
Figure 14. Annotation details for every single class in training data after image level augmentation.
Figure 14. Annotation details for every single class in training data after image level augmentation.
Sensors 23 02966 g014
Figure 15. Annotation heatmap details for all classes.
Figure 15. Annotation heatmap details for all classes.
Sensors 23 02966 g015
Figure 16. Object count histogram for: (a) grease, (b) tree roots, (c) plastic, and (d) all classes.
Figure 16. Object count histogram for: (a) grease, (b) tree roots, (c) plastic, and (d) all classes.
Sensors 23 02966 g016
Figure 17. Conformation of object detector models.
Figure 17. Conformation of object detector models.
Sensors 23 02966 g017
Figure 18. YOLOX detection results for all classes in S-BIRD.
Figure 18. YOLOX detection results for all classes in S-BIRD.
Sensors 23 02966 g018
Figure 19. Visual upshots of detected tree roots, plastic and grease types of sewer blocks.
Figure 19. Visual upshots of detected tree roots, plastic and grease types of sewer blocks.
Sensors 23 02966 g019
Figure 20. Block diagram of an automated system.
Figure 20. Block diagram of an automated system.
Sensors 23 02966 g020
Figure 21. Embedded vision platform for sewer robotic system.
Figure 21. Embedded vision platform for sewer robotic system.
Sensors 23 02966 g021
Table 1. Specifications of a utilized sewer camera.
Table 1. Specifications of a utilized sewer camera.
FacetsDetails
camera dimension23 mm × 120 mm
camera light12 modifiable white LEDs
watertight gradeIP68
vision angle140 degree
Table 2. Arithmetical details of captured frames.
Table 2. Arithmetical details of captured frames.
Captured framesObject Class (Sewer Blockage Type)Captured Frames
Tree roots2295
Plastic2392
Grease2353
Total frames7040
Annotations10,233 (Average = 1.5 per frame)
Average frame size0.08 Megapixels
Mean frame ratio352 × 240 (wide)
Angle of diagonal0.598 radian = 34.3°
Length of diagonal426 pixels
Aspect ratio Class1.467:1
Pixel density9 pixels/mm or 230 pixels/inch
Table 3. Annotations for training data.
Table 3. Annotations for training data.
Object Class (Sewer Blockage Type)Annotations
Grease2920
Tree roots2455
Plastic1821
Total7196 (Average = 1.5 per frame)
Table 4. Arithmetical details of training frames in S-BIRD after preprocessing and augmentation.
Table 4. Arithmetical details of training frames in S-BIRD after preprocessing and augmentation.
TermsDetails
Total frames14,765
Annotations69,061 (Average = 4.7 per frame)
Average frame size0.173 Megapixels
Mean frame ratio416 × 416 (square)
Aspect ratio Class1:1
Angle of diagonal0.785 radian = 45°
Length of diagonal588 pixels
Pixel density12 pixels/mm or 290 pixels/inch
Table 5. Instances of conformation parts in the object detector models.
Table 5. Instances of conformation parts in the object detector models.
Conformation PartsDetails
Inputframes, multi-scaled frames, frame patches
BackbonesCSPDarknet-53 [15], Darknet53 [16], ResNet-50, ResNet-152, ResNet-10, GoogLeNet, Inception-ResNet-V2, EfficientNet-B0/B7, DetNet-59, ThunderNet, CBNet, VGG16, ViT, etc.
NeckBi-FPN, FPN, SFAM, PAN, etc.
HeadsDenseYOLO [17], SqueezeDet, DetectNet, SSD, RetinaNet, MatrixNet, CenterNet, etc.
SparseMask R-CNN, R-FCN, Faster R-CNN [18], Cascade R-CNN, etc.
Table 6. Crucial traits in training.
Table 6. Crucial traits in training.
TraitsValues
learning modelYOLOX-s
Annotation data typeVOC
max_epoch300
batch_size16
fp16True
num_classes3
Params8.94 M
Gflops26.64
depth0.33
width0.5
input_size(640, 640)
random_size(14, 26)
nmsthre0.65
degrees10.0
translate0.1
scale(0.1, 2)
mscale(0.8, 1.6)
shear2.0
warmup_epochs5
weight_decay0.0005
momentum0.9
Table 7. Time results of the trained model.
Table 7. Time results of the trained model.
Timing ParametersOutturns (Milliseconds)
Average forward time3.19 ms
Average NMS time0.88 ms
Average inference time4.07 ms
Table 8. Precision results of the trained model.
Table 8. Precision results of the trained model.
Class (Sewer Block Type)Average PrecisionMap_5095Map_50
grease0.90040.78850.9005
tree roots0.8930
plastic0.9081
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Patil, R.R.; Mustafa, M.Y.; Calay, R.K.; Ansari, S.M. S-BIRD: A Novel Critical Multi-Class Imagery Dataset for Sewer Monitoring and Maintenance Systems. Sensors 2023, 23, 2966. https://doi.org/10.3390/s23062966

AMA Style

Patil RR, Mustafa MY, Calay RK, Ansari SM. S-BIRD: A Novel Critical Multi-Class Imagery Dataset for Sewer Monitoring and Maintenance Systems. Sensors. 2023; 23(6):2966. https://doi.org/10.3390/s23062966

Chicago/Turabian Style

Patil, Ravindra R., Mohamad Y. Mustafa, Rajnish Kaur Calay, and Saniya M. Ansari. 2023. "S-BIRD: A Novel Critical Multi-Class Imagery Dataset for Sewer Monitoring and Maintenance Systems" Sensors 23, no. 6: 2966. https://doi.org/10.3390/s23062966

APA Style

Patil, R. R., Mustafa, M. Y., Calay, R. K., & Ansari, S. M. (2023). S-BIRD: A Novel Critical Multi-Class Imagery Dataset for Sewer Monitoring and Maintenance Systems. Sensors, 23(6), 2966. https://doi.org/10.3390/s23062966

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop