Advancing the Robotic Vision Revolution: Development and Evaluation of a Bionic Binocular System for Enhanced Robotic Vision
Abstract
:1. Introduction
2. Methods
2.1. Structural Design of Bionic Binocular Camera
2.2. Basis of Binocular Vision
2.3. Binocular Vision with Degrees of Freedom
2.3.1. The Model with Only the Camera Rotating
2.3.2. The Model of Co-Rotation of Axis and Camera
3. Experiments
3.1. Experimental Platform
3.2. Binocular Ranging Accuracy Experiment
3.3. Camera Parameters and Error Testing
3.4. Distance Testing
3.5. Bionic Binocular Camera Dynamic Viewing Angle Test
3.6. Target Detection Algorithm
3.7. Comparison of Existing Robotic Bionic Eye Devices
3.8. Experimental Summary
- Camera Specifications: JEWXON BC200 Bionic Eye Binocular Camera: Features a maximum resolution of 3296 × 2512 at 30 fps, Static angle FOV: H65°, V51°, Best dynamic angle FOV: H206°, V192°, integrated IMU, a working distance of 0.2–10 m, and a power consumption of 2.8 W. This camera surpasses the control group’s other cameras in resolution, maximum viewing angle, effective working distance, and energy efficiency.
- Vision System Combination: The JEWXON BC200, paired with the Jetson Orin NX, achieves an AI performance of 100 TOPS. This combination represents one of the more advanced embedded vision systems currently available, better suited for mobile robots due to its lower power consumption compared to PCs.
- Error Experiment: The JEWXON BC200 bionic eye binocular camera has the best distortion value, the average error is 0.08 pixels, and the performance is excellent, performing better than the Fuayun A100 standard binocular camera.
- Vision Ranging: The measurement accuracy of the JEWXON BC200 is closer to the world coordinate system distance compared to the ORBBEC Dabai DW depth camera. Using only the camera rotation model during the OpenCV panoramic image synthesis process can lead to image precision errors resulting in distortions and pixel misalignments that cause parts of the composite image to be missing. However, the combined rotation model of the axis and camera significantly improves image distortion, enhancing image precision and thus creating a more perfect composite image.
- Enhancing Robotic Vision: To make robotic vision more akin to human vision, the introduction of YOLO-V8 improves the autonomous recognition ability and recognition accuracy of the bionic eye. The BC200 can accurately identify objects such as clocks, laptops, keyboards, mice, cups, pedestrians, and so on in the real world. The bionic eye binocular camera not only provides a broader field of view and more stable images but also achieves autonomous target tracking and precise identification, surpassing the capabilities of fixed binocular and structured light cameras.
4. Conclusions
5. Future Work Focus
- A.
- The hardware platform for the binocular-vision system with degrees of freedom requires improvements, including adopting more advanced visual platforms, enhancing computational power while achieving lower operational power consumption, and increasing both installation and control precision. This will reduce the introduction of errors and enhance the accuracy of depth computation.
- B.
- While it is feasible to calculate distances using functional methods when the camera has degrees of freedom, it also introduces a considerable computational model that increases processing time; there is still room for future optimization.
- C.
- A solution is needed for the problem of absolute error variation with distance. One potential method to reduce errors is to attempt to replace the existing YOLO-V8 with a more precise and advanced target detection model.
Author Contributions
Funding
Institutional Review Board Statement
Informed Consent Statement
Data Availability Statement
Acknowledgments
Conflicts of Interest
References
- Lu, H.M.; Li, Y.J.; Chen, M.; Kim, H.; Serikawa, S. Brain Intelligence: Go beyond Artificial Intelligence. Mobile Netw. Appl. 2018, 23, 368–375. [Google Scholar] [CrossRef]
- Bathla, G.; Bhadane, K.; Singh, R.K.; Kumar, R.; Aluvalu, R.; Krishnamurthi, R.; Kumar, A.; Thakur, R.N.; Basheer, S. Autonomous Vehicles and Intelligent Automation: Applications, Challenges, and Opportunities. Mob. Inf. Syst. 2022, 2022, 36. [Google Scholar] [CrossRef]
- Kober, J.; Bagnell, J.A.; Peters, J. Reinforcement learning in robotics: A survey. Int. J. Robot. Res. 2013, 32, 1238–1274. [Google Scholar] [CrossRef]
- Maimaitijiang, M.; Sagan, V.; Sidike, P.; Hartling, S.; Esposito, F.; Fritschi, F.B. Soybean yield prediction from UAV using multimodal data fusion and deep learning. Remote Sens. Environ. 2020, 237, 20. [Google Scholar] [CrossRef]
- Peng, Y.Y.; Yang, N.; Xu, Q.; Dai, Y.; Wang, Z.Q. Recent Advances in Flexible Tactile Sensors for Intelligent Systems. Sensors 2021, 21, 5392. [Google Scholar] [CrossRef]
- Li, G.; Xie, D.G.; Zhong, H.; Zhang, Z.Y.; Fu, X.K.; Zhou, Q.L.; Li, Q.; Ni, H.; Wang, J.O.; Guo, E.J.; et al. Photo-induced non-volatile VO2 phase transition for neuromorphic ultraviolet sensors. Nat. Commun. 2022, 13, 9. [Google Scholar] [CrossRef]
- Golnabi, H.; Asadpour, A. Design and application of industrial machine vision systems. Robot. Comput.-Integr. Manuf. 2007, 23, 630–637. [Google Scholar] [CrossRef]
- Perez, L.; Rodriguez, I.; Rodriguez, N.; Usamentiaga, R.; Garcia, D.F. Robot Guidance Using Machine Vision Techniques in Industrial Environments: A Comparative Review. Sensors 2016, 16, 335. [Google Scholar] [CrossRef]
- Xie, M.; Lai, T.; Fang, Y. A New Principle Toward Robust Matching in Human-like Stereovision. Biomimetics 2023, 8, 285. [Google Scholar] [CrossRef]
- LeCun, Y.; Bottou, L.; Bengio, Y.; Haffner, P. Gradient-based learning applied to document recognition. Proc. IEEE 1998, 86, 2278–2324. [Google Scholar] [CrossRef]
- Deng, J.; Dong, W.; Socher, R.; Li, L.J.; Li, K.; Li, F.F. ImageNet: A Large-Scale Hierarchical Image Database. In Proceedings of the IEEE-Computer-Society Conference on Computer Vision and Pattern Recognition Workshops, Miami Beach, FL, USA, 20–25 June 2009. [Google Scholar] [CrossRef]
- Krizhevsky, A.; Sutskever, I.; Hinton, G.E. Imagenet classification with deep convolutional neural networks. Adv. Neural Inf. Process. Syst. 2012, 25. [Google Scholar] [CrossRef]
- Simonyan, K.; Zisserman, A. Very deep convolutional networks for large-scale image recognition. arXiv 2014, arXiv:1409.1556. [Google Scholar]
- He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 26 June–1 July 2016. [Google Scholar] [CrossRef]
- Hou, J.; You, B.; Xu, J.; Wang, T.; Cao, M. Surface Defect Detection of Preform Based on Improved YOLOv5. Appl. Sci. 2023, 13, 7860. [Google Scholar] [CrossRef]
- Zheng, Y.; Zhou, G.; Lu, B. A Multi-Scale Rebar Detection Network with an Embedded Attention Mechanism. Appl. Sci. 2023, 13, 8233. [Google Scholar] [CrossRef]
- LeCun, Y.; Kavukcuoglu, K.; Farabet, C. Convolutional Networks and Applications in Vision. In Proceedings of the International Symposium on Circuits and Systems Nano-Bio Circuit Fabrics and Systems (ISCAS 2010), Paris, France, 30 May–2 June 2010. [Google Scholar] [CrossRef]
- Wang, J.D.; Sun, K.; Cheng, T.H.; Jiang, B.R.; Deng, C.R.; Zhao, Y.; Liu, D.; Mu, Y.D.; Tan, M.K.; Wang, X.G.; et al. Deep High-Resolution Representation Learning for Visual Recognition. IEEE Trans. Pattern Anal. Mach. Intell. 2021, 43, 3349–3364. [Google Scholar] [CrossRef]
- Shijie, F.; Chao, Z.; Wei, Y.; Qian, C. Application of deep learning technology to fringe projection 3D imaging. Infrared Laser Eng. 2020, 49, 0303018-0303018-17. [Google Scholar] [CrossRef]
- Sun, X.; Jiang, Y.; Ji, Y.; Fu, W.; Yan, S.; Chen, Q.; Yu, B.; Gan, X. Distance measurement system based on binocular stereo vision. In Proceedings of the IOP Conference Series: Earth and Environmental Science, Xiamen, China, 7–9 June 2019. [Google Scholar] [CrossRef]
- Hansard, M.; Lee, S.; Choi, O.; Horaud, R.P. Time-of-Flight Cameras: Principles, Methods and Applications; Springer Science & Business Media: Berlin/Heidelberg, Germany, 2012. [Google Scholar]
- Xia, X.; Dai, S.; Qi, H.; Xu, Z.; Wang, S.; Zhang, M. Research on Object Measurement Based on 3D Stereo Vision. In Proceedings of the 2021 33rd Chinese Control and Decision Conference (CCDC), Kunming, China, 22–24 May 2021. [Google Scholar] [CrossRef]
- Zhang, S. High-speed 3D shape measurement with structured light methods: A review. Opt. Lasers Eng. 2018, 106, 119–131. [Google Scholar] [CrossRef]
- Zhang, S. High-Speed 3D Imaging with Digital Fringe Projection Techniques; Proc. SPIE 11813, Tribute to James C. Wyant: The Extraordinaire in Optical Metrology and Optics Education, 118130V (9 September 2021). Available online: https://www.spiedigitallibrary.org/conference-proceedings-of-spie/11813/118130V/High-speed-3D-imaging-with-digital-fringe-projection-techniques/10.1117/12.2567675.short (accessed on 14 June 2024).
- Rowley, D. A holographic interference camera. J. Phys. E Sci. Instrum. 1979, 12, 971. [Google Scholar] [CrossRef]
- Qian, J.F.; Ma, S.W.; Xu, Y.L.; Li, X.; Shen, Y.J. Application of Binocular Vision System to Face Detection and Tracking in Service Robot. Proc. SPIE 8335, 2012 International Workshop on Image Processing and Optical Engineering, 83350P (15 November 2011). Available online: https://www.spiedigitallibrary.org/conference-proceedings-of-spie/8335/83350P/Application-of-binocular-vision-system-to-face-detection-and-tracking/10.1117/12.918122.short (accessed on 14 June 2024).
- Shang, C.C.; Ma, H.W. Coal Mine Robot Binocular Vision Recognition System Based on Fuzzy Neural Network. In Proceedings of the International Conference on Automation, Mechanical and Electrical Engineering (AMEE), Phuket, Thailand, 26–27 July 2015. [Google Scholar] [CrossRef]
- Guo, S.X.; Chen, S.Z.; Liu, F.G.; Ye, X.F.; Yang, H.B. Binocular Vision-based Underwater Ranging Methods. In Proceedings of the IEEE International Conference on Mechatronics and Automation (ICMA), Takamatsu, Japan, 6–9 August 2017. [Google Scholar] [CrossRef]
- Sheng, H.Y.; Wei, S.M.; Yu, X.L.; Tang, L. Research on Binocular Visual System of Robotic Arm Based on Improved SURF Algorithm. IEEE Sens. J. 2020, 20, 11849–11855. [Google Scholar] [CrossRef]
- Noton, D.; Stark, L. Eye movements and visual perception. Sci. Am. 1971, 224, 34–43. [Google Scholar]
- Henderson, J.M. Eye movements. In The Oxford Handbook of Cognitive Psychology; Reisberg, D., Ed.; Oxford Library of Psychology; Oxford University Press: Hongkong, China, 2013; pp. 69–82. [Google Scholar]
- Cornsweet, T. Visual Perception; Academic Press: Cambridge, MA, USA, 2012. [Google Scholar]
- Gordon, I.E. Theories of Visual Perception; Psychology Press: London, UK, 2004. [Google Scholar] [CrossRef]
- Scassellati, B. Eye Finding via Face Detection for a Foveated Active Vision System; AAAI/IAAI: Cambridge, MA, USA, 1998. [Google Scholar]
- Murata, A. Ergonomics and cognitive engineering for robot-human cooperation. In Proceedings of the 9th IEEE International Workshop on Robot and Human Interactive Communication. IEEE RO-MAN 2000 (Cat. No.00TH8499), Osaka, Japan, 27–29 September 2000; pp. 206–211. [Google Scholar] [CrossRef]
- Breazeal, C. Socially intelligent robots: Research, development, and applications. In Proceedings of the 2001 IEEE International Conference on Systems, Man and Cybernetics. e-Systems and e-Man for Cybernetics in Cyberspace (Cat.No.01CH37236), Tucson, AZ, USA, 7–10 October 2001; Volume 4, pp. 2121–2126. [Google Scholar] [CrossRef]
- Shaw-Garlock, G. Looking forward to sociable robots. Int. J. Soc. Robot. 2009, 1, 249–260. [Google Scholar] [CrossRef]
- Lou, H.; Duan, X.; Guo, J.; Liu, H.; Gu, J.; Bi, L.; Chen, H. DC-YOLOv8: Small-Size Object Detection Algorithm Based on Camera Sensor. Electronics 2023, 12, 2323. [Google Scholar] [CrossRef]
- Wang, H.; Liu, C.; Cai, Y.; Chen, L.; Li, Y. YOLOv8-QSD: An Improved Small Object Detection Algorithm for Autonomous Vehicles Based on YOLOv8. IEEE Trans. Instrum. Meas. 2024, 73, 1–16. [Google Scholar] [CrossRef]
- Zhang, H.; Lee, S. Robot Bionic Vision Technologies: A Review. Appl. Sci. 2022, 12, 7970. [Google Scholar] [CrossRef]
- Wang, Q.; Yin, Y.; Zou, W.; Xu, D. Measurement error analysis of binocular stereo vision: Effective guidelines for bionic eyes. IET Sci. Meas. Technol. 2017, 11, 829–838. [Google Scholar] [CrossRef]
- Chen, X.; Wang, C.; Zhang, T.; Hua, C.; Fu, S.; Huang, Q. Hybrid Image Stabilization of Robotic Bionic Eyes. In Proceedings of the 2018 IEEE International Conference on Robotics and Biomimetics (ROBIO), Kuala Lumpur, Malaysia, 12–15 December 2018. [Google Scholar] [CrossRef]
- Zhang, H.; Lee, S. Robot Bionic Eye Motion Posture Control System. Electronics 2023, 12, 698. [Google Scholar] [CrossRef]
- Zhu, Z.; Wang, Q.; Zou, W.; Zhang, F. Motion Control on Bionic Eyes: A Comprehensive Review. arXiv 2019, arXiv:1901.01494. [Google Scholar]
Technology Category | Monocular Vision | Binocular Stereo Vision | Structured Light | Bionic binocular Vision |
---|---|---|---|---|
Product pictures | ||||
Technology principle | ||||
Principle of work | Single camera | Dual camera | Camera and infrared projection patterns | Autonomous motion dual camera |
Response time | Fast | Medium | Slow | Algorithm determines speed |
Weak light | Weak | Weak | Good | Weak |
Bright light | Good | Good | Weak | Good |
Identification precision | Low | Low | Medium | Algorithm determination accuracy |
Resolving capability | High | High | Medium | High |
Identification distance | Medium | Medium | Very short | Medium |
Operation difficulty | Low | High | Medium | High |
Cost | Low | Medium | High | High |
Power consumption | Low | Low | Medium | Low |
Disadvantages | Low recognition accuracy, poor dark light | Dark light features are not obvious | High requirements for ambient light, short recognition distance | Automatic target tracking, super wide viewing Angle |
Representative company | Cognex, Honda, Keyence | LeapMoTion, Fuayun | Intel, Microsoft, ORBBEC | Huawei, Eyevolution, Jewxon |
Point | x-Coordinate | y-Coordinate |
---|---|---|
Point | x-Coordinate | y-Coordinate |
---|---|---|
Point | x-Coordinate | y-Coordinate |
---|---|---|
Metric/Camera | Fuayun A100 | JEWXON BC200 | Comparison |
---|---|---|---|
Focal Length (mm) | Left: 417.0386, Right: 456.7638 | Left: 595.6271, Right: 594.2564 | BC200 has a longer focal length, suitable for long-distance shooting and providing a wider field of view. |
Distortion Coefficients | Left: [−0.057164, 0.095798, 0.011117, 0.000282, 0], Right: [−0.006848, −0.054429, 0.010137, 0.001460, 0] | Left: [0.0403, −0.095, −0.001167, 0.000428, 0], Right: [0.0299, −0.0322, −0.000880, −0.002340, 0] | BC200 has overall lower distortion coefficients, indicating higher image quality and less optical distortion. |
Image Size (pixels) | 640 × 480 | 640 × 480 | Equivalent test environment |
Rotation Matrix Stability | Lefit: [326.8532231, 252.4517895] Right: [323.8062822, 256.9767981] | Lefit: [321.5142126, 181.3599961] Right: [303.7775542 183.9382323] | BC200 may deliver images with less distortion, beneficial for applications that require high precision and image fidelity. |
Translation Matrix | [16.9321,0.0315, 1.4005] | [−48.7250, −0.2644, 1.8614] | BC200 larger Z-axis offset is suitable for long-distance imaging. |
Feature/Metric | Only Camera Rotates Model | Axis and Camera Rotate Model |
---|---|---|
Stability of Image | Moderate | High |
Image Distortion | Significant | Minimal |
Coverage Area | Limited | Extensive |
Resolution Consistency | Consistent | Consistent |
Suitability for Dynamic Scenes | Moderate | Excellent |
Zou Wei et al. R&D Team [44] | Chen et al. R&D Team [42] | JEWXON BC200 |
---|---|---|
Features: The device consists of two CCD cameras and stepping motors, which can basically realize the movement function of the human eye. | Features: each eye contains two cameras (long-focus lens and short-focus lens) for simulating the perception of human eye features. And a 3-degree-of-freedom neck mechanism is designed with an integrated IMU. | Features: 4K HD mini-camera integrated in each eye, integrated IMU, and due to the latest levitation conduction technology used in the data collector, the eye rotates without being interfered by cables during the eye rotation, resulting in a larger angle of eye rotation. |
Disadvantages: The model uses larger stepper motors, resulting in an excessively large product, and multiple large motors and loads running will inevitably lead to higher overall operating power consumption of the device, which is not conducive to mobility and portability. | Disadvantages: Although the motor controlling the rotation of the eyeball is reduced, the overall design of the bionic eye shape is too large and a larger motor is used for the neck mechanism, which leads to higher power consumption of the device and is not suitable for use with mobile devices. | Disadvantages: Although the overall use of mini cameras and mini motors, the overall power consumption drops a lot, but there is still room for improvement, such as the realization of the head and neck movement function. |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2024 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Zhang, H.; Lee, S. Advancing the Robotic Vision Revolution: Development and Evaluation of a Bionic Binocular System for Enhanced Robotic Vision. Biomimetics 2024, 9, 371. https://doi.org/10.3390/biomimetics9060371
Zhang H, Lee S. Advancing the Robotic Vision Revolution: Development and Evaluation of a Bionic Binocular System for Enhanced Robotic Vision. Biomimetics. 2024; 9(6):371. https://doi.org/10.3390/biomimetics9060371
Chicago/Turabian StyleZhang, Hongxin, and Suan Lee. 2024. "Advancing the Robotic Vision Revolution: Development and Evaluation of a Bionic Binocular System for Enhanced Robotic Vision" Biomimetics 9, no. 6: 371. https://doi.org/10.3390/biomimetics9060371
APA StyleZhang, H., & Lee, S. (2024). Advancing the Robotic Vision Revolution: Development and Evaluation of a Bionic Binocular System for Enhanced Robotic Vision. Biomimetics, 9(6), 371. https://doi.org/10.3390/biomimetics9060371