Next Article in Journal
Developing a Biomimetic 3D Neointimal Layer as a Prothrombotic Substrate for a Humanized In Vitro Model of Atherothrombosis
Previous Article in Journal
Design and Control of a Tendon-Driven Robotic Finger Based on Grasping Task Analysis
Previous Article in Special Issue
Innate Orientating Behavior of a Multi-Legged Robot Driven by the Neural Circuits of C. elegans
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Advancing the Robotic Vision Revolution: Development and Evaluation of a Bionic Binocular System for Enhanced Robotic Vision

by
Hongxin Zhang
1,* and
Suan Lee
2,*
1
Jewxon Intelligent Technology Co., Ltd., No. 293, Gongren East Road, Jiaojiang District, Taizhou 318000, China
2
School of Computer Science, Semyung University, Jecheon 27136, Republic of Korea
*
Authors to whom correspondence should be addressed.
Biomimetics 2024, 9(6), 371; https://doi.org/10.3390/biomimetics9060371
Submission received: 8 May 2024 / Revised: 12 June 2024 / Accepted: 15 June 2024 / Published: 19 June 2024
(This article belongs to the Special Issue Bio-Inspired and Biomimetic Intelligence in Robotics)

Abstract

:
This paper describes a novel bionic eye binocular vision system designed to mimic the natural movements of the human eye. The system provides a broader field of view and enhances visual perception in complex environments. Compared with similar bionic binocular cameras, the JEWXON BC200 bionic binocular camera developed in this study is more miniature. It consumes only 2.8 W of power, which makes it ideal for mobile robots. Combining axis and camera rotation enables more seamless panoramic image synthesis and is therefore suitable for self-rotating bionic binocular cameras. In addition, combined with the YOLO-V8 model, the camera can accurately recognize objects such as clocks and keyboards. This research provides new ideas for the development of robotic vision systems.

1. Introduction

In the context of the new technological revolution and industrial transformation, the deep integration of artificial intelligence and robotics is essential for unleashing human productivity and accelerating the implementation of AI technologies. This profound empowerment of traditional industries effectively drives their transformation and upgrade [1,2,3]. Sensors perceive the external world as critical components of robots and provide essential information to support AI’s accurate decision-making [4,5]. Among these, vision sensors play a crucial role in the deep integration of AI and robotics. They are widely used, enabling robots equipped with vision systems to perceive and interpret their surroundings, thus allowing them to interact with the world and efficiently perform various tasks.
As one of the most crucial human organs, the eye plays a vital role in shaping our understanding of the world [6]. Given its significance, robots with vision systems can perceive and interpret their surroundings, enabling them to interact with the world and perform various tasks effectively [7,8,9]. The development of applications that combine machine vision as the eyes and artificial intelligence as the brain has been a focal point of research. In the 1990s, Yann LeCun pioneered the convolutional neural network (CNN) with his creation of LeNet-5, which endowed computers with the ability to learn and recognize image data [10]. Effectively, ImageNet, a large-scale image dataset, has acted as a catalyst for the development of computer vision, showcasing the impressive performance of deep learning in image recognition and laying the foundation for further research into vision-intelligent robots [11,12,13,14,15,16,17,18]. However, these applications typically utilize a single camera and focus on analyzing two-dimensional images on a flat plane. The material world humans inhabit is three-dimensional. The technology to capture and process three-dimensional information reflects a significant grasp of the target world and is a hallmark of intelligence [19]. Consequently, bionic eyes enable robots to obtain three-dimensional coordinates by mimicking the structure of the human eye, propelling artificial intelligence into a new phase of development.
Monocular vision, binocular stereoscopic vision, and structured light technology have all been extensively researched; for more details, see Table 1. Monocular vision systems are known for their simple setup, fast response times, low power consumption, and cost efficiency. However, their main disadvantage is lower recognition accuracy. Representative companies in this field include Cognex, Honda, and Keyence. Instruments capable of three-dimensional measurement include binocular stereoscopic cameras [20], Time-of-Flight (TOF) cameras [21,22], structured light imaging cameras [23,24], holographic interferometry cameras [25], and more. Among these, the structure of binocular-vision cameras most closely resembles that of the human eye. Binocular stereoscopic vision technology uses two cameras to simulate the disparity principle of human eyes, obtaining planar and depth information from the real environment through triangulation. Companies like Luster (Bumblebee) and Fuayun (A100) represent the early stages of vision processing technology when algorithm and device precision were not very high; binocular-vision cameras do not emit light actively but rely entirely on two captured images to calculate depth, utilizing passive binocular imaging techniques. Consequently, structured light cameras were developed, using an infrared projection mode and binocular cameras to calculate the features projected onto object surfaces, thus significantly enhancing recognition precision compared to traditional stereoscopic cameras. However, this method requires high ambient light conditions and has a limited projection range, failing to function properly in reflective light conditions. Companies like Intel (Realsense), ORBBEC, and Microsoft (Kinect) represent this product category.
Binocular-vision cameras are more cost-effective compared to other 3D optical devices. They can also adapt to complex lighting conditions and capture dynamic scenes and moving targets. Consequently, binocular cameras are often used as bionic eyes and installed on mobile robots, where they have found widespread applications across various fields. In service robots, Qian implemented a binocular-vision system combined with the AdaBoost algorithm to detect facial regions in real time, facilitating tasks such as conversing with people, tracking faces, and coordinating mechanical arms to grasp objects [26]. In coal mine rescue robots, robots equipped with binocular-vision systems accurately collect information about collapses and obstacles within mines, providing critical feedback to rescue personnel to prevent secondary accidents [27]. Underwater rescue robots equipped with binocular stereovision, despite challenging and variable lighting conditions, demonstrate exceptional distance measurement capabilities and can perform underwater search and rescue tasks [28]. In an essential area of robotics—robotic arm vision—Sheng utilized the SURF algorithm to extract and match image feature points based on binocular vision, accomplishing the mechanical arm’s grasping of targets [29]. In the research on autonomous forestry robots for trunk distance measurement, Zhao developed a trunk-measuring system based on binocular vision theory using TI’s DaVincDM37x system. Zhao calculated three-dimensional information from images captured by the binocular cameras, obtaining accurate target location and distance measurements. To enhance the accuracy of obstacle detection algorithms in autonomous vehicles, Liang developed a binocular vision-based obstacle detection algorithm, which minimizes recognition errors and accurately measures the relative position between the vehicle and obstacles.
Binocular-vision technologies have been developed in recent years. Bionic binocular-vision technology mimics the visual mechanisms of humans and animals. However, most binocular cameras cannot rotate autonomously, thus missing autonomous flexibility. This paper primarily focuses on designing a binocular-vision system that can imitate human eye mechanisms, achieving flexibility similar to human eyes and automatically tracking targets for dynamic recognition, thereby providing a broader field of view, as shown in Figure 1. Eye movement is crucial for visual perception. Collecting more information, achieving a more comprehensive view, and eliminating blind spots all depend on the rotation of the eyes [30,31]. Eyes with degrees of freedom can smoothly track and locate targets, keeping the area of interest at the center of the frame and allowing quick responses to changes in the tracked target [32,33]. Particularly for robots performing target tracking tasks, when the target swiftly changes direction and the robot body cannot quickly turn, a binocular-vision system fixed on the mobile robot body is highly likely to lose track of the target. Therefore, researching a binocular-vision system with degrees of freedom is essential.
Basic binocular-vision systems have inherent limitations in wide-angle coverage and image quality, making it difficult for robots to meet performance requirements in complex and challenging environments. Therefore, this paper has developed two models of binocular-vision systems with degrees of freedom, including one model where only the camera rotates and another where both the axis and camera rotate together. The model where only the camera rotates is considered the ideal model. The proposed models can effectively improve the robotic vision performance, enhancing the system’s capabilities in challenging and complex environments, increasing its adaptability, and expanding its applications. Additionally, it is essential to note that traditional binocular-vision depth calculation methods, which rely on the principles of similar triangles, are no longer applicable when there are degrees of freedom in the binocular system. Thus, this paper also presents the depth calculation formulas corresponding to the two models.

2. Methods

2.1. Structural Design of Bionic Binocular Camera

To mimic the visual motion functions of human eyes, this study developed the JEWXON BC200 bionic binocular camera, as shown in Figure 2. This device consists mainly of precise miniature components, including the left eyeball and its associated motors. The motor for the up-and-down movement of the left eyeball is responsible for its vertical movement, simulating the up-and-down observation function of the human eye. In contrast, the left eyeball left/right movement motor handles horizontal movements, allowing the left eyeball to move from side to side, thus expanding the field of view. The right eyeball and its motors coordinate with the left eyeball to achieve binocular vision. The motor for the up and down movement of the right eyeball and the right eyeball left/right movement motor control the vertical and horizontal movements of the right eyeball, respectively. Both eyeballs are equipped with an IMU, which is responsible for collecting the attitude data of the eyeballs. The micro camera, embedded within the left and right bionic eyeballs, captures images and, along with the IMU data, transmits them in real time to the data acquisition sensors. These sensors are connected to the PCB via a cable, which transmits the data to the MCU for real-time processing. This lets the camera respond quickly in dynamic environments, capturing clear images and accurate eyeball attitude angles. After processing the data, the MCU communicates with the drive unit to perform the necessary operations. The drive is connected to the motors, controlling the multiple movement motors, allowing the left and right eyeballs to move up and down independently and left and right, significantly expanding the camera’s field of view. The PCB is designed for low power consumption, with a power usage of only 2.8 W, which is particularly important for mobile robots as low power consumption translates to longer operational times and higher efficiency. The PCB is mounted on the back of the bionic eyeballs and integrates data input/output interfaces. Through the coordinated work of these components, the JEWXON BC200 bionic binocular camera operates efficiently in dynamic environments, capturing high-quality images and providing precise eyeball attitude data, thus offering reliable visual perception capabilities for mobile robots.

2.2. Basis of Binocular Vision

Monocular cameras can capture two-dimensional images or videos but lack the ability to directly perceive depth information. As shown in Figure 3, on the same line, three different targets at varying distances project onto the same position in camera A. Therefore, camera A cannot distinguish which point is farther or closer based on the formed images. Due to the lack of a human eye-like stereoscopic sensing system and parallax for the target, monocular cameras are unable to directly acquire depth information. Thanks to studies of the human visual system, the researchers found that both eyes have different views of the same observation target. The brain could infer the depth of the target by comparing these two view differences.
The development of camera depth perception technology can be traced back to the 1980s. In 1998, based on the human eye’s mechanism, the Massachusetts Institute of Technology (MIT) started developing robot bionic eyes and created the Kismet robot [34]. The robot could interact by capturing facial expressions using two Charge-coupled Device (CCD) cameras. After that, an increasing number of researchers focused on developing robot vision [35,36,37], primarily centered around the research of bionic eyes composed of binocular cameras. Utilizing the parallax between the left and right cameras for the same observation target, binocular imaging techniques could calculate the distance from the camera to the target. As the target moves away from the camera, the parallax between the images seen by the left and right cameras decreases, and the difference increases as the target moves closer to the camera. To accurately calculate the distance to a target, it is necessary to calibrate the left and right cameras to determine their geometric relationship and distortion parameters. Subsequently, the feature points in the left and right images would be matched, and the depth information of the target would be obtained by utilizing the parallax of the image and the principle of triangulation.
A schematic diagram of the binocular-vision system is shown in Figure 4. In the ideal scenario of a robot bionic eye, two identical cameras are placed in parallel and simultaneously capture images. Due to the different positions of the two cameras (one left and one right), the image points formed by the same observation target in the left and right camera are not in the same position, resulting in a particular parallax. Based on the parallax between the left and right cameras, the depth information of the target distance from the binocular camera can be calculated by utilizing the principle of similar triangles.
p is the center of the observed target. L 1 and L 2 are the projection planes of the left and right cameras, respectively. L 1 is the plane after rotating L 1 , and L 2 is the plane after rotating L 2 . o l and o r are the optical centers of the two cameras, respectively. The installation height of the two cameras is A , the installation distance is B , and both cameras have a focal length of f . The horizontal length of the camera’s projection plane is C . To facilitate the understanding of the calculation in the presence of degrees of freedom, the two planes L 1 and L 2 are chosen as the reference planes. According to the triangle similarity theorem, it is easy to prove: p x l x r p o l o r . The distance from fixed point p to o l   o r is labeled Z . Thus, there is the following relation:
B B ( ( B C ) + N 2 + C N 1 ) = Z f
The equation simplifies to:
Z = f B N 1 N 2
From this, the target depth distance can be measured as:
Z = Z + f + A
f , A , and B can be obtained through measurement; thus, they are known parameters. By multiplying the pixel disparity of matched points with the real size of each pixel, the value of N 1 N 2 can be obtained. This is the depth calculation method employed for basic binocular vision.

2.3. Binocular Vision with Degrees of Freedom

However, in many specialized and complex environments, basic binocular-vision systems often fail to achieve the desired results; for example, those scenes that require wider viewing angles and higher image quality. Therefore, in this paper, two models of binocular-vision systems with degrees of freedom were built, including the model with only the camera rotating and the model of co-rotation of the axis and camera. Also, in the paper, the corresponding depth calculation formulas for these two models were provided.

2.3.1. The Model with Only the Camera Rotating

Ideal binocular vision is shown in Figure 5. The binocular camera is free to rotate while the connecting axis is immobile perpendicular to the substrate.
A simple coordinate system was constructed in order to calculate the distance from the center p of the observed target to the substrate. α represents the angle of camera rotation, f denotes the camera focal length, A corresponds to the distance from the fixed point o 1 to the projection plane L 1 , and C represents the horizontal length of the projection plane. o l was chosen as the origin coordinate ( 0 , 0 ) . At the same time, clockwise rotation was considered positive, counterclockwise rotation was considered negative, and | α | < 60 . The coordinates of the points of the left camera are shown in Table 2:
Let the line function of the line containing points x l and p be g 2 ( x ) = p 1 x + q 1 . The slope p 1 of g 1 ( x ) could be calculated by the points x l and x l :
p 1 = cos α f sin α ( N 1 C 2 ) sin α f + cos α ( N 1 C 2 )
By substituting point o l into g 1 ( x ) , the bias term q 1 was obtained as:
q 1 = f ( x ) p 1 x = C o s α f S i n α C o s α f 2 S i n 2 α f ( N 1 C / 2 ) S i n α f + C o s α ( N 1 C / 2 ) = C o s 2 α f N 1 C / 2 + S i n 2 α f N 1 C / 2 S i n α f + C o s α N 1 C / 2 = f ( N 1 C / 2 ) S i n α f + C o s α ( N 1 C / 2 )
Similarly, let the points x r and p on the right lie on the line g 2 ( x ) = p 2 x + q 2 . Similarly, the slope p 2 of the right camera was as follows:
p 2 = C o s α f s i n α ( N 2 C / 2 ) S i n α f + C o s α ( N 2 C / 2 )
The coordinates of o r could be calculated as ( s i n α f + B , c o s α f ) . The bias term q 2 could be obtained by substituting point o r into the linear function g 2 ( x ) :
q 2 = f ( N 2 C / 2 ) S i n α f + C o s α ( N 2 C / 2 ) B p 2
The coordinates of the intersection of g 1 and g 2 are the coordinates of point p .

2.3.2. The Model of Co-Rotation of Axis and Camera

However, in practical application situations, the camera often needs to be fixed on top of a substrate with a certain thickness. Therefore, the model where the camera rotates alone and the axis does not rotate is the ideal case. The model of co-rotation of the axis and camera is more common and realizable. The schematic of the model of co-rotation of the axis and camera is shown in Figure 6.
A simple coordinate system was constructed in order to calculate the distance from the center p of the observed target to the substrate. α represents the angle of camera rotation, f denotes the camera focal length, A corresponds to the distance from the rotation point o 1 to the projection plane L 1 , and C represents the horizontal length of the projection plane. o 1 was chosen as the origin coordinate ( 0 , 0 ) , then o 2 coordinates were ( B , 0 ) . At the same time, clockwise rotation was considered positive, counterclockwise rotation was considered negative, and | α | < 60 . When the projection plane was parallel to the mounting substrate, α was considered 0 . The length of the projection plane and the mounting substrate were the same as when they are perpendicular, so the coordinates of the points on the left can be easily determined as shown in Table 3:
Let p and o l be the function of this line, denoted as f 1 ( x ) = a 1 x + b 1 . Since the points p, o l , x l , and x l were collinear, any two sets of coordinates could be substituted into the line equation to determine the function of this line. It followed that:
a 1 = C o s α A + S i n α ( N 1 C / 2 ) C o s α ( A + 2 f ) S i n α N 1 C / 2 S i n α A C o s α N 1 C / 2 S i n α ( A + 2 f ) + C o s α N 1 C / 2 = 2 f C o s α + 2 S i n α ( N 1 C / 2 ) 2 f S i n α 2 C o s α ( N 1 C / 2 ) = S i n α ( N 1 C / 2 ) f C o s α C o s α ( N 1 C / 2 ) f S i n α
By substituting a 1 and o L into f 1 ( x ) , b 1 could be determined as follows:
b 1 = C o s α ( A + f ) S i n α N 1 C / 2 f C o s α C o s α N 1 C / 2 f S i n α S i n α ( A + f )
When α = 0 , the expression for f 1 ( x ) could be simplified to:
f 1 ( x ) = f N 1 C / 2 x + A + f
Similarly, the coordinates of each point of the right camera could be obtained, as shown in Table 4:
Compared to the left camera, the right camera differs mainly in the extra bias term B in the horizontal coordinate. Therefore, the calculation method for the slope a 2 in the corresponding function f 2 ( x ) = a 2 x + b 2 was similar, and it was as follows:
a 2 = S i n α N 2 C / 2 f C o s α C o s α N 2 C / 2 f S i n α
Since there was a bias term B , b 2 was:
b 2 = C o s α ( A + f ) S i n α ( N 2 C / 2 ) f C o s α C o s α ( N 2 C / 2 ) f S i n α ( B + S i n α ( A + f ) )
Similarly, when α = 0 , the function f 2 ( x ) was as follows:
f 2 ( x ) = f ( N 2 C / 2 ) x + A + f B ( f ( N 2 C / 2 ) )
Point p was the intersection of functions f 1 ( x ) and f 2 ( x ) , and the y-coordinate of point p represented the depth value of the observed target. The fixed point p had coordinates ( x , y ) , f 1 ( x ) = a 1 x + b 1 , f 2 ( x ) = a 2 x + b 2 , and then the following relationship held:
x = b 1 b 2 a 2 a 1 y = a 2 b 1 a 1 b 2 a 2 a 1
When the camera performed up and down sweeps, there was an impact on depth perception. However, the vertical coordinates of the two cameras remained consistent, so the main impact was on the relative height, as shown in Figure 7.
Obviously, when there was a rotation angle along the vertical axis and the two cameras were parallel, the distance A relative to the mounting base plate ensured that the vertical rotation had little impact on the distance. Its effect was on the relative height of points o 1 and o 2 . This relative height could be derived by constructing the function in the same way, and afterward, the distance between point p and the mounting base plate could be obtained.

3. Experiments

To assess the technological advantages of the bionic eye in next-generation vision devices, this project has procured mainstream industry depth cameras for experimental comparison. The cameras include Intel Realsense SR300 Depth Camera:Working distance: 0.3–2 m, Depth Field of View (FOV): H73°, V59°, D90°, Frame rate: VGA 30 fps, Power consumption: 1.8 W; ORBBEC Dabai DW Depth Camera:Working distance: 0.2–2.5 m, Power consumption: 4.0 W, Depth FOV: H79°, V55°, D88.5°, Operating temperature: 10 °C to 40 °C; Fuayun A100 Binocular Camera, Maximum resolution: 2560 × 960 30 fps, Angle FOV: H52°, V49°, Working distance: 0.2–8 m, Power consumption: 3.6 W; SXHDR 300 Binocular Camera, Maximum resolution: 1920 × 1080 30 fps, Angle FOV: H80°, V92°, Working distance: 0.2–5 m, Power consumption: 4.7 W; JEWXON BC200 Bionic Eye Binocular Camera, Maximum resolution: 3296 × 2512 30 fps, Static angle FOV: H65°, V51°, Best dynamic angle FOV: H206°, V192°, Baseline Length: 35 mm, Distortion: <1%, Built-in IMU, Working distance: 0.2–10 m, Low Power consumption: 1.8 W. The experimental results comparing the performance of these four cameras with the project-developed BC200 bionic eye binocular camera are illustrated in Figure 8.
During the preliminary operational tests, it was found that the Intel SR300 depth camera and the ORBBEC Dabai DW depth camera had similar specifications, leading to the decision to phase out the SR300 camera. Additionally, the SXHDR 300 binocular camera was found to be inferior to the Fuayun A100 binocular camera in terms of resolution, viewing angle, and power consumption, resulting in the elimination of the SXHDR 300 binocular camera from further consideration.

3.1. Experimental Platform

To achieve optimal performance of the cameras, this experiment involves connecting the cameras to different system platforms as detailed in Figure 9. The JEWXON BC200 bionic eye binocular camera is connected via USB 3.0 to Nvidia’s Jetson Orin NX 16 GB embedded vision development board. The operating system used is Ubuntu 20.04, with CUDA version 11.4.19. The AI performance of the setup reaches 100 TOPS.
The Fuayun A100 binocular camera is connected via USB 3.0 to the Allwinnertech H6 embedded development board, which features a quad-core Cortex A53 processor, Mali-T720 GPU, and 2 GB LPDDR3 memory. The operating system used is Android 7.0, as shown in Figure 10.
The ORBBEC Dabai DW depth camera is connected via USB 3.0 to an ASUS FX50 laptop, which features an i7 CPU 12,700 H at 2.30 GHz, 16 GB of memory, and an RTX 3070 GPU. The operating system used is Windows 11, as shown in Figure 11.

3.2. Binocular Ranging Accuracy Experiment

Accurately obtaining distances in the world coordinate system is a major feature of binocular stereovision. For this experiment, a high-precision DELIXI brand laser rangefinder, model D100, was selected, which has a maximum measuring distance of 120 M, as shown in Figure 12. This device was used to measure the pose of the cameras, thus allowing for a comparison of the actual measurement accuracy of the cameras.
In this study, the laser rangefinder and the cameras being tested were aligned with a reference target (a chessboard) to perform depth imaging measurements. The lenses of the three cameras and the emitter of the laser rangefinder were kept at a horizontal level. The distances measured were as follows: the Fuayun A100 binocular camera was 1.328 M from the chessboard; the ORBBEC Dabai DW depth camera was 1.339 M from the chessboard; and the JEWXON BC200 bionic eye binocular camera was 1.338 M from the chessboard, as shown in Figure 13.

3.3. Camera Parameters and Error Testing

Since the ORBBEC Dabai DW depth camera comes factory-calibrated with built-in direct ranging capabilities, as shown in Figure 14, this experiment only compares the two binocular cameras. Table 5 presents the basic parameters of the two experimental cameras obtained through the Matlab Stereo Camera Calibrator tool. BC200 may deliver images with less distortion, which is beneficial for applications that require high precision and image fidelity. From the camera error experiments shown in Figure 15 and Figure 16, it is evident that the average error for the Fuayun A100 binocular camera is 3.53 pixels, while the JEWXON BC200 bionic eye binocular camera has an average error of 0.08 pixels.

3.4. Distance Testing

Each of the three cameras acquired images on their respective platforms, and after shooting, calibration, and correction, the following data were obtained: ORBBEC Dabai DW Depth Camera: The actual distance from the camera to the target object was 1.339 M, with the camera’s measurement data showing 1.269 M; Fuayun A100 Binocular Camera: The actual distance from the camera to the target object was 1.328 M, with the camera’s measurement data showing 1.015 M; JEWXON BC200 Bionic Eye Binocular Camera: The actual distance from the camera to the target object was 1.338 M, with the camera’s measurement data showing 1.299 M. The experimental results indicate that the JEWXON BC200 bionic eye binocular camera’s measurement accuracy is closer to the world coordinate system distance compared to the ORBBEC Dabai DW camera. However, the Fuayun A100 binocular camera recorded measurement data of 1.015 M, which shows a significant deviation from the actual distance, as depicted in Figure 17.

3.5. Bionic Binocular Camera Dynamic Viewing Angle Test

Due to the JEWXON BC200 bionic eye binocular camera’s capability to rotate freely, similar to human eyeballs, it maintains a wider field of view compared to other fixed-depth and binocular cameras, even when the camera itself is in a fixed position. This feature allows for more comprehensive visual coverage, as depicted in Figure 18.
In dynamic environments, objects do not always align perfectly with the camera’s field of view, making it crucial to measure depth data at various angles. However, binocular-vision systems inherently have installation errors that inevitably affect the calculation of depth values. Additionally, the inability to guarantee perfectly equal rotation angles for the left and right cameras is one source of error. As the bionic binocular camera’s rotation angle increases, the distance error caused by pixel differences at each point also increases. Minor pixel position errors can lead to significant distance measurement errors, causing errors to increase with the target distance. Using a basic model alone cannot achieve dynamic visual measurement acquisition.
The bionic binocular camera conducts real-time online calibration of images, which allows for the rapid acquisition of camera pose and image data. After obtaining left and right disparity images, the system corrects and processes them into images with depth information. However, during the rotation process of the binocular camera, using only the camera rotation model (Moda1) can cause image distortion during multi-angle motion shooting. This leads to incorrect pixel alignment, causing errors in the entire panoramic image composition.
On the other hand, the model where both the axis and camera rotate together (Moda2) results in images that are excessively smoothed during the JEWXON BC200 bionic eye binocular camera’s motion, reducing distortion and errors. The resulting panoramic images are more detailed, and the degree of image warping is significantly improved.
Experiments show that traditional binocular cameras experience increasing errors as the rotation angle increases. Furthermore, the improved model demonstrates smaller errors compared to the traditional model, proving the stability and effectiveness of the new model, as seen in Table 6. The comparison and results of these models and their impact on image quality and accuracy can be seen in Figure 19.

3.6. Target Detection Algorithm

In the field of machine vision, YOLO-V8 is a lightweight and highly efficient visual detector based on deep learning and computer vision, making it particularly suitable for embedded devices with strict power consumption requirements. YOLO-V8 outperforms other detectors such as YOLO-V5 and YOLO-V7 in both accuracy and speed and is well-suited for capturing targets in various complex scenarios. As an advanced and user-friendly object detection algorithm, it has attracted wide attention from researchers and developers [38]. This novel algorithm has been broadly applied in various fields, including autonomous driving, pedestrian tracking, robotics, and industrial inspection [39]. Therefore, this paper applies YOLO-V8 in conjunction with image data obtained from the JEWXON BC200 bionic eye binocular camera. As seen in Figure 20, after applying the algorithm that involves joint rotation of the axis and camera, the rotation process of the binocular camera is stable and smooth. The images are stable without significant jello effects or blurring. The accuracy of measurement and recognition is improved; it can accurately identify objects such as clocks, laptops, keyboards, mice, cups, pedestrians, and so on in the real world. This demonstrates that the bionic eye binocular camera not only provides a wider field of view and more stable images but also achieves high precision in autonomous target tracking, recognition, and ranging when combined with YOLO-V8.

3.7. Comparison of Existing Robotic Bionic Eye Devices

Binocular cameras are discussed in the article “Robot Bionic Vision Technologies: A Review” [40], as shown in Table 7. The bionic eye developed by Zou Wei et al. [41] consists of two CCD cameras and stepper motors, which essentially achieve the movement functions of the human eye. However, using large stepper motors results in an excessively large device with high power consumption, which is detrimental to mobility and portability. The bionic eye developed by Chen et al. [42] incorporates both long-focus and short-focus lenses in each eye, along with a three-degree-of-freedom neck mechanism integrated with an IMU. Although this design enhances perceptual capabilities, the overall size of the device and the high power consumption of the neck mechanism’s motor make it unsuitable for mobile applications. The JEWXON BC200, an improvement over previous generations of bionic eyes [43], features 4K HD micro-cameras and IMUs in a more compact form. It also utilizes the latest magnetic levitation conduction technology, allowing the eyeballs to rotate without cable interference, thus increasing the range of motion. Despite the significant reduction in overall power consumption, further improvements are needed to achieve additional functions such as head and neck movement.

3.8. Experimental Summary

The bionic eye can mimic the natural movements of the human eye and provide a wider field of view. When applied to humanoid robots, this enhanced vision technology significantly improves the robot’s visual perception capabilities, making it easier to track targets and make autonomous decisions in complex environments. The avoidance of repetitive head-turning motions allows for quicker responses in complex and dynamic scenarios. This paper introduces an innovative design for a bionic binocular-vision system, aimed at overcoming the limitations of existing technology by incorporating a model of binocular-vision systems with degrees of freedom, thereby enhancing the robot’s visual performance. Two models were developed: one involving only camera rotation and the other involving the rotation of both the axis and the camera. Various depth vision cameras were selected to conduct experiments on measurement accuracy and image quality. These experiments demonstrate that the designed bionic binocular-vision system and models not only simulate the flexible movement of human eyes but also significantly enhance the visual processing capabilities in dynamic environments by introducing degrees of freedom. The contributions of the proposed degree-of-freedom binocular-vision system models to the fields of machine vision and bionic eyes are as follows:
  • Camera Specifications: JEWXON BC200 Bionic Eye Binocular Camera: Features a maximum resolution of 3296 × 2512 at 30 fps, Static angle FOV: H65°, V51°, Best dynamic angle FOV: H206°, V192°, integrated IMU, a working distance of 0.2–10 m, and a power consumption of 2.8 W. This camera surpasses the control group’s other cameras in resolution, maximum viewing angle, effective working distance, and energy efficiency.
  • Vision System Combination: The JEWXON BC200, paired with the Jetson Orin NX, achieves an AI performance of 100 TOPS. This combination represents one of the more advanced embedded vision systems currently available, better suited for mobile robots due to its lower power consumption compared to PCs.
  • Error Experiment: The JEWXON BC200 bionic eye binocular camera has the best distortion value, the average error is 0.08 pixels, and the performance is excellent, performing better than the Fuayun A100 standard binocular camera.
  • Vision Ranging: The measurement accuracy of the JEWXON BC200 is closer to the world coordinate system distance compared to the ORBBEC Dabai DW depth camera. Using only the camera rotation model during the OpenCV panoramic image synthesis process can lead to image precision errors resulting in distortions and pixel misalignments that cause parts of the composite image to be missing. However, the combined rotation model of the axis and camera significantly improves image distortion, enhancing image precision and thus creating a more perfect composite image.
  • Enhancing Robotic Vision: To make robotic vision more akin to human vision, the introduction of YOLO-V8 improves the autonomous recognition ability and recognition accuracy of the bionic eye. The BC200 can accurately identify objects such as clocks, laptops, keyboards, mice, cups, pedestrians, and so on in the real world. The bionic eye binocular camera not only provides a broader field of view and more stable images but also achieves autonomous target tracking and precise identification, surpassing the capabilities of fixed binocular and structured light cameras.

4. Conclusions

In this study, we developed the JEWXON BC200 bionic binocular camera and validated its effectiveness and innovation through comparative evaluation experiments with commonly used depth cameras and bionic binocular cameras. The JEWXON BC200 features a maximum resolution of 3296 × 2512 at 30 fps, a static angle FOV of H65° and V51°, and an optimal dynamic field of view of H206° and V192°. It integrates an IMU, operates within a distance of 0.2 m to 10 m, and consumes only 2.8 W, outperforming other cameras in terms of resolution, viewing angle, working distance, and energy efficiency. Combined with the Jetson Orin NX, which offers AI performance of up to 100 TOPS, the JEWXON BC200 forms one of the most advanced vision systems for embedded platforms, particularly suited for mobile robots. The camera demonstrated superior distortion performance with an average error of 0.08 pixels and achieved higher measurement accuracy compared to the ORBBEC Dabai DW depth camera. When using a model that rotates both the axis and the camera, the image distortion and precision errors observed in fixed installations were significantly reduced, resulting in more accurate and seamless composite images. Additionally, incorporating YOLO-V8 enhanced the camera’s ability to accurately identify objects such as clocks, laptops, and pedestrians, making the BC200 not only smaller and more energy-efficient but also more capable in autonomous target tracking and precise recognition, surpassing the capabilities of fixed binocular and structured light cameras.

5. Future Work Focus

A.
The hardware platform for the binocular-vision system with degrees of freedom requires improvements, including adopting more advanced visual platforms, enhancing computational power while achieving lower operational power consumption, and increasing both installation and control precision. This will reduce the introduction of errors and enhance the accuracy of depth computation.
B.
While it is feasible to calculate distances using functional methods when the camera has degrees of freedom, it also introduces a considerable computational model that increases processing time; there is still room for future optimization.
C.
A solution is needed for the problem of absolute error variation with distance. One potential method to reduce errors is to attempt to replace the existing YOLO-V8 with a more precise and advanced target detection model.
The experiments conducted demonstrate that the designed bionic eye binocular-vision system offers better target depth calculation performance and a broader field of view compared to traditional binocular-vision systems. The experimental results confirm that the designed bionic binocular-vision system achieves flexible movements similar to human eyes while enhancing image quality and depth calculation accuracy during motion. This advancement not only enhances humanoid robots’ adaptability and application range in complex environments but also provides a crucial theoretical basis and technical pathway for the future development of robotic vision systems.

Author Contributions

Conceptualization, H.Z.; validation, S.L.; preparation, H.Z.; supervision, H.Z.; project administration, H.Z. All authors have read and agreed to the published version of the manuscript.

Funding

This study received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data presented in this study are available upon request from the corresponding author.

Acknowledgments

Thanks to Dongmin Jin for his contribution to the experimental pictures in this article.

Conflicts of Interest

Authors Hongxin Zhang and Suan Lee were employed by the company Jewxon Intelligent Technology Co., Ltd. The remaining authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

References

  1. Lu, H.M.; Li, Y.J.; Chen, M.; Kim, H.; Serikawa, S. Brain Intelligence: Go beyond Artificial Intelligence. Mobile Netw. Appl. 2018, 23, 368–375. [Google Scholar] [CrossRef]
  2. Bathla, G.; Bhadane, K.; Singh, R.K.; Kumar, R.; Aluvalu, R.; Krishnamurthi, R.; Kumar, A.; Thakur, R.N.; Basheer, S. Autonomous Vehicles and Intelligent Automation: Applications, Challenges, and Opportunities. Mob. Inf. Syst. 2022, 2022, 36. [Google Scholar] [CrossRef]
  3. Kober, J.; Bagnell, J.A.; Peters, J. Reinforcement learning in robotics: A survey. Int. J. Robot. Res. 2013, 32, 1238–1274. [Google Scholar] [CrossRef]
  4. Maimaitijiang, M.; Sagan, V.; Sidike, P.; Hartling, S.; Esposito, F.; Fritschi, F.B. Soybean yield prediction from UAV using multimodal data fusion and deep learning. Remote Sens. Environ. 2020, 237, 20. [Google Scholar] [CrossRef]
  5. Peng, Y.Y.; Yang, N.; Xu, Q.; Dai, Y.; Wang, Z.Q. Recent Advances in Flexible Tactile Sensors for Intelligent Systems. Sensors 2021, 21, 5392. [Google Scholar] [CrossRef]
  6. Li, G.; Xie, D.G.; Zhong, H.; Zhang, Z.Y.; Fu, X.K.; Zhou, Q.L.; Li, Q.; Ni, H.; Wang, J.O.; Guo, E.J.; et al. Photo-induced non-volatile VO2 phase transition for neuromorphic ultraviolet sensors. Nat. Commun. 2022, 13, 9. [Google Scholar] [CrossRef]
  7. Golnabi, H.; Asadpour, A. Design and application of industrial machine vision systems. Robot. Comput.-Integr. Manuf. 2007, 23, 630–637. [Google Scholar] [CrossRef]
  8. Perez, L.; Rodriguez, I.; Rodriguez, N.; Usamentiaga, R.; Garcia, D.F. Robot Guidance Using Machine Vision Techniques in Industrial Environments: A Comparative Review. Sensors 2016, 16, 335. [Google Scholar] [CrossRef]
  9. Xie, M.; Lai, T.; Fang, Y. A New Principle Toward Robust Matching in Human-like Stereovision. Biomimetics 2023, 8, 285. [Google Scholar] [CrossRef]
  10. LeCun, Y.; Bottou, L.; Bengio, Y.; Haffner, P. Gradient-based learning applied to document recognition. Proc. IEEE 1998, 86, 2278–2324. [Google Scholar] [CrossRef]
  11. Deng, J.; Dong, W.; Socher, R.; Li, L.J.; Li, K.; Li, F.F. ImageNet: A Large-Scale Hierarchical Image Database. In Proceedings of the IEEE-Computer-Society Conference on Computer Vision and Pattern Recognition Workshops, Miami Beach, FL, USA, 20–25 June 2009. [Google Scholar] [CrossRef]
  12. Krizhevsky, A.; Sutskever, I.; Hinton, G.E. Imagenet classification with deep convolutional neural networks. Adv. Neural Inf. Process. Syst. 2012, 25. [Google Scholar] [CrossRef]
  13. Simonyan, K.; Zisserman, A. Very deep convolutional networks for large-scale image recognition. arXiv 2014, arXiv:1409.1556. [Google Scholar]
  14. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 26 June–1 July 2016. [Google Scholar] [CrossRef]
  15. Hou, J.; You, B.; Xu, J.; Wang, T.; Cao, M. Surface Defect Detection of Preform Based on Improved YOLOv5. Appl. Sci. 2023, 13, 7860. [Google Scholar] [CrossRef]
  16. Zheng, Y.; Zhou, G.; Lu, B. A Multi-Scale Rebar Detection Network with an Embedded Attention Mechanism. Appl. Sci. 2023, 13, 8233. [Google Scholar] [CrossRef]
  17. LeCun, Y.; Kavukcuoglu, K.; Farabet, C. Convolutional Networks and Applications in Vision. In Proceedings of the International Symposium on Circuits and Systems Nano-Bio Circuit Fabrics and Systems (ISCAS 2010), Paris, France, 30 May–2 June 2010. [Google Scholar] [CrossRef]
  18. Wang, J.D.; Sun, K.; Cheng, T.H.; Jiang, B.R.; Deng, C.R.; Zhao, Y.; Liu, D.; Mu, Y.D.; Tan, M.K.; Wang, X.G.; et al. Deep High-Resolution Representation Learning for Visual Recognition. IEEE Trans. Pattern Anal. Mach. Intell. 2021, 43, 3349–3364. [Google Scholar] [CrossRef]
  19. Shijie, F.; Chao, Z.; Wei, Y.; Qian, C. Application of deep learning technology to fringe projection 3D imaging. Infrared Laser Eng. 2020, 49, 0303018-0303018-17. [Google Scholar] [CrossRef]
  20. Sun, X.; Jiang, Y.; Ji, Y.; Fu, W.; Yan, S.; Chen, Q.; Yu, B.; Gan, X. Distance measurement system based on binocular stereo vision. In Proceedings of the IOP Conference Series: Earth and Environmental Science, Xiamen, China, 7–9 June 2019. [Google Scholar] [CrossRef]
  21. Hansard, M.; Lee, S.; Choi, O.; Horaud, R.P. Time-of-Flight Cameras: Principles, Methods and Applications; Springer Science & Business Media: Berlin/Heidelberg, Germany, 2012. [Google Scholar]
  22. Xia, X.; Dai, S.; Qi, H.; Xu, Z.; Wang, S.; Zhang, M. Research on Object Measurement Based on 3D Stereo Vision. In Proceedings of the 2021 33rd Chinese Control and Decision Conference (CCDC), Kunming, China, 22–24 May 2021. [Google Scholar] [CrossRef]
  23. Zhang, S. High-speed 3D shape measurement with structured light methods: A review. Opt. Lasers Eng. 2018, 106, 119–131. [Google Scholar] [CrossRef]
  24. Zhang, S. High-Speed 3D Imaging with Digital Fringe Projection Techniques; Proc. SPIE 11813, Tribute to James C. Wyant: The Extraordinaire in Optical Metrology and Optics Education, 118130V (9 September 2021). Available online: https://www.spiedigitallibrary.org/conference-proceedings-of-spie/11813/118130V/High-speed-3D-imaging-with-digital-fringe-projection-techniques/10.1117/12.2567675.short (accessed on 14 June 2024).
  25. Rowley, D. A holographic interference camera. J. Phys. E Sci. Instrum. 1979, 12, 971. [Google Scholar] [CrossRef]
  26. Qian, J.F.; Ma, S.W.; Xu, Y.L.; Li, X.; Shen, Y.J. Application of Binocular Vision System to Face Detection and Tracking in Service Robot. Proc. SPIE 8335, 2012 International Workshop on Image Processing and Optical Engineering, 83350P (15 November 2011). Available online: https://www.spiedigitallibrary.org/conference-proceedings-of-spie/8335/83350P/Application-of-binocular-vision-system-to-face-detection-and-tracking/10.1117/12.918122.short (accessed on 14 June 2024).
  27. Shang, C.C.; Ma, H.W. Coal Mine Robot Binocular Vision Recognition System Based on Fuzzy Neural Network. In Proceedings of the International Conference on Automation, Mechanical and Electrical Engineering (AMEE), Phuket, Thailand, 26–27 July 2015. [Google Scholar] [CrossRef]
  28. Guo, S.X.; Chen, S.Z.; Liu, F.G.; Ye, X.F.; Yang, H.B. Binocular Vision-based Underwater Ranging Methods. In Proceedings of the IEEE International Conference on Mechatronics and Automation (ICMA), Takamatsu, Japan, 6–9 August 2017. [Google Scholar] [CrossRef]
  29. Sheng, H.Y.; Wei, S.M.; Yu, X.L.; Tang, L. Research on Binocular Visual System of Robotic Arm Based on Improved SURF Algorithm. IEEE Sens. J. 2020, 20, 11849–11855. [Google Scholar] [CrossRef]
  30. Noton, D.; Stark, L. Eye movements and visual perception. Sci. Am. 1971, 224, 34–43. [Google Scholar]
  31. Henderson, J.M. Eye movements. In The Oxford Handbook of Cognitive Psychology; Reisberg, D., Ed.; Oxford Library of Psychology; Oxford University Press: Hongkong, China, 2013; pp. 69–82. [Google Scholar]
  32. Cornsweet, T. Visual Perception; Academic Press: Cambridge, MA, USA, 2012. [Google Scholar]
  33. Gordon, I.E. Theories of Visual Perception; Psychology Press: London, UK, 2004. [Google Scholar] [CrossRef]
  34. Scassellati, B. Eye Finding via Face Detection for a Foveated Active Vision System; AAAI/IAAI: Cambridge, MA, USA, 1998. [Google Scholar]
  35. Murata, A. Ergonomics and cognitive engineering for robot-human cooperation. In Proceedings of the 9th IEEE International Workshop on Robot and Human Interactive Communication. IEEE RO-MAN 2000 (Cat. No.00TH8499), Osaka, Japan, 27–29 September 2000; pp. 206–211. [Google Scholar] [CrossRef]
  36. Breazeal, C. Socially intelligent robots: Research, development, and applications. In Proceedings of the 2001 IEEE International Conference on Systems, Man and Cybernetics. e-Systems and e-Man for Cybernetics in Cyberspace (Cat.No.01CH37236), Tucson, AZ, USA, 7–10 October 2001; Volume 4, pp. 2121–2126. [Google Scholar] [CrossRef]
  37. Shaw-Garlock, G. Looking forward to sociable robots. Int. J. Soc. Robot. 2009, 1, 249–260. [Google Scholar] [CrossRef]
  38. Lou, H.; Duan, X.; Guo, J.; Liu, H.; Gu, J.; Bi, L.; Chen, H. DC-YOLOv8: Small-Size Object Detection Algorithm Based on Camera Sensor. Electronics 2023, 12, 2323. [Google Scholar] [CrossRef]
  39. Wang, H.; Liu, C.; Cai, Y.; Chen, L.; Li, Y. YOLOv8-QSD: An Improved Small Object Detection Algorithm for Autonomous Vehicles Based on YOLOv8. IEEE Trans. Instrum. Meas. 2024, 73, 1–16. [Google Scholar] [CrossRef]
  40. Zhang, H.; Lee, S. Robot Bionic Vision Technologies: A Review. Appl. Sci. 2022, 12, 7970. [Google Scholar] [CrossRef]
  41. Wang, Q.; Yin, Y.; Zou, W.; Xu, D. Measurement error analysis of binocular stereo vision: Effective guidelines for bionic eyes. IET Sci. Meas. Technol. 2017, 11, 829–838. [Google Scholar] [CrossRef]
  42. Chen, X.; Wang, C.; Zhang, T.; Hua, C.; Fu, S.; Huang, Q. Hybrid Image Stabilization of Robotic Bionic Eyes. In Proceedings of the 2018 IEEE International Conference on Robotics and Biomimetics (ROBIO), Kuala Lumpur, Malaysia, 12–15 December 2018. [Google Scholar] [CrossRef]
  43. Zhang, H.; Lee, S. Robot Bionic Eye Motion Posture Control System. Electronics 2023, 12, 698. [Google Scholar] [CrossRef]
  44. Zhu, Z.; Wang, Q.; Zou, W.; Zhang, F. Motion Control on Bionic Eyes: A Comprehensive Review. arXiv 2019, arXiv:1901.01494. [Google Scholar]
Figure 1. Bionic binocular-vision system designed based on the principles of simulating human eye movements.
Figure 1. Bionic binocular-vision system designed based on the principles of simulating human eye movements.
Biomimetics 09 00371 g001
Figure 2. JEWXON BC200 Bionic eyes structure diagram.
Figure 2. JEWXON BC200 Bionic eyes structure diagram.
Biomimetics 09 00371 g002
Figure 3. The schematic of the inability of the monocular camera to determine depth.
Figure 3. The schematic of the inability of the monocular camera to determine depth.
Biomimetics 09 00371 g003
Figure 4. The schematic diagram of binocular-vision system.
Figure 4. The schematic diagram of binocular-vision system.
Biomimetics 09 00371 g004
Figure 5. The schematic diagram of the model with only the camera rotating.
Figure 5. The schematic diagram of the model with only the camera rotating.
Biomimetics 09 00371 g005
Figure 6. The schematic of the model of co-rotation of axis and camera.
Figure 6. The schematic of the model of co-rotation of axis and camera.
Biomimetics 09 00371 g006
Figure 7. The relationship between the eyeball and the observation point p during vertical movement.
Figure 7. The relationship between the eyeball and the observation point p during vertical movement.
Biomimetics 09 00371 g007
Figure 8. Camera evaluation photos.
Figure 8. Camera evaluation photos.
Biomimetics 09 00371 g008
Figure 9. JEWXON BC200 bionic eye binocular camera system platform.
Figure 9. JEWXON BC200 bionic eye binocular camera system platform.
Biomimetics 09 00371 g009
Figure 10. Fuayun A100 binocular camera system platform.
Figure 10. Fuayun A100 binocular camera system platform.
Biomimetics 09 00371 g010
Figure 11. ORBBEC Dabai DW depth camera system platform.
Figure 11. ORBBEC Dabai DW depth camera system platform.
Biomimetics 09 00371 g011
Figure 12. Laser rangefinder.
Figure 12. Laser rangefinder.
Biomimetics 09 00371 g012
Figure 13. Reference to the actual distance from the target to the test camera, field experimental environment diagram.
Figure 13. Reference to the actual distance from the target to the test camera, field experimental environment diagram.
Biomimetics 09 00371 g013
Figure 14. ORBBEC Dabai DW depth camera ranging experiment.
Figure 14. ORBBEC Dabai DW depth camera ranging experiment.
Biomimetics 09 00371 g014
Figure 15. Fuayun A100 binocular camera error experiment. Reprojection Errors Image (Image Pairs 8): left: camera 1, right: camera 2.
Figure 15. Fuayun A100 binocular camera error experiment. Reprojection Errors Image (Image Pairs 8): left: camera 1, right: camera 2.
Biomimetics 09 00371 g015
Figure 16. JEWXON BC200 bionic eye binocular camera error experiment. Reprojection Errors Image (Image Pairs 7): left: camera 1, right: camera 2.
Figure 16. JEWXON BC200 bionic eye binocular camera error experiment. Reprojection Errors Image (Image Pairs 7): left: camera 1, right: camera 2.
Biomimetics 09 00371 g016
Figure 17. Display of measurement accuracy for each camera.
Figure 17. Display of measurement accuracy for each camera.
Biomimetics 09 00371 g017
Figure 18. Multi-angle testing of the JEWXON BC200 bionic eye binocular camera after fixed installation. This test evaluates the performance of the JEWXON BC200 bionic eye binocular camera in a fixed installation, focusing on its ability to cover multiple angles due to its internal mechanisms that simulate human eye movements. The testing aims to demonstrate how the camera maintains a comprehensive field of vision across various orientations and conditions.
Figure 18. Multi-angle testing of the JEWXON BC200 bionic eye binocular camera after fixed installation. This test evaluates the performance of the JEWXON BC200 bionic eye binocular camera in a fixed installation, focusing on its ability to cover multiple angles due to its internal mechanisms that simulate human eye movements. The testing aims to demonstrate how the camera maintains a comprehensive field of vision across various orientations and conditions.
Biomimetics 09 00371 g018
Figure 19. Panoramic synthesis comparison: only camera rotates model vs. axis and camera rotate model.
Figure 19. Panoramic synthesis comparison: only camera rotates model vs. axis and camera rotate model.
Biomimetics 09 00371 g019
Figure 20. Bionic eye binocular camera combined with YOLO-V8 target detection and recognition effect.
Figure 20. Bionic eye binocular camera combined with YOLO-V8 target detection and recognition effect.
Biomimetics 09 00371 g020
Table 1. Performance comparison table of mainstream visual technology.
Table 1. Performance comparison table of mainstream visual technology.
Technology CategoryMonocular
Vision
Binocular Stereo
Vision
Structured LightBionic binocular
Vision
Product
pictures
Biomimetics 09 00371 i001Biomimetics 09 00371 i002Biomimetics 09 00371 i003Biomimetics 09 00371 i004
Technology principleBiomimetics 09 00371 i005Biomimetics 09 00371 i006Biomimetics 09 00371 i007Biomimetics 09 00371 i008
Principle
of work
Single cameraDual cameraCamera and infrared
projection patterns
Autonomous motion
dual camera
Response timeFastMediumSlowAlgorithm
determines speed
Weak lightWeakWeakGoodWeak
Bright lightGoodGoodWeakGood
Identification precisionLowLowMediumAlgorithm
determination accuracy
Resolving
capability
HighHighMediumHigh
Identification distanceMediumMediumVery shortMedium
Operation difficultyLowHighMediumHigh
CostLowMediumHighHigh
Power consumptionLowLowMediumLow
DisadvantagesLow recognition accuracy, poor dark lightDark light features are not obviousHigh requirements for ambient light, short recognition distanceAutomatic target tracking, super wide viewing Angle
Representative companyCognex, Honda, KeyenceLeapMoTion, FuayunIntel, Microsoft, ORBBECHuawei, Eyevolution, Jewxon
Data sources: Official websites of related products and patent databases.
Table 2. The coordinates of the points.
Table 2. The coordinates of the points.
Pointx-Coordinatey-Coordinate
o l 0 0
o l sin α f cos α f
o l 2 sin α f 2 cos α f
x l cos α N 1 C 2 sin α N 1 C 2
x l 2 sin α f + cos α N 1 C 2 2 cos α f sin α N 1 C 2
Table 3. The coordinates of the points of the left camera.
Table 3. The coordinates of the points of the left camera.
Pointx-Coordinatey-Coordinate
o 1 0 0
o l sin α A cos α A
o l sin α ( A + f ) cos α ( A + f )
o l sin α ( A + 2 f ) cos α ( A + 2 f )
x l sin α A cos α ( N 1 C 2 ) cos α A + sin α ( N 1 C 2 )
x l sin α ( A + 2 f ) + cos α ( N 1 C 2 ) cos α A + sin α ( N 1 C 2 )
Table 4. The coordinates of the points of the right camera.
Table 4. The coordinates of the points of the right camera.
Pointx-Coordinatey-Coordinate
o 2 B 0
o r B + sin α A cos α A
o r B + sin α ( A + f ) cos α ( A + f )
o r B + sin α ( A + 2 f ) cos α ( A + 2 f )
x r B + sin α A cos α ( N 1 C 2 ) cos α A + sin α ( N 1 C 2 )
x r B + sin α ( A + 2 f ) + cos α ( N 1 C 2 ) cos α ( A + 2 f ) sin α ( N 1 C 2 )
Table 5. Test group stereo camera data table.
Table 5. Test group stereo camera data table.
Metric/CameraFuayun A100JEWXON BC200Comparison
Focal Length (mm)Left: 417.0386,
Right: 456.7638
Left: 595.6271,
Right: 594.2564
BC200 has a longer focal length, suitable for long-distance shooting and providing a wider field of view.
Distortion
Coefficients
Left: [−0.057164, 0.095798, 0.011117, 0.000282, 0],
Right: [−0.006848, −0.054429, 0.010137, 0.001460, 0]
Left: [0.0403, −0.095, −0.001167, 0.000428, 0],
Right: [0.0299, −0.0322, −0.000880, −0.002340, 0]
BC200 has overall lower distortion coefficients, indicating higher image quality and less optical distortion.
Image Size (pixels)640 × 480640 × 480Equivalent test environment
Rotation Matrix StabilityLefit: [326.8532231, 252.4517895]
Right: [323.8062822, 256.9767981]
Lefit: [321.5142126,
181.3599961]
Right: [303.7775542
183.9382323]
BC200 may deliver images with less distortion, beneficial for applications that require high precision and image fidelity.
Translation Matrix[16.9321,0.0315, 1.4005][−48.7250, −0.2644, 1.8614]BC200 larger Z-axis offset is suitable for long-distance imaging.
Table 6. Comparison effect table of two models.
Table 6. Comparison effect table of two models.
Feature/MetricOnly Camera Rotates ModelAxis and Camera Rotate Model
Stability of ImageModerateHigh
Image DistortionSignificantMinimal
Coverage AreaLimitedExtensive
Resolution ConsistencyConsistentConsistent
Suitability for Dynamic ScenesModerateExcellent
Table 7. Comparison of existing robotic bionic eye devices.
Table 7. Comparison of existing robotic bionic eye devices.
Zou Wei et al. R&D Team [44]Chen et al. R&D Team [42]JEWXON BC200
Biomimetics 09 00371 i009Biomimetics 09 00371 i010Biomimetics 09 00371 i011
Features: The device consists of two CCD cameras and stepping motors, which can basically realize the movement function of the human eye.Features: each eye contains two cameras (long-focus lens and short-focus lens) for simulating the perception of human eye features. And a 3-degree-of-freedom neck mechanism is designed with an integrated IMU.Features: 4K HD mini-camera integrated in each eye, integrated IMU, and due to the latest levitation conduction technology used in the data collector, the eye rotates without being interfered by cables during the eye rotation, resulting in a larger angle of eye rotation.
Disadvantages: The model uses larger stepper motors, resulting in an excessively large product, and multiple large motors and loads running will inevitably lead to higher overall operating power consumption of the device, which is not conducive to mobility and portability.Disadvantages: Although the motor controlling the rotation of the eyeball is reduced, the overall design of the bionic eye shape is too large and a larger motor is used for the neck mechanism, which leads to higher power consumption of the device and is not suitable for use with mobile devices.Disadvantages: Although the overall use of mini cameras and mini motors, the overall power consumption drops a lot, but there is still room for improvement, such as the realization of the head and neck movement function.
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Zhang, H.; Lee, S. Advancing the Robotic Vision Revolution: Development and Evaluation of a Bionic Binocular System for Enhanced Robotic Vision. Biomimetics 2024, 9, 371. https://doi.org/10.3390/biomimetics9060371

AMA Style

Zhang H, Lee S. Advancing the Robotic Vision Revolution: Development and Evaluation of a Bionic Binocular System for Enhanced Robotic Vision. Biomimetics. 2024; 9(6):371. https://doi.org/10.3390/biomimetics9060371

Chicago/Turabian Style

Zhang, Hongxin, and Suan Lee. 2024. "Advancing the Robotic Vision Revolution: Development and Evaluation of a Bionic Binocular System for Enhanced Robotic Vision" Biomimetics 9, no. 6: 371. https://doi.org/10.3390/biomimetics9060371

APA Style

Zhang, H., & Lee, S. (2024). Advancing the Robotic Vision Revolution: Development and Evaluation of a Bionic Binocular System for Enhanced Robotic Vision. Biomimetics, 9(6), 371. https://doi.org/10.3390/biomimetics9060371

Article Metrics

Back to TopTop