Skip to main page content
U.S. flag

An official website of the United States government

Dot gov

The .gov means it’s official.
Federal government websites often end in .gov or .mil. Before sharing sensitive information, make sure you’re on a federal government site.

Https

The site is secure.
The https:// ensures that you are connecting to the official website and that any information you provide is encrypted and transmitted securely.

Access keys NCBI Homepage MyNCBI Homepage Main Content Main Navigation
. 2019 Jan 21;19(2):428.
doi: 10.3390/s19020428.

Guava Detection and Pose Estimation Using a Low-Cost RGB-D Sensor in the Field

Affiliations

Guava Detection and Pose Estimation Using a Low-Cost RGB-D Sensor in the Field

Guichao Lin et al. Sensors (Basel). .

Abstract

Fruit detection in real outdoor conditions is necessary for automatic guava harvesting, and the branch-dependent pose of fruits is also crucial to guide a robot to approach and detach the target fruit without colliding with its mother branch. To conduct automatic, collision-free picking, this study investigates a fruit detection and pose estimation method by using a low-cost red⁻green⁻blue⁻depth (RGB-D) sensor. A state-of-the-art fully convolutional network is first deployed to segment the RGB image to output a fruit and branch binary map. Based on the fruit binary map and RGB-D depth image, Euclidean clustering is then applied to group the point cloud into a set of individual fruits. Next, a multiple three-dimensional (3D) line-segments detection method is developed to reconstruct the segmented branches. Finally, the 3D pose of the fruit is estimated using its center position and nearest branch information. A dataset was acquired in an outdoor orchard to evaluate the performance of the proposed method. Quantitative experiments showed that the precision and recall of guava fruit detection were 0.983 and 0.948, respectively; the 3D pose error was 23.43° ± 14.18°; and the execution time per fruit was 0.565 s. The results demonstrate that the developed method can be applied to a guava-harvesting robot.

Keywords: RGB-D sensor; branch reconstruction; fully convolutional network; guava detection; pose estimation.

PubMed Disclaimer

Conflict of interest statement

The authors declare no conflict of interest.

Figures

Figure 1
Figure 1
The guava-harvesting robot and its vision sensing system. (a) Robot system; (b) vision sensing system.
Figure 2
Figure 2
Flow diagram of the developed vision sensing algorithm.
Figure 3
Figure 3
Fully convolutional network (FCN) configuration. The first row uses a deconvolution stride of 32, resulting in a coarse prediction. The second row fuses the outputs from the conv7 layer, the pool3 layer, and the pool4 layer at stride 8, leading to a finer prediction. The deconvolution parameter is defined as ‘(stride) × deconv’.
Figure 4
Figure 4
Segmentation results of the FCN model. (a) An aligned red–green–blue (RGB) image where black pixels represent objects outside the working range of the Kinect V2 sensor; (b) segmentation result where the red parts represent the fruits, and the green parts are the branches.
Figure 5
Figure 5
Fruit detection results. (a) Fruit point cloud extracted from Figure 4b; (b) clustering results, where each cluster is marked with a random color.
Figure 6
Figure 6
Branch reconstruction process. (a) Branch skeletons extracted from Figure 4b; (b) branch point cloud; (c) detected line segments, where each segment is marked with a random color.
Figure 7
Figure 7
Principle of fruit pose estimation. (a) Schematic diagram; (b) three-dimensional (3D) pose estimation result, where the red array represents the fruit pose.
Figure 8
Figure 8
Example showing ground-truth labels. (a) Ground-truth labels for three classes: fruit (blue), branch (red), and background. (b) Ground-truth fruit (blue) and the corresponding mother branch (red).
Figure 9
Figure 9
Examples illustrating unsuccessful detections.
Figure 10
Figure 10
Examples illustrating the fruit poses estimated by the proposed algorithm. The yellow array represents the fruit pose.
Figure 11
Figure 11
Failure examples. The yellow array represents the estimated pose, while the white array is the ground-truth pose.

References

    1. Dong Q., Murakami T., Nakashima Y. Recalculating the agricultural labor force in china. China Econ. J. 2018;11:151–169. doi: 10.1080/17538963.2018.1458431. - DOI
    1. Silwal A., Davidson J.R., Karkee M., Mo C., Zhang Q., Lewis K. Design, integration, and field evaluation of a robotic apple harvester. J. Field Robot. 2017;34:1140–1159. doi: 10.1002/rob.21715. - DOI
    1. Bac C.W., Hemming J., Van Tuijl J., Barth R., Wais E., Van Henten E.J. Performance evaluation of a harvesting robot for sweet pepper. J. Field Robot. 2017;36:1123–1139. doi: 10.1002/rob.21709. - DOI
    1. Kusumam K., Krajník T., Pearson S., Duckett T., Cielniak G. 3d-vision based detection, localization, and sizing of broccoli heads in the field. J. Field Robot. 2017;34:1505–1518. doi: 10.1002/rob.21726. - DOI
    1. Song Y., Glasbey C.A., Horgan G.W., Polder G., Dieleman J.A., van der Heijden G.W.A.M. Automatic fruit recognition and counting from multiple images. Biosyst. Eng. 2014;118:203–215. doi: 10.1016/j.biosystemseng.2013.12.008. - DOI

LinkOut - more resources