Original Articles

Automated tea shoot picking using the YOLO network and Mamba images segmentation for top-view detection with a monocular camera

Publisher's note
All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.
Published: 18 November 2025
7
Views
8
Downloads

Authors

Detection and localization of tea shoots (one bud with two leaves) are critical steps in the automation of tea harvesting. Using red, green, blue-depth (RGB-D) camera to detect and locate tea shoots from side angles results in significant occlusion of tea shoots, as well as loss of depth information. To achieve automated, intelligent, and precise tea harvesting, this paper proposes a method for detecting and locating tea shoots from the top using a monocular camera. Firstly, the “You Only Look Once” (YOLO) network is employed to detect tea shoots regions in images collected by the monocular camera and to crop individual tea shoot top images. For these cropped images, a U-shaped images segmentation model based on Mamba is proposed. This model achieves a mean intersection over union (MIoU) of 87.80% and an accuracy (ACC) of 95.63%, precisely locating the specific tea shoots top regions. The center of the circumscribed circle of this region is used as the position for the next step in the picking process, accurately guiding the picking effector to the top of the tea shoot. Finally, the picking effector, controlled by feedback signals from infrared sensors, performs up-and-down reciprocation and cutting actions to complete the picking process. This method effectively avoids the problem of depth information loss during localization with RGB-D camera. To verify the effectiveness of the proposed approach, picking experiments were conducted on HouKui tea within a simulated tea garden environment, achieving a tea shoot picking success rate of 75.54%. The results indicate that this method offers significant application value and provides a new perspective for the development of automated tea shoots picking.

Downloads

Download data is not yet available.

Citations

Fu, L., Gao, F., Wu, J., Li, R., Karkee, M., Zhang, Q., 2020. Application of consumer RGB-D cameras for fruit detection and localization in field: A critical review. Comput. Electron. Agr. 177:105687. DOI: https://doi.org/10.1016/j.compag.2020.105687
Gu, A., Dao, T., 2024. Mamba: Linear-time sequence modeling with selective state spaces. arXiv:2312.00752.
Han, Y., Xiao, H., Qin, G., Song, Z., Ding, W., Mei, S., 2014. Developing situations of tea plucking machine. Engineering.6:268-273. DOI: https://doi.org/10.4236/eng.2014.66031
Kamilaris, A., Prenafeta-Boldú, F.X., 2018. Deep learning in agriculture: A survey. Comput. Electron. Agr. 147:70–90. DOI: https://doi.org/10.1016/j.compag.2018.02.016
Karunasena, G.M.K.B., Priyankara, H.D.N.S., 2020. Tea bud leaf identification by using machine learning and image processing techniques. Int. J. Sci. Eng. Res. 11:624-628. DOI: https://doi.org/10.14299/ijser.2020.08.02
Li, J., Gao, W., Wu, Y., Liu, Y., Shen, Y., 2022. High-quality indoor scene 3D reconstruction with RGB-D cameras: A brief review. Comput. Vis. Media 8:369-393. DOI: https://doi.org/10.1007/s41095-021-0250-8
Li, Y., He, L., Jia, J., Lv, J., Chen, J., Qiao, X., Wu, C., 2021. In-field tea shoot detection and 3D localization using an RGB-D camera. Comput. Electron. Agr. 185:106149. DOI: https://doi.org/10.1016/j.compag.2021.106149
Liu, L., Ouyang, W., Wang, X., Fieguth, P., Chen, J., Liu, X., Pietikäinen, M., 2020. Deep learning for generic object detection: a survey. Int. J. Comput. Vis. 128:261-318. DOI: https://doi.org/10.1007/s11263-019-01247-4
Long, Z., Jiang, Q., Wang, J., Zhu, H.L., Li, B., Wen, F., 2022. Research on method of tea flushes vision recognition and picking point localization. Transducers Microsyst. Technol. 41:39-41+45.
Mai, C.Y., Zheng, L.H., Sun, H., Yang, W., 2015. Research on 3D reconstruction of fruit tree and fruit recognition and location method based on RGB-D Camera. J. Agric. Mech. Eng. 46:35-40.
Santos, T.T., de Souza, L.L., dos Santos, A.A., Avila, S., 2020. Grape detection, segmentation, and tracking using deep neural networks and three-dimensional association. Comput. Electron. Agr. 170:105247. DOI: https://doi.org/10.1016/j.compag.2020.105247
Tang Y.P., Han W.M., Hu A.G., Wang W.Y., 2016. Design and experiment of intelligentized tea-plucking machine for human riding based on machine vision. Trans. Chin. Soc. Agric. Mach. 47:15–20.
Wang, C.Y., Bochkovskiy, A., Liao, H.Y.M., 2023. YOLOv7: Trainable bag-of-freebies sets new state-of-the-art for real-time object detectors. Proc. IEEE/CVF Conf. on Computer Vision and Pattern Recognition (CVPR), Vancouver. pp. 7464–7475. DOI: https://doi.org/10.1109/CVPR52729.2023.00721
Wang, J., Zhang, Z., Luo, L., Wei, H., Wang, W., Chen, M., Luo, S., 2023. DualSeg: Fusing transformer and CNN structure for image segmentation in complex vineyard environment. Comput. Electron. Agr. 206:107682. DOI: https://doi.org/10.1016/j.compag.2023.107682
Wang, Z.F., Lu, H.L., Geng, W.T., Sun, Z.Q., 2022. Research on object detection and positioning system of fruit picking robot based on OpenCV. Electr. Technol. Softw. Eng. 137-140.
Wu, S.P., Ca, Y.M., Wang, S., Fu, Q.Y., Feng, Y., Lv, L.Z., 2018. Effects of different picking periods on biochemical components and quality of Xinyang Maojian tea. Shandong Agric. Sci. 50:57-60.
Wu, X.M., Tang, X., Zhang, F.G., Gu, J.M., 2015. Tea buds image identification based on lab color model and K-means clustering. Trans. Chin. Soc. Agric. Machin. 36:161-164.
Yang, C.H., Liu, Y.P., Wang, Y., Xiong, L.Y., Xu, H.B., Zhao, W.H., 2019. Research and experiment on recognition and location system for citrus picking robot in natural environment. Trans. Chin. Soc. Agric. Machin. 50:14-22.
Yang, H., Chen, L., Chen, M., Ma, Z., Deng, F., Li, M., Li, X., 2019. Tender tea shoots recognition and positioning for picking robot using improved YOLO-V3 model. IEEE Access 7:180998-181011. DOI: https://doi.org/10.1109/ACCESS.2019.2958614
Zhang, L., Zhang, H., Chen, Y., Dai, S., Li, X., Kenji, I., Liu, Z., Li, M., 2019. Real-time monitoring of optimum timing for harvesting fresh tea leaves based on machine vision. Int. J. Agric. Biol. Eng. 12:6–9. DOI: https://doi.org/10.25165/j.ijabe.20191201.3418
Zhou, S.Y., Ca, L., 2022. Effects of different picking periods and tenderness of tea leaves on the quality of black tea. Chin. Food Saf. Mag. 149-152.
Zhu, H.C., Li, X., Meng, Y., Yang, H.B., Xu, Z., Li, Z.H., 2022. Tea bud detection based on Faster R CNN network. Trans. Chin. Soc. Agric. Mach. 53:217–224.
Zhu, L., Liao, B., Zhang, Q., Wang, Xinlong, Liu, W., Wang, Xinggang, 2024. Vision Mamba: Efficient visual representation learning with bidirectional state space model. arXiv:2401.09417.
Zou, Q., Lu, A.J., Zhou, H., Zhao, Q., 2022. An improved YOLOV3 algorithm model for tea bud detection. Laser J. 43:70-75.

Supporting Agencies

Anhui Key Laboratory of Smart Agricultural Technology and Equipment, The National Natural Science Foundation of China, Key Research and Development Project of Anhui Province
Wu Zhang, School of Information and Artificial Intelligence, Anhui Agricultural University, Hefei

Key Laboratory of Agricultural Sensors, Ministry of Agriculture and Rural Affairs, Hefei;
Anhui Province Key Laboratory of Smart Agricultural Technology and Equipment, Hefei, China

How to Cite



“Automated tea shoot picking using the YOLO network and Mamba images segmentation for top-view detection with a monocular camera” (2025) Journal of Agricultural Engineering [Preprint]. doi:10.4081/jae.2025.1637.