Detection of multi-class coconut clusters for robotic picking under occlusion conditions

Yuxing Fu, Hongcheng Zheng, Zongbin Wang, Jinyang Huang, Wei Fu

Abstract


With the development of tree-climbing robots and robotic end-effectors, it is possible to develop automated coconutpicking
robots with the help of machine vision technology. Coconuts grow in clusters in the canopy and are easily occluded by
leaves. Therefore, the detection of multi-class coconut clusters according to the occlusion condition is necessary for robots to
develop picking strategies. The coconut detection model, named YOLO-Coco, was developed based on the YOLOv7-tiny
network. It detected coconuts in different conditions such as not-occluded, leaves-occluded, and trunk-occluded fruit. The
developed model used Efficient Channel Attention (ECA) to enhance the feature weights extracted by the backbone network.
Re-parameterization Convolution (RepConv) made the model convolution layers deeper and provided more semantic
information for the detection head. Finally, the Bi-directional Feature Pyramid Network (BiFPN) was used to optimize the head
network structure of YOLO-Coco to achieve the balanced fusion of multi-scale features. The results showed that the mean
average precision (mAP) of YOLO-Coco for detecting multi-class coconut clusters was 93.6%, and the average precision (AP)
of not-occluded, leaves-occluded, and trunk-occluded fruit were 90.5%, 93.8%, and 96.4%, respectively. The detection
accuracy of YOLO-Coco for yellow coconuts was 5.1% higher than that for green coconuts. Compared with seven mainstream
deep learning networks, YOLO-Coco achieved the highest detection accuracy in detecting multi-class coconut clusters, while
maintaining advantages in detection speed and model size. The developed model can accurately detect coconuts in complex
canopy environments, providing technical support for the visual system of coconut-picking robots.
Keywords: coconut clusters, picking robot, leaves-occluded, multi-class detection, YOLOv7-tiny
DOI: 10.25165/j.ijabe.20251801.9031

Citation: Fu Y X, Zheng H C, Wang Z B, Huang J Y, Fu W. Detection of multi-class coconut clusters for robotic picking
under occlusion conditions. Int J Agric & Biol Eng, 2025; 18(1): 267–278.

Keywords


coconut clusters, picking robot, leaves-occluded, multi-class detection, YOLOv7-tiny

References


Mat K, Abdul Kari Z, Rusli N D, Che Harun H, Wei L S, Rahman M M, et al. Coconut palm: Food, feed, and nutraceutical properties. Animals, 2022; 12(16): 2107.

Ru S F, Wang J F, Fan J Q. Design and parameter optimization of removing coconut fiber device by bionic steel wire roller brush based on characteristics of claw-toe. Transactions of the CSAE, 2018; 22(34): 27–35. (in Chinese)

Hainan Provincial Bureau of Statistics. Hainan Statistical Yearbook, 2023. Available: https://stats.hainan.gov.cn/tjj/tjsu/ndsj/2023/202311/P020231129587527396831.pdf. Accessed on [2023-12-02].

Caladcad J A, Cabahug S, Catamco M R, Villaceran P E, Cosgafa L, Cabizares K N, et al. Determining Philippine coconut maturity level using machine learning algorithms based on acoustic signal. Computers and Electronics in Agriculture, 2020; 172: 105327.

Ignacio I F, Miguel T S. Research opportunities on the coconut (Cocos nucifera L.) using new technologies. South African Journal of Botany, 2021; 141: 414–420.

Zhou H Y, Wang X, Au W, Kang H W, Chen C. Intelligent robots for fruit harvesting: Recent developments and future challenges. Precision Agriculture, 2022; 23(5): 1856–1907.

Gené-Mola J, Vilaplana V, Rosell-Polo J R, Morros J R, Ruiz-Hidalgo J, Gregorio E. Multi-modal deep learning for Fuji apple detection using RGB-D cameras and their radiometric capabilities. Computers and Electronics in Agriculture, 2019; 162: 689–698.

Megalingam R K, Manoharan S K, Mohandas S M, Vadivel S R R, Gangireddy R, Ghanta S, et al. Amaran: An unmanned robotic coconut tree climber and harvester. IEEE-ASME Transactions on Mechatronics, 2021; 26(1): 288–299.

Wibowo T S, Sulistijono I A, Risnumawan A. End-to-end coconut harvesting robot. In: 2016 International Electronics Symposium (IES), 2016; pp.444–449. doi: 10.1109/ELECSYM.2016.7861047.

Parvathi S, Selvi S T. Detection of maturity stages of coconuts in complex background using Faster R-CNN model. Biosystems Engineering, 2021; 202: 119–132.

Fu L S, Majeed Y, Zhang X, Karkee M, Zhang Q. Faster R-CNN-based apple detection in dense-foliage fruiting-wall trees using RGB and depth features for robotic harvesting. Biosystems Engineering, 2020; 197: 245–256.

Tang Y C, Zhou H, Wang H J, Zhang Y Q. Fruit detection and positioning technology for a Camellia oleifera C. Abel orchard based on improved YOLOv4-tiny model and binocular stereo vision. Expert Systems with Applications, 2023; 211: 118573.

Zhu X Y, Chen F J, Zhang X W, Zheng Y L, Peng X D, Chen C. Detection the maturity of multi-cultivar olive fruit in orchard environments based on Olive-EfficientDet. Scientia Horticulturae, 2024; 324: 112607.

Patrício D I, Rieder R. Computer vision and artificial intelligence in precision agriculture for grain crops: A systematic review. Computers and Electronics in Agriculture, 2018; 153: 69–81.

Shaikh T A, Rasool T, Lone F R. Towards leveraging the role of machine learning and artificial intelligence in precision agriculture and smart farming. Computers and Electronics in Agriculture, 2022; 198: 107119.

Zhu X Y, Li J J, Jia R C, Liu B, Yao Z H, Yuan A H, et al. LAD-Net: A novel light weight model for early apple leaf pests and diseases classification. IEEE/ACM Transactions on Computational Biology and Bioinformatics, 2023; 20(2): 1156–1169.

Wang Z, Hua Z X, Wen Y C, Zhang S J, Xu X S, Song H B. E-YOLO: Recognition of estrus cow based on improved YOLOv8n model. Expert Systems with Applications, 2024; 238: 122212.

Montoya-Cavero L E, Torres R D D, Gómez-Espinosa A, Cabello J A E. Vision systems for harvesting robots: Produce detection and localization. Computers and Electronics in Agriculture, 2022; 192: 106562.

Yang Y Y, Han Y X, Li S, Yang Y D, Zhang M, Li H. Vision based fruit recognition and positioning technology for harvesting robots. Computers and Electronics in Agriculture, 2023; 213: 108258.

Zhang B, Wang R R, Zhang H M, Yin C H, Xia Y Y, Fu M, et al. Dragon fruit detection in natural orchard environment by integrating lightweight network and attention mechanism. Frontiers in Plant Science, 2022; 13: 1040923.

Zhang J, Karkee M, Zhang Q, Zhang X, Yaqoob M, Fu L S, et al. Multi-class object detection using faster R-CNN and estimation of shaking locations for automated shake-and-catch apple harvesting. Computers and Electronics in Agriculture, 2020; 173: 105384.

Gao F, Fu L S, Zhang X, Majeed Y, Li R, Karkee M, Zhang Q. Multi-class fruit-on-plant detection for apple in SNAP system using Faster R-CNN. Computers and Electronics in Agriculture, 2020; 176: 105634.

Song H B, Shang Y Y, He D J. Review on deep learning technology for fruit target recognition. Transactions of the CSAM, 2023; 54(1): 1–19. (in Chinese)

Peng H X, Huang B, Shao Y Y, Li Z S, Zhang C W, Chen Y, et al. General improved SSD model for picking object recognition of multiple fruits in natural environment. Transactions of the CSAE, 2018; 16(34): 155–162. (in Chinese)

Redmon J, Divvala S, Girshick R, Farhadi A. You Only Look Once: Unified, real-time object detection. In: 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas: IEEE, 2016; pp.779–788. doi: 10.1109/CVPR.2016.91.

Li T H, Sun M, He Q H, Zhang G S, Shi G, Ding X, et al. Tomato recognition and location algorithm based on improved YOLOv5. Computers and Electronics in Agriculture, 2023; 208: 107759.

Xu L J, Wang Y H, Shi X S, Tang Z L, Chen X Y, Wang Y C, et al. Real-time and accurate detection of citrus in complex scenes based on HPL-YOLOv4. Computers and Electronics in Agriculture, 2023; 205: 107590.

Suo R, Gao F F, Zhou Z X, Fu L S, Song Z Z, Dhupia J, et al. Improved multi-classes kiwifruit detection in orchard to avoid collisions during robotic picking. Computers and Electronics in Agriculture, 2021; 182: 106052.

Zhang F, Chen Z J, Ali S, Yang N, Fu S L, Zhang Y K. Multi-class detection of cherry tomatoes using improved YOLOv4-Tiny. Int J Agric & Biol Eng, 2023; 16(2): 225–231.

Wang C Y, Bochkovskiy A, Liao H Y M. YOLOv7: Trainable Bag-of-Freebies Sets New State-of-the-Art for Real-Time Object Detectors. In: 2023 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2023; pp.7464–7475. doi: 10.1109/CVPR52729.2023.00721.

Wang Y, Yao X Z, Li B, Xu S, Yi Z F, Zhao J H. Malformed sweet pepper fruit identification algorithm based on improved YOLOv7-tiny. Transactions of the CSAM, 2023; 11(54): 236–246. (in Chinese)

Wang Q L, Wu B G, Zhu P F, Li P H, Zuo W M, Hu Q H. ECA-Net: Efficient Channel Attention for Deep Convolutional Neural Networks. In: 2020 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2020; pp.11531–11539. doi: 10.1109/CVPR42600.2020.01155.

Li Q M, Han Z C, Wu X M. Deeper insights into graph convolutional networks for semi-supervised learning. In: 32nd AAAI Conference on Artificial Intelligence, 2018; pp.3538–3545. doi: 10.48550/arXiv.1801.07606.

Ding X H, Zhang X Y, Ma N N, Han J G, Ding G G, Sun J. RepVGG: Making VGG-style ConvNets Great Again. In: 2021 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2021; pp.13728–13737. doi: 10.1109/CVPR46437.2021.01352.

Tan M X, Pang R Y, Le Q V. EfficientDet: Scalable and Efficient Object Detection. In: 2020 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2020; pp.10778–10787. doi: 10.1109/CVPR42600.2020.01079.

Yao T, Tan Z P, Cheng E, Wu L G. Method for daylily detection and segmentation based on improved YOLOv7-seg. Transactions of the CSAE, 2024; 40(9): 146–153. (in Chinese)

Divyanth L G, Soni P, Pareek C M, Machavaram R, Nadimi M, Paliwal J, et al. Detection of coconut clusters based on occlusion condition using attention-guided Faster R-CNN for robotic harvesting. Foods, 2022; 11(23): 3903.




Copyright (c) 2025 International Journal of Agricultural and Biological Engineering

Creative Commons License
This work is licensed under a Creative Commons Attribution 4.0 International License.

2023-2026 Copyright IJABE Editing and Publishing Office