Original Articles
18 July 2025

GPPK4PCM: pest classification model integrating growth period prior knowledge

Publisher's note
All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.
7
Views
1
Downloads

Authors

Recent advancements in computer vision technology have significantly improved pest classification. However, pests of the same species exhibit distinct morphological changes throughout different life periods. Traditional methods apply the same feature extraction techniques to all periods, limiting classification precision. In addition to its inherent visual characteristics, pest images contain implicit growth period information. To address this issue, we propose a Pest Classification Model Integrating Growth Period Prior Knowledge. The model is composed of three sub-modules where: i) A deep learning network first identifies the growth periods of pests, and this prior knowledge is then used to guide the text encoder of the CLIP pre-trained model in generating period-specific textual features. ii) A parallel deep learning network extracts visual features from pest images. iii) An efficient low-rank multimodal fusion module integrates textual and visual features through parameter-optimized tensor decomposition, significantly improving classification accuracy across pest developmental phases. To evaluate its effectiveness, a dataset containing pests at different growth periods was constructed from Sichuan Agricultural University's pest dataset. Experimental results show that GPPK4PCM outperforms well-established deep learning neural networks. Compared to other advanced models, the proposed model excels in pest and disease classification tasks, effectively handling significant morphological differences across life periods.

Altmetrics

Downloads

Download data is not yet available.

Citations

Albattah, W., Masood, M., Javed, A., 2023. Custom CornerNet: a drone-based improved deep learning technique for large-scale multiclass pest localization and classification. Compl. Intell. Syst. 9:1299-1316. DOI: https://doi.org/10.1007/s40747-022-00847-x
Bai, J., Liu, X., Wang, Y., 2024. Integrating prior knowledge and contrast feature for signal modulation classification. IEEE Internet Things 11:21461-21473. DOI: https://doi.org/10.1109/JIOT.2024.3377916
Chollet, F., 2017. Xception: Deep learning with depthwise separable convolutions. IEEE Conf. on Computer Vision and Pattern Recognition (CVPR), Honolulu, 2017, pp. 1800-1807. DOI: https://doi.org/10.1109/CVPR.2017.195
Dai, G., Fan, J., Dewi, C., 2023. ITF-WPI: Image and text based cross-modal feature fusion model for wolfberry pest recognition. Comput. Electron. Agr. 212: 108129. DOI: https://doi.org/10.1016/j.compag.2023.108129
Deng, X., Feng, S., Lyu, G., 2022. Beyond word embeddings: Heterogeneous prior knowledge driven multi-label image classification. IEEE T. Multimedia 25: 4013-4025. DOI: https://doi.org/10.1109/TMM.2022.3171095
Dosovitskiy, A., Beyer, L., Kolesnikov, A., 2020. An image is worth 16x16 words: Transformers for image recognition at scale. arXiv:2010.11929.
Ebrahimi, M.A., Khoshtaghaza, M.H., Minaei, S., 2017. Vision-based pest detection based on SVM classification method. Comput. Electron. Agr. 137: 52-58. DOI: https://doi.org/10.1016/j.compag.2017.03.016
Guo, Q., Wang, C., Xiao, D., 2024. A lightweight open-world pest image classifier using ResNet8-based matching network and NT-Xent loss function. Expert Syst. Appl. 237:121395. DOI: https://doi.org/10.1016/j.eswa.2023.121395
Han, K., Wang, Y., Tian, Q., 2020. Ghostnet: More features from cheap operations. IEEE/CVF Conf. Computer Vision and Pattern Recognition (CVPR), Seattle, pp. 1577-1586. DOI: https://doi.org/10.1109/CVPR42600.2020.00165
He, K., Zhang, X., Ren, S., 2016. Deep residual learning for image recognition. IEEE/CVF Conf. Computer Vision and Pattern Recognition (CVPR), Las Vegas, pp. 770-778. DOI: https://doi.org/10.1109/CVPR.2016.90
Howard, A.G., Zhu, M., Chen, B., 2017. Mobilenets: Efficient convolutional neural networks for mobile vision applications. arXiv:1704.04861.
Kasinathan T, Singaraju D. 2021. Uyyala S R. Insect classification and detection in field crops using modern machine learning techniques. Available from: https://papers.nips.cc/paper_files/paper/2012/file/c399862d3b9d6b76c8436e924a68c45b-Paper.pdf
Khanramaki, M., Asli-Ardeh, E.A., Kozegar, E., 2021. Citrus pests classification using an ensemble of deep learning models. Comput. Electron. Agr. 186:106192. DOI: https://doi.org/10.1016/j.compag.2021.106192
Krizhevsky A, Sutskever I. 2012. Hinton G E. Imagenet classification with deep convolutional neural networks. Adv. Neural Information Processing Syst. 25.
Li, S., Sun, L., Li, Q., 2023. CLIP-ReID: exploiting vision-language model for image re-identification without concrete text labels. Proc. AAAI Conf. on Artificial Intelligence 37:1405-1413. DOI: https://doi.org/10.1609/aaai.v37i1.25225
Lin, Y., Chen, M., Wang, W., 2023. CLIP is also an efficient segmenter: A text-driven approach for weakly supervised semantic segmentation. IEEE/CVF Conf. Computer Vision and Pattern Recognition (CVPR), Vancouver, pp. 15305-15314. DOI: https://doi.org/10.1109/CVPR52729.2023.01469
Liu, W., Wu, G., Ren, F., 2020. DFF-ResNet: An insect pest recognition model based on residual networks. Big Data Mining Anal. 3:300-310. DOI: https://doi.org/10.26599/BDMA.2020.9020021
Liu, Z., Shen, Y., Lakshminarasimhan, V.B., 2018. Efficient low-rank multimodal fusion with modality-specific factors. Proc. 56th Annual Meet. Assoc. Computational Linguistics 1:2247-2256. DOI: https://doi.org/10.18653/v1/P18-1209
Lu, W., Wang, X., Jia, W. 2022. Root hair image processing based on deep learning and prior knowledge. Comput. Electron. Agr. 202:107397. DOI: https://doi.org/10.1016/j.compag.2022.107397
Radford, A., Kim, J.W., Hallacy, C., 2021. Learning transferable visual models from natural language supervision. Proc. 38th Int. Conf. Machine Learning, PMLR 8748-8763.
Schuler, J.P.S, Romani, S., Abdel-Nasser, M., Rashwan, H., Puig, D., 2022. Color-aware two-branch DCNN for efficient plant disease classification. Mendel 28:55-62. DOI: https://doi.org/10.13164/mendel.2022.1.055
Setiawan, A., Yudistira, N., Wihandika, R.C., 2022. Large scale pest classification using efficient Convolutional Neural Network with augmentation and regularizers. Comput. Electron. Agr. 200:107204. DOI: https://doi.org/10.1016/j.compag.2022.107204
Sichuan Agriculture University, 2020. Sichuan Agricultural University plant diseases and pests open dataset. Accessed 15 July 2020. Available from: https://github.com/SAUTEG/version_1.0
Simonyan, K., Zisserman, A., 2014. Very deep convolutional networks for large-scale image recognition. arXiv: 1409.1556.
Szegedy, C., Liu, W., Jia, Y., 2015. Going deeper with convolutions. IEEE/CVF Conf. Computer Vision and Pattern Recognition (CVPR), Boston, pp. 1-9. DOI: https://doi.org/10.1109/CVPR.2015.7298594
Szegedy, C., Ioffe, S., Vanhoucke, V., 2017. Inception-v4, inception-resnet and the impact of residual connections on learning. Proc. AAAI Conf. on Artificial Intelligence 31:4278-4284. DOI: https://doi.org/10.1609/aaai.v31i1.11231
Tan, M., Le, Q., 2019. Efficientnet: Rethinking model scaling for convolutional neural networks. Proc. 36th Int. Conf. Machine Learning, PMLR 6105-6114.
Tuda, M., Luna-Maldonado, A.I., 2020. Image-based insect species and gender classification by trained supervised machine learning algorithms. Ecol. Inform. 60:101135. DOI: https://doi.org/10.1016/j.ecoinf.2020.101135
Wang, Q., Wang, C., Lai, Z., 2024. Insectmamba: Insect pest classification with state space model. arXiv:2404.0361.
Wei, D., Chen, J., Luo, T., 2021. Classification of crop pests based on multi-scale feature fusion. Comput. Electron. Agr. 194:106736. DOI: https://doi.org/10.1016/j.compag.2022.106736
Xia, W., Han, D., Li, D., 2023. An ensemble learning integration of multiple CNN with improved vision transformer models for pest classification. Ann. Appl. Biol. 182:144-158. DOI: https://doi.org/10.1111/aab.12804
Yi, C., Ren, L., Zhan, D.C., 2024. Leveraging cross-modal neighbor representation for improved CLIP classification. IEEE/CVF Conf. Computer Vision and Pattern Recognition (CVPR), Seattle, pp. 27402-27411. DOI: https://doi.org/10.1109/CVPR52733.2024.02587
Zadeh, A., Chen, M., Poria, S., 2017. Tensor fusion network for multimodal sentiment analysis. arXiv:1707.07250. DOI: https://doi.org/10.18653/v1/D17-1115
Zhang, Y., Chen, L., Yuan, Y., 2023. Multimodal fine-grained transformer model for pest recognition. Electronics 12:2620. DOI: https://doi.org/10.3390/electronics12122620
Zhou, J., Li, J., Wang, C., 2021. Crop disease identification and interpretation method based on multimodal deep learning. Comput. Electron. Agr. 189:106408. DOI: https://doi.org/10.1016/j.compag.2021.106408
Jianhua Zheng, College of Information Science and Technology, Zhongkai University of Agriculture and Engineering, Guangzhou

Guangzhou Key Laboratory of Agricultural Products Quality & Safety Traceability Information Technology Zhongkai University of Agriculture and Engineering, Guangzhou;
Smart Agriculture Innovation Research Institute, Zhongkai University of Agriculture and Engineering, Guangzhou, China

Zhijie Luo, College of Information Science and Technology, Zhongkai University of Agriculture and Engineering, Guangzhou

Guangzhou Key Laboratory of Agricultural Products Quality & Safety Traceability Information Technology Zhongkai University of Agriculture and Engineering, Guangzhou;
Smart Agriculture Innovation Research Institute, Zhongkai University of Agriculture and Engineering, Guangzhou, China

How to Cite



“GPPK4PCM: pest classification model integrating growth period prior knowledge” (2025) Journal of Agricultural Engineering [Preprint]. doi:10.4081/jae.2025.1814.