رویکردی متفاوت در تشخیص علایم خسارت تریپس روی برگ خیار به کمک بینایی ماشین مبتنی بر هوش مصنوعی

نوع مقاله : مقاله پژوهشی

نویسندگان

1 گروه مهندسی بیوسیستم، دانشکده کشاورزی، دانشگاه تبریز ، تبریز، ایران.

2 گروه گیاهپزشکی، دانشکده کشاورزی، دانشگاه تبریز، تبریز، ایران.

10.22034/jam.2026.68067.1332

چکیده

خیار از جمله محصولات پرطرفداری است که نقش مهمی در سبد غذایی مردم دارد. به همین دلیل، تولید آن باید همگام با میزان تقاضای جامعه مدیریت شود. با این حال، فرایند تولید خیار با چالش‌های متعددی همراه است. یکی از چالش‌های مطرح در این حوزه، خسارات ناشی از انواع آفات و بیماری‌ها می‌باشد. شناسایی سریع و دقیق آفات و بیماری‌ها نقش مهمی در مدیریت بهینه و افزایش عملکرد محصول خواهد داشت. استفاده از فناوری‌های پیشرفته بینایی ماشین و هوش مصنوعی می‌تواند جایگزین مؤثری برای تشخیص بیماری‌ها توسط نیروی انسانی باشد. در این پژوهش، برخلاف بسیاری از تحقیقات پیشین که بر طبقه‌بندی بیماری‌ها بر اساس تصاویر ثابت متمرکز بودند، رویکردی نوین مبتنی بر پردازش ویدئو بر اساس شبکه‌های عصبی کانولوشنی (CNNs) مورد بررسی قرار گرفته است. مجموعه تصاویر برگ خیار شامل برگ‌های سالم و برگ‌هایی با علائم ناشی از آلودگی به تریپس، به عنوان یکی از آفات رایج این گیاه، بود. پس از اعمال چند تکنیک پیش‌پردازش بر روی تصاویر، از دو مدل مبتنی بر شبکه عصبی کانولوشنیMovieNet  و SlowFast برای آموزش و ارزیابی طبقه‌بندی برگ سالم و برگ آلوده استفاده گردید. پیش‌پردازش شامل تکنیک‌های داده افزایی و نرمال‌سازی و چند فرآیند جهت تغذیه به مدل‌های مذکور بود. علاوه براین، تقسیم بندی دا‌ده‌ها به زیرمجموعه‌های آموزش، اعتبارسنجی و آزمایش از دیگر عملیات مربوط به پیش‌پردازش داده‌ها بود. تنظیم فراپارامترها و بهینه‌سازی توابع مختلف از جمله کارهایی بود که در حین آموزش مدل‌ها مورد مورد بررسی قرار گرفت. تحلیل نتایج در مراحل آموزش و ارزیابی نشان داد که مدل SlowFast عملکرد بسیار دقیقی در شناسایی و طبقه‌بندی علایم خسارت تریپس از برگ‌های سالم دارد. این مدل با دستیابی به دقت 27/99 درصد روی داده‌های آزمایشی، توانست بهترین عملکرد را ارائه و در مقایسه با مدل MovieNet نتایج برتری از خود نشان دهد.

کلیدواژه‌ها

موضوعات


عنوان مقاله [English]

A Novel AI-Based Machine Vision Approach for Detecting Thrips Damage Symptoms on Cucumber Leaves

نویسندگان [English]

  • Hossein Akhtari 1
  • Hossein Navid 1
  • Roghaiyeh Karimzadeh 2
  • Nayer Etminanfar 1
1 Department of Biosystem Engineering, Faculty of Agriculture, University of Tabriz, Tabriz. Iran.
2 Department of Plant Protection, Faculty of Agriculture, University of Tabriz, Tabriz, Iran.
چکیده [English]

Introduction
The integration of artificial intelligence and machine vision into agriculture has opened new horizons for the early detection of pests and plant diseases. These technologies are particularly valuable in greenhouse environments, where rapid intervention is essential to minimize crop losses. Traditional approaches have predominantly relied on static image classification, which lacks the ability to capture temporal dynamics of symptoms such as progressive leaf damage. In this context, this study explores the use of video-based deep learning for the detection of thrips induced damage on cucumber leaves. By comparing the performance of two modern spatiotemporal models MovieNet and SlowFast, the research aims to identify a reliable solution for real-time, accurate detection of pest damage in greenhouse conditions.
Materials and Methods
This study aimed to investigate the effectiveness of deep learning models, particularly convolutional neural networks, in detecting thrips-infested cucumber leaves using video-based input rather than still images. To this end, a dataset of 606 images was collected from the research greenhouse of the University of Tabriz, including 274 images of healthy leaves and 332 images of thrips-infested leaves.Given the widespread presence of this pest in the greenhouse, images were captured under natural conditions from various angles and distances to enhance data diversity and improve the model’s generalization to real-world scenarios. Care was also taken to ensure similar conditions when capturing images of healthy leaves to maintain class balance.
To simulate temporal dynamics and enable video-based learning, the collected images were converted into short video clips. Specifically, every four randomly selected images were combined sequentially to form one video. The frame rate was deliberately set to a low value of 3 frames per second to facilitate meaningful temporal feature extraction. After this preprocessing step and applying several offline data augmentation techniques, the final dataset comprised 909 videos, including 411 videos of healthy leaves and 498 videos of thrips-infected leaves.
For the learning task, two deep spatiotemporal architectures were employed MovieNet and SlowFast. Both models are known for their ability to capture motion and spatial patterns effectively. Prior to training, the video data were split into training (70%), validation (15%), and test (15%) sets using stratified sampling to preserve class distribution across subsets. All videos were resized and normalized according to the input requirements of the respective architectures. The models were trained with learning hyperparameters optimally tuned to ensure effective convergence and to minimize overfitting.
Performance evaluation was conducted using common classification metrics, including accuracy, precision, recall, and F1-score, computed on the test set. Additionally, confusion matrices and training-validation loss curves were analyzed to further assess model behavior during training and generalization capability.
Results and Discussion
Training-validation loss curves highlighted key differences between the two models. In the case of MovieNet, the training and validation loss both decreased rapidly at the beginning, indicating effective learning. However, during later epochs, the validation loss diverged from the training loss, suggesting overfitting. The model achieved 100% test accuracy, but this was considered unreliable due to the relatively small test set and the model’s tendency to memorize rather than generalize. Conversely, SlowFast demonstrated fluctuating loss values during the initial training phases, possibly due to its more complex architecture and optimization strategy. Despite the instability early on, both training and validation losses eventually converged, indicating improved generalization. This model achieved a final test accuracy of 99.27%, with a test loss of 0.0425, reflecting strong performance.
Detailed classwise evaluation revealed that the healthy leaf class achieved 98.41% precision and 100% recall, indicating that no healthy samples were misclassified. The thrips-damaged class recorded 100% precision and 98.67% recall, suggesting high detection accuracy with minimal false negatives. The overall F1-score, precision, and recall all exceeded 99%, confirming balanced and accurate performance across both classes.
The confusion matrix further validated these results. All 62 healthy samples were correctly classified, with zero misclassifications. Among the 75 thrips-damaged samples, 74 were correctly identified, with only one instance misclassified as healthy. This minimal error highlights the robustness of the SlowFast model in binary classification of pest damage.
Conclusion
This research demonstrates the efficacy of video-based deep learning methods for detecting thrips damage on cucumber leaves in greenhouse environments. Unlike conventional static image approaches, video enables the capture of dynamic changes and subtle visual cues over time, enhancing model accuracy and reliability.
Between the two models tested, SlowFast outperformed MovieNet, providing superior generalization and higher classification accuracy without overfitting. Its architectural design, particularly the dual-pathway temporal processing and ResNet-50 backbone, enabled it to achieve a final test accuracy of 99.27% and excellent precision-recall balance across both classes.
This video-based approach demonstrated several key advantages over traditional image-based methods, including enhanced accuracy through the capture of temporal symptom progression, reduced misclassification caused by static noise, and improved pattern recognition in dynamic real-world scenarios. These strengths highlight the potential of video-based deep learning techniques for integration into intelligent monitoring systems in modern greenhouses, offering farmers the ability to detect and respond to pest infestations more promptly and effectively
Future work should explore multi-class detection of various pests and diseases, as well as the incorporation of attention mechanisms or transformer-based video models to further improve accuracy. Additionally, developing mobile or cloud-based platforms for model deployment could make this technology more accessible for real-world agricultural applications.
Acknowledgement
The authors would like to thank the students working in the greenhouse for their cooperation and for allowing data collection during their research activities.

کلیدواژه‌ها [English]

  • Cucumber Thrips
  • Deep Learning
  • Machine Vision
  • Smart Agriculture
  • Artificial Intelligence in Agriculture
Akbar, J. U. M., Kamarulzaman, S. F., Muzahid, A. J. M., Rahman, M. A., & Uddin, M. (2024). A comprehensive review on deep learning assisted computer vision techniques for smart greenhouse agriculture. IEEE Access, 12, 4485-4522. https:/doi.org/10.1109/ACCESS.2024.3349418
Bora, D. J., Gupta, A. K., & Khan, F. A. (2015). Comparing the performance of L* A* B* and HSV color spaces with respect to color image segmentation. arXiv preprint arXiv:1506.01472. https:/doi.org/10.48550/arXiv.1506.01472
Cloyd, R. A. (2009). Western flower thrips (Frankliniella occidentalis) management on ornamental crops grown in greenhouses: have we reached an impasse. Pest Technology, 3(1), 1-9.
Deligeorgidis, P., Ipsilandis, C., Kaltsoudas, G., Giakalis, L., Petkou, D., & Deligeorgidis, N. (2006). The evaluation of damage caused by Frankliniella occidentalis (Pergande)(Thysanoptera: Thripidae) on cucumber grown in glass greenhouse. https:/doi.org/10.3923/je.2006.124.129
Escamilla-García, A., Soto-Zarazúa, G. M., Toledano-Ayala, M., Rivas-Araiza, E., & Gastélum-Barrios, A. (2020). Applications of artificial neural networks in greenhouse technology and overview for smart agriculture development. Applied Sciences, 10(11), 3835. https:/doi.org/10.3390/app10113835
Feichtenhofer, C., Fan, H., Malik, J., & He, K. (2019). Slowfast networks for video recognition. Proceedings of the IEEE/CVF international conference on computer vision, https://doi.org/10.48550/arXiv.1812.03982
Gerland, P., Raftery, A. E., Ševčíková, H., Li, N., Gu, D., Spoorenberg, T., Alkema, L., Fosdick, B. K., Chunn, J., & Lalic, N. (2014). World population stabilization unlikely this century. science, 346(6206), 234-237. https:/doi.org/10.1126/science.1257469
Guo, Q., Jin, S., Li, M., Yang, Q., Xu, K., Ju, Y., Zhang, J., Xuan, J., Liu, J., & Su, Y. (2020). Application of deep learning in ecological resource research: Theories, methods, and challenges. Science China Earth Sciences, 63, 1457-1474. https:/doi.org/10.1007/s11430-019-9584-9
He, K., Zhang, X., Ren, S., & Sun, J. (2016). Deep residual learning for image recognition. Proceedings of the IEEE conference on computer vision and pattern recognition, https://doi.org/10.48550/arXiv.1512.03385
Hearst, M. A., Dumais, S. T., Osuna, E., Platt, J., & Scholkopf, B. (1998). Support vector machines. IEEE Intelligent Systems and their applications, 13(4), 18-28. https:/doi.org/10.1109/5254.708428
Hecht-Nielsen, R. (1992). Theory of the backpropagation neural network. In Neural networks for perception (pp. 65-93). Elsevier. https:/doi.org/10.1016/B978-0-12-741252-8.50010-8
Howard, A., Sandler, M., Chu, G., Chen, L.-C., Chen, B., Tan, M., Wang, W., Zhu, Y., Pang, R., & Vasudevan, V. (2019). Searching for mobilenetv3. Proceedings of the IEEE/CVF international conference on computer vision, https://doi.org/10.48550/arXiv.1905.02244
Huang, Q., Xiong, Y., Rao, A., Wang, J., & Lin, D. (2020). Movienet: A holistic dataset for movie understanding. Computer Vision–ECCV 2020: 16th European Conference, Glasgow, UK, August 23–28, 2020, Proceedings, Part IV 16, https://doi.org/10.48550/arXiv.2007.10937
Hussain, N., Khan, M. A., Tariq, U., Kadry, S., Yar, M. A. E., Mostafa, A. M., Alnuaim, A. A., & Ahmad, S. (2022). Multiclass Cucumber Leaf Diseases Recognition Using Best Feature Selection. Computers, Materials & Continua, 70(2). https:/doi.org/10.32604/cmc.2022.019036
Kainat, J., Sajid Ullah, S., Alharithi, F. S., Alroobaea, R., Hussain, S., & Nazir, S. (2021). Blended Features Classification of Leaf‐Based Cucumber Disease Using Image Processing Techniques. Complexity, 2021(1), 9736179. https:/doi.org/10.1155/2021/9736179
Li, D., Wang, R., Xie, C., Liu, L., Zhang, J., Li, R., ... & Liu, W. (2020). A recognition method for rice plant diseases and pests video detection based on deep convolutional neural network. Sensors, 20(3), 578. https://doi.org/10.3390/s20030578
Li, J., Zhang, H., Wan, W., & Sun, J. (2020). Two-class 3D-CNN classifiers combination for video copy detection. Multimedia Tools and Applications, 79(7), 4749-4761. https:/doi.org/10.1007/s11042-018-6047-9
Li, K., Song, Y., Zhu, X., & Zhang, L. (2024). A severity estimation method for lightweight cucumber leaf disease based on DM-BiSeNet. Information Processing in Agriculture. https:/doi.org/10.1016/j.inpa.2024.03.003
Lin, K., Gong, L., Huang, Y., Liu, C., & Pan, J. (2019). Deep learning-based segmentation and quantification of cucumber powdery mildew using convolutional neural network. Frontiers in plant science, 10, 155. https:/doi.org/10.3389/fpls.2019.00155
Liu, J., & Wang, X. (2020). Tomato diseases and pests detection based on improved Yolo V3 convolutional neural network. Frontiers in plant science, 11, 898. https:/doi.org/10.3389/fpls.2020.00898
Liu, J., Wang, X., Chen, Q., Yan, P., & Liu, X. (2025). Deep learning method for cucumber disease detection in complex environments for new agricultural productivity. BMC Plant Biology, 25(1), 888. https://doi.org/10.1186/s12870-025-06841-y
Loshchilov, I., & Hutter, F. (2017). Decoupled weight decay regularization. arXiv preprint arXiv:1711.05101. https:/doi.org/10.48550/arXiv.1711.05101
Louppe, G. (2014). Understanding random forests: From theory to practice Universite de Liege (Belgium)]. https:/doi.org/10.48550/arXiv.1407.7502
Ma, J., Du, K., Zheng, F., Zhang, L., Gong, Z., & Sun, Z. (2018). A recognition method for cucumber diseases using leaf symptom images based on deep convolutional neural network. Computers and electronics in agriculture, 154, 18-24. https:/doi.org/10.1016/j.compag.2018.08.048
Marques, M. J. M., Mingxin, L., Shiojiri, K. T., Hagiwara, T., Hirose, K., Shiojiri, K., & Umezu, S. (2024). Plant Doctor: A hybrid machine learning and image segmentation software to quantify plant damage in video footage. arXiv preprint arXiv:2407.02853. https://doi.org/10.48550/arXiv.2407.02853
Mia, M. J., Maria, S. K., Taki, S. S., & Biswas, A. A. (2021). Cucumber disease recognition using machine learning and transfer learning. Bulletin of Electrical Engineering and Informatics, 10(6), 3432-3443. https:/doi.org/10.11591/EEI.V10I6.3096
Nateq Golestan, M., Shafaghi, F., & Sheikhi Gorjan, A. (2021). Efficacy of several insecticides against western flower Thrips Frankliniella occidentalis by introducing its Two new effectiveness indices in greenhouse cucumber. Journal of Applied Research in Plant Protection, 10(4), 17-23. https:/doi.org/10.22034/arpp.2021.12905
O'shea, K., & Nash, R. (2015). An introduction to convolutional neural networks. arXiv preprint arXiv:1511.08458. https:/doi.org/10.48550/arXiv.1511.08458
Ozguven, M. M. (2020). Deep learning algorithms for automatic detection and classification of mildew disease in cucumber. Fresenius Environ Bull, 29(8), 7081-7087.
Pandian, J. A., Kumar, V. D., Geman, O., Hnatiuc, M., Arif, M., & Kanchanadevi, K. (2022). Plant disease detection using deep convolutional neural network. Applied Sciences, 12(14), 6982. https://doi.org/10.3390/app12146982
Rahmawati, S., Devita, R., Zain, R. H., Rianti, E., Lubis, N., & Wanto, A. (2021). Prewitt and canny methods on inversion image edge detection: an evaluation. Journal of Physics: Conference Series, https://doi.org/10.1088/1742-6596/1933/1/012039
Ranganathan, G. (2021). A study to find facts behind preprocessing on deep learning algorithms. Journal of Innovative Image Processing, 3(1), 66-74. https:/doi.org/10.36548/jiip.2021.1.006
Rohlfs, C. (2025). Generalization in neural networks: A broad survey. Neurocomputing, 611, 128701. https:/doi.org/10.1016/j.neucom.2024.128701
Sandler, M., Howard, A., Zhu, M., Zhmoginov, A., & Chen, L.-C. (2018). Mobilenetv2: Inverted residuals and linear bottlenecks. Proceedings of the IEEE conference on computer vision and pattern recognition, https://doi.org/10.48550/arXiv.1801.04381
Shipp, J., Wang, K., & Binns, M. (2000). Economic injury levels for western flower thrips (Thysanoptera: Thripidae) on greenhouse cucumber. Journal of economic entomology, 93(6), 1732-1740. https:/doi.org/10.1603/0022-0493-93.6.1732
Simonyan, K., & Zisserman, A. (2014). Very deep convolutional networks for large-scale image recognition. arXiv preprint arXiv:1409.1556. https:/doi.org/10.48550/arXiv.1409.1556
Sun, Y., Cheng, H., Cheng, Q., Zhou, H., Li, M., Fan, Y., Shan, G., Damerow, L., Lammers, P. S., & Jones, S. B. (2017). A smart-vision algorithm for counting whiteflies and thrips on sticky traps using two-dimensional Fourier transform spectrum. Biosystems Engineering, 153, 82-88. https:/doi.org/10.1016/j.biosystemseng.2016.11.001
Szegedy, C., Vanhoucke, V., Ioffe, S., Shlens, J., & Wojna, Z. (2016). Rethinking the inception architecture for computer vision. Proceedings of the IEEE conference on computer vision and pattern recognition, https://doi.org/10.48550/arXiv.1512.00567
Uoc, N., Duong, N., Son, L., & Thanh, B. (2022). A novel automatic detecting system for cucumber disease based on the convolution neural network algorithm. GMSARN Int. J, 16(3), 295-301.
Yang, L., Huang, B., Mao, M., Yao, L., Niedermann, S., Hu, W., & Chen, Y. (2016). Sustainability assessment of greenhouse vegetable farming practices from environmental, economic, and socio-institutional perspectives in China. Environmental Science and Pollution Research, 23, 17287-17297. https://doi.org/10.1007/s11356-016-6937-1
 Yang, X., Liu, M., Xu, J., Zhao, L., Wei, S., Li, W., Chen, M., Chen, M., & Li, M. (2018). Image segmentation and recognition algorithm of greenhouse whitefly and thrip adults for automatic monitoring device. Transactions of the Chinese Society of Agricultural Engineering, 34(1), 164-170. https://doi.org/10.11975/j.issn.1002-6819.2018.01.022
Yu, C., Gao, C., Wang, J., Yu, G., Shen, C., & Sang, N. (2021). Bisenet v2: Bilateral network with guided aggregation for real-time semantic segmentation. International journal of computer vision, 129, 3051-3068. https://doi.org/10.48550/arXiv.2004.02147
Zhang, J., Rao, Y., Man, C., Jiang, Z., & Li, S. (2021). Identification of cucumber leaf diseases using deep learning and small sample size for agricultural Internet of Things. International Journal of Distributed Sensor Networks, 17(4), 15501477211007407. https://doi.org/10.1177/15501477211007407
Zhang, S., Wu, X., You, Z., & Zhang, L. (2017). Leaf image based cucumber disease recognition using sparse representation classification. Computers and electronics in agriculture, 134, 135-141. https://doi.org/10.1016/j.compag.2017.01.014