Unveiling the Advancements: YOLOv7 vs YOLOv8 in Pulmonary Carcinoma Detection

Moulieswaran Elavarasu, Kalpana Govindaraju

Abstract


In this work, precision and recall measures are used to assess the performance of YOLOv7 and YOLOv8 models in identifying pulmonary carcinoma on a distinct collection of 700 photos. The necessity of early disease detection is increasing, thus choosing a reliable object detection model is essential. The goal of the research is to determine which model works best for this purpose, taking into account the unique difficulties that pulmonary cancer presents. The work makes a contribution to the field by showcasing the improvements made to YOLOv8 and underlining how well it detects both benign and malignant. YOLOv7 and YOLOv8 were used to independently train custom models using the pulmonary carcinoma dataset. The models' performance was measured using precision, recall, and mean average precision measures, which allowed for a comprehensive comparison examination. When it came to precision (58.2%), recall (61.2%), and mean average precision at both the 0.5:0.95 (33.3%) and 0.5 (53.3%) criteria, YOLOv8 outperformed YOLOv7. The 3.0% accuracy gain highlights YOLOv8's improved capabilities, especially in identifying small objects. YOLOv8's enhanced accuracy can be attributed to the optimisation of the detection process through its anchor-free design. According to this study, YOLOv8 is a more reliable model for pulmonary carcinoma identification than YOLOv7. The results indicate that YOLOv8 is the better option because of its higher recall, precision, and enhanced capacity to detect smaller objects—all of which are critical for early illness detection in medical imaging.


Keywords


YOLOv7; YOLOv8; Object Detection; Computer Vision; Pulmonary Carcinoma Detection; Medical Image Analysis.

Full Text:

PDF

References


A. Kumar, A. Kalia, A. Sharma, and M. Kaushal, “A hybrid tiny YOLO v4-SPP module based improved face mask detection vision system,” J Ambient Intell Humaniz Comput, vol. 14, no. 6, pp. 6783–6796, Jun. 2023, doi: 10.1007/s12652-021-03541-x.

O. Gedik and A. Demirhan, “Comparison of the effectiveness of deep learning methods for face mask detection,” Traitement du Signal, vol. 38, no. 4, pp. 947–953, Aug. 2021, doi: 10.18280/ts.380404.

Z. Zou, K. Chen, Z. Shi, Y. Guo, and J. Ye, “Object Detection in 20 Years: A Survey,” Proceedings of the IEEE, vol. 111, no. 3, pp. 257–276, Mar. 2023, doi: 10.1109/JPROC.2023.3238524.

A. Mohan, C. Papageorgiou, and T. Poggio, “Example-based object detection in images by components,” IEEE Trans Pattern Anal Mach Intell, vol. 23, no. 4, pp. 349–361, Apr. 2001, doi: 10.1109/34.917571.

Z. Li, H. Zhang, and J. Deng, “Traffic sign detection algorithm based on improved Faster R-CNN,” Chinese Journal of Liquid Crystal and Displays, vol. 36, no. 3, pp. 484–492, 2021, doi: 10.37188/CJLCD.2020-0195.

S. Ren, K. He, R. Girshick, and J. Sun, “Faster R-CNN: Towards Real-Time Object Detection with Region Proposal Networks,” IEEE Trans Pattern Anal Mach Intell, vol. 39, no. 6, pp. 1137–1149, Jun. 2017, doi: 10.1109/TPAMI.2016.2577031.

Y. Wang, X. Jia, M. Zhou, L. Xie, and Z. Tian, “A novel F-RCNN based hand gesture detection approach for FMCW systems,” Wireless Networks, pp. 1-14, Jul. 2019, doi: 10.1007/s11276-019-02096-2.

R. Girshick, “Fast R-CNN,” in 2015 IEEE International Conference on Computer Vision (ICCV), pp. 1440–1448, 2015, doi: 10.1109/ICCV.2015.169.

R. Girshick, J. Donahue, T. Darrell, and J. Malik, “Rich feature hierarchies for accurate object detection and semantic segmentation,” in Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition, pp. 580–587, 2014, doi: 10.1109/CVPR.2014.81.

Z. Tian, C. Shen, H. Chen, and T. He, “FCOS: Fully Convolutional One-Stage Object Detection,” in 2019 IEEE/CVF International Conference on Computer Vision (ICCV), pp. 9626–9635, 2019, doi: 10.1109/ICCV.2019.00972.

E. Daoud, N. Khalil, and M. Gaedke, “Implementation of a One-Stage Object Detection Solution to Detect Counterfeit Products,” IADIS International Journal on Computer Science and Information Systems, vol. 17, no. 1, pp. 37–49, 2022.

Z. Zheng, P. Wang, W. Liu, J. Li, R. Ye, and D. Ren, “Distance-IoU Loss: Faster and Better Learning for Bounding Box Regression,” Proceedings of the AAAI Conference on Artificial Intelligence, vol. 34, no. 7, pp. 12993–13000, 2020, doi: 10.1609/aaai.v34i07.6999.

J. Redmon, S. Divvala, R. Girshick, and A. Farhadi, “You only look once: Unified, real-time object detection,” in Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition, pp. 779–788, 2016, doi: 10.1109/CVPR.2016.91.

W. Liu et al., “SSD: Single Shot MultiBox Detector,” in European Conference on Computer Vision, pp. 21–37, 2016, doi: 10.1007/978-3-319-46448-0_2.

M. Tan, R. Pang, and Q. V. Le, “EfficientDet: Scalable and efficient object detection,” in Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition, pp. 10778–10787, 2020, doi: 10.1109/CVPR42600.2020.01079.

T. Y. Lin, P. Goyal, R. Girshick, K. He, and P. Dollar, “Focal Loss for Dense Object Detection,” IEEE Trans Pattern Anal Mach Intell, vol. 42, no. 2, pp. 318–327, Feb. 2020, doi: 10.1109/TPAMI.2018.2858826.

J. Redmon and A. Farhadi, “YOLO9000: Better, faster, stronger,” in Proceedings - 30th IEEE Conference on Computer Vision and Pattern Recognition, CVPR 2017, pp. 6517–6525, 2017, doi: 10.1109/CVPR.2017.690.

J. Redmon and A. Farhadi, “YOLOv3: An Incremental Improvement,” arXiv preprint arXiv:1804.02767, Apr. 2018.

C.-Y. Wang, A. Bochkovskiy, and H.-Y. M. Liao, “YOLOv7: Trainable bag-of-freebies sets new state-of-the-art for real-time object detectors,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 7464-7475 Jul. 2022.

C. Li et al., “YOLOv6: A Single-Stage Object Detection Framework for Industrial Applications,” arXiv preprint arXiv:2209.02976, Sep. 2022.

J. M. Górriz et al., “Artificial intelligence within the interplay between natural and artificial computation: Advances in data science, trends and applications,” Neurocomputing, vol. 410, pp. 237–270, Oct. 2020, doi: 10.1016/j.neucom.2020.05.078.

M. F. İnkaya and H. Gürkan, “A YOLOv3 Based Smart City Application For Children’s Playgrounds,” Journal of Innovative Science and Engineering (JISE), vol. 5, no. 1, pp. 25-40, Dec. 2020, doi: 10.38088/jise.813664.

U. Nepal and H. Eslamiat, “Comparing YOLOv3, YOLOv4 and YOLOv5 for Autonomous Landing Spot Detection in Faulty UAVs,” Sensors, vol. 22, no. 2, Jan. 2022, doi: 10.3390/s22020464.

B. Song, R. Li, X. Pan, X. Liu, and Y. Xu, “Improved YOLOv5 Detection Algorithm of Contraband in X-ray Security Inspection Image,” in 2022 5th International Conference on Pattern Recognition and Artificial Intelligence (PRAI), pp. 169–174, 2022, doi: 10.1109/PRAI55851.2022.9904110.

Z. Ge, S. Liu, F. Wang, Z. Li, and J. Sun, “YOLOX: Exceeding YOLO Series in 2021,” arXiv preprint arXiv:2107.08430, Jul. 2021.

J. Terven and D. Cordova-Esparza, “A Comprehensive Review of YOLO: From YOLOv1 and Beyond,” arXiv preprint arXiv:2304.00501, Apr. 2023.

J. Zhou, Y. Zhang, and J. Wang, “RDE-YOLOv7: An Improved Model Based on YOLOv7 for Better Performance in Detecting Dragon Fruits,” Agronomy, vol. 13, no. 4, p. 1042, Mar. 2023, doi: 10.3390/agronomy13041042.

Y. Lai, R. Ma, Y. Chen, T. Wan, R. Jiao, and H. He, “A Pineapple Target Detection Method in a Field Environment Based on Improved YOLOv7,” Applied Sciences, vol. 13, no. 4, p. 2691, Feb. 2023, doi: 10.3390/app13042691.

A. G. Howard et al., “MobileNets: Efficient Convolutional Neural Networks for Mobile Vision Applications,” arXiv preprint arXiv:1704.04861, Apr. 2017.

H. Rezatofighi, N. Tsoi, J. Gwak, A. Sadeghian, I. Reid, and S. Savarese, “Generalized Intersection over Union: A Metric and A Loss for Bounding Box Regression,” in Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp. 658-666, 2019.

K. J. Oguine, O. C. Oguine, and H. I. Bisallah, “YOLO v3: Visual and Real-Time Object Detection Model for Smart Surveillance Systems(3s),” in 2022 5th Information Technology for Education and Development (ITED), pp. 1-8, 2022.

T. Yin, X. Zhou, and P. Krähenbühl, “Center-based 3D Object Detection and Tracking,” in Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp. 11784-11793, Jun. 2020.

G. Jocher, A. Chaurasia, and J. Qiu. Ultralytics YOLO (Version 8.0.0) [Computer software], 2023, https://github.com/ultralytics/ultralytics.

Sahla Muhammed Ali, “Comparative Analysis of YOLOv3, YOLOv4 and YOLOv5 for Sign Language Detection,” International Journal Of Advance Research And Innovative Ideas In Education, vol. 7, no. 4, 2021.

A. Bochkovskiy, C.-Y. Wang, and H.-Y. M. Liao, “YOLOv4: Optimal Speed and Accuracy of Object Detection,” arXiv preprint arXiv:2004.10934, Apr. 2020.

C.-Y. Wang, A. Bochkovskiy, and H.-Y. M. Liao, “Scaled-YOLOv4: Scaling Cross Stage Partial Network,” in 2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 13024–13033, 2021, doi: 10.1109/CVPR46437.2021.01283.

C. Liu, H. Sui, J. Wang, Z. Ni, and L. Ge, “Real-Time Ground-Level Building Damage Detection Based on Lightweight and Accurate YOLOv5 Using Terrestrial Images,” Remote Sens (Basel), vol. 14, no. 12, p. 2763, Jun. 2022, doi: 10.3390/rs14122763.

W. Jia et al., “Real‐time automatic helmet detection of motorcyclists in urban traffic using improved YOLOv5 detector,” IET Image Process, vol. 15, no. 14, pp. 3623–3637, Dec. 2021, doi: 10.1049/ipr2.12295.

R. Li, Z. Ji, S. Hu, X. Huang, J. Yang, and W. Li, “Tomato Maturity Recognition Model Based on Improved YOLOv5 in Greenhouse,” Agronomy, vol. 13, no. 2, p. 603, Feb. 2023, doi: 10.3390/agronomy13020603.

Z. Cheng and Y. Li, “Improved YOLOv7 Algorithm for Detecting Bone Marrow Cells,” Sensors, vol. 23, no. 17, Sep. 2023, doi: 10.3390/s23177640.

B. Xiao, M. Nguyen, and W. Q. Yan, “Fruit ripeness identification using YOLOv8 model,” Multimed. Tools Appl., pp. 1-18, Aug. 2023, doi: 10.1007/s11042-023-16570-9.

G. Yang, J. Wang, Z. Nie, H. Yang, and S. Yu, “A Lightweight YOLOv8 Tomato Detection Algorithm Combining Feature Enhancement and Attention,” Agronomy, vol. 13, no. 7, Jul. 2023, doi: 10.3390/agronomy13071824.

C. Şaman and Ş. Çelikbaş, “YOLOv8-Based Lung Nodule Detection: A Novel Hybrid Deep Learning Model Proposal,” International Research Journal of Engineering and Technology, vol. 10, no. 8, pp. 230-237, 2023.

Y. Niu, W. Cheng, C. Shi, and S. Fan, “YOLOv8-CGRNet: A Lightweight Object Detection Network Leveraging Context Guidance and Deep Residual Learning,” Electronics, vol. 13, no. 1, p. 43, Dec. 2023, doi: 10.3390/electronics13010043.

D. Han, J. Kim, and J. Kim, “Deep Pyramidal Residual Networks,” in Proceedings of the IEEE conference on computer vision and pattern recognition, pp. 5927-5935, 2017.

D. Reis, J. Kupec, J. Hong, and A. Daoudi, “Real-Time Flying Object Detection with YOLOv8,” arXiv preprint arXiv: 2305.09972, 2023.

Z. Huang, L. Li, G. C. Krizek, and L. Sun, “Research on Traffic Sign Detection Based on Improved YOLOv8,” Journal of Computer and Communications, vol. 11, no. 7, pp. 226–232, 2023, doi: 10.4236/jcc.2023.117014.

Z. Gevorgyan, “SIoU Loss: More Powerful Learning for Bounding Box Regression,” arXiv preprint arXiv:2205.12740, May 2022.

J. Woo, J. H. Baek, S. H. Jo, S. Y. Kim, and J. H. Jeong, “A Study on Object Detection Performance of YOLOv4 for Autonomous Driving of Tram,” Sensors, vol. 22, no. 22, Nov. 2022, doi: 10.3390/s22229026.

A. Benjumea, I. Teeti, F. Cuzzolin, and A. Bradley, “YOLO-Z: Improving small object detection in YOLOv5 for autonomous vehicles,” arXiv preprint arXiv:2112.11798, Dec. 2021.

Y.-F. Zhang, W. Ren, Z. Zhang, Z. Jia, L. Wang, and T. Tan, “Focal and efficient IOU loss for accurate bounding box regression,” Neurocomputing, vol. 506, pp. 146–157, Sep. 2022, doi: 10.1016/j.neucom.2022.07.042.

Y. Guo et al., “Improved YOLOV4-CSP Algorithm for Detection of Bamboo Surface Sliver Defects With Extreme Aspect Ratio,” IEEE Access, vol. 10, pp. 29810–29820, 2022, doi: 10.1109/ACCESS.2022.3152552.

O. E. Olorunshola, M. E. Irhebhude, and A. E. Evwiekpaefe, “A Comparative Study of YOLOv5 and YOLOv7 Object Detection Algorithms,” Journal of Computing and Social Informatics, vol. 2, no. 1, pp. 1-12, 2023.

J. Terven, D. M. Córdova-Esparza, and J. A. Romero-González, “A comprehensive review of yolo architectures in computer vision: From yolov1 to yolov8 and yolo-nas,” Machine Learning and Knowledge Extraction, vol. 5, no. 4, pp. 1680-1716, 2023.

L. Nkuzo, M. Sibiya, and E. D. Markus, “A Comprehensive Analysis of Real-Time Car Safety Belt Detection Using the YOLOv7 Algorithm,” Algorithms, vol. 16, no. 9, p. 400, Aug. 2023, doi: 10.3390/a16090400.

C. Yi, J. Liu, T. Huang, H. Xiao, and H. Guan, “An efficient method of pavement distress detection based on improved YOLOv7,” Meas Sci Technol, vol. 34, no. 11, Aug. 2023, doi: 10.1088/1361-6501/ace929.

Y. Jiao, Z. Wang, Y. Shang, R. Li, Z. Hua, and H. Song, “Detecting endosperm cracks in soaked maize using μCT technology and R-YOLOv7-tiny,” Comput. Electron. Agric., vol. 213, p. 108232, Oct. 2023, doi: 10.1016/j.compag.2023.108232.

P. K. S and N. K. K, “Drone-based apple detection: Finding the depth of apples using YOLOv7 architecture with multi-head attention mechanism,” Smart Agricultural Technology, vol. 5, Oct. 2023, doi: 10.1016/j.atech.2023.100311.

L. Ma, L. Zhao, Z. Wang, J. Zhang, and G. Chen, “Detection and Counting of Small Target Apples under Complicated Environments by Using Improved YOLOv7-tiny,” Agronomy, vol. 13, no. 5, p. 1419, May 2023, doi: 10.3390/agronomy13051419.

G. Wei et al., “BFD-YOLO: A YOLOv7-Based Detection Method for Building Façade Defects,” Electronics (Switzerland), vol. 12, no. 17, Sep. 2023, doi: 10.3390/electronics12173612.

Y. Gao et al., “A Study on the Rapid Detection of Steering Markers in Orchard Management Robots Based on Improved YOLOv7,” Electronics (Switzerland), vol. 12, no. 17, Sep. 2023, doi: 10.3390/electronics12173614.

N. Gan et al., “YOLO-CID: Improved YOLOv7 for X-ray Contraband Image Detection,” Electronics (Switzerland), vol. 12, no. 17, Sep. 2023, doi: 10.3390/electronics12173636.

M. J. A. Soeb et al., “Tea leaf disease detection and identification based on YOLOv7 (YOLO-T),” Sci. Rep., vol. 13, no. 1, Dec. 2023, doi: 10.1038/s41598-023-33270-4.

K. Liu, Q. Sun, D. Sun, L. Peng, M. Yang, and N. Wang, “Underwater Target Detection Based on Improved YOLOv7,” J. Mar. Sci. Eng., vol. 11, no. 3, p. 677, Mar. 2023, doi: 10.3390/jmse11030677.

K. Li, Y. Wang, and Z. Hu, “Improved YOLOv7 for Small Object Detection Algorithm Based on Attention and Dynamic Convolution,” Applied Sciences, vol. 13, no. 16, p. 9316, Aug. 2023, doi: 10.3390/app13169316.

T. Kage, K. Sugimoto, and S. ichiro Kamata, “PCA based guided bilateral filter for medical color images,” in ACM International Conference Proceeding Series, pp. 142–148, 2019, doi: 10.1145/3326172.3326201.

S. Paris, “A gentle introduction to bilateral filtering and its applications,” in ACM SIGGRAPH 2007 courses, pp. 3-es, 2007.

C. Tomasi and R. Manduchi, “Bilateral Filtering for Gray and Color Images,” in Sixth International Conference on Computer Vision, pp. 839-846, 1998.

B. H. Chen, Y. S. Tseng, and J. L. Yin, “Gaussian-adaptive bilateral filter,” IEEE Signal Process Lett., vol. 27, pp. 1670–1674, 2020, doi: 10.1109/LSP.2020.3024990.

O. Ronneberger, P. Fischer, and T. Brox, “U-Net: Convolutional Networks for Biomedical Image Segmentation,” in Medical Image Computing and Computer-Assisted Intervention–MICCAI 2015: 18th International Conference, Munich, Germany, October 5-9, 2015, Proceedings, Part III 18, pp. 234-241, 2015.

H. C. Shin et al., “Deep Convolutional Neural Networks for Computer-Aided Detection: CNN Architectures, Dataset Characteristics and Transfer Learning,” IEEE Trans. Med. Imaging, vol. 35, no. 5, pp. 1285–1298, May 2016, doi: 10.1109/TMI.2016.2528162.

P. Adarsh, P. Rathi, and M. Kumar, “YOLO v3-Tiny: Object Detection and Recognition using one stage improved model,” in 6th International Conference on Advanced Computing and Communication Systems (ICACCS), pp. 687-694, 2020, doi: 10.1109/ICACCS48705.2020.9074315.

W. Fang, L. Wang, and P. Ren, “Tinier-YOLO: A Real-Time Object Detection Method for Constrained Environments,” IEEE Access, vol. 8, pp. 1935–1944, 2020, doi: 10.1109/ACCESS.2019.2961959.

D. Dua and C. Graff, “(UCI) Machine Learning Repository,” 2019.

S. G. Armato et al., “The Lung Image Database Consortium (LIDC) and Image Database Resource Initiative (IDRI): A completed reference database of lung nodules on CT scans,” Med. Phys., vol. 38, pp. 915–931, 2011.

M. Sokolova, N. Japkowicz, and S. Szpakowicz, “Beyond Accuracy, F-Score and ROC: A Family of Discriminant Measures for Performance Evaluation,” in AI 2006: Advances in Artificial Intelligence, pp. 1015–1021, 2006, doi: 10.1007/11941439_114.

V. Cerqueira, L. Torgo, and I. Mozetič, “Evaluating time series forecasting models: an empirical study on performance estimation methods,” Mach. Learn., vol. 109, no. 11, pp. 1997–2028, Nov. 2020, doi: 10.1007/s10994-020-05910-7.

P. Nanglia, S. Kumar, A. N. Mahajan, P. Singh, and D. Rathee, “A hybrid algorithm for lung cancer classification using SVM and Neural Networks,” ICT Express, vol. 7, no. 3, pp. 335–341, Sep. 2021, doi: 10.1016/j.icte.2020.06.007.

M. Everingham, L. Van Gool, C. K. I. Williams, J. Winn, and A. Zisserman, “The Pascal Visual Object Classes (VOC) Challenge,” Int. J. Comput. Vis., vol. 88, no. 2, pp. 303–338, Jun. 2010, doi: 10.1007/s11263-009-0275-4.

T.-Y. Lin et al., “Microsoft COCO: Common Objects in Context,” in Computer Vision – ECCV 2014, pp. 740–755, 2014, doi: 10.1007/978-3-319-10602-1_48.

A. Kirillov, R. Girshick, K. He, and P. Dollar, “Panoptic Feature Pyramid Networks,” in 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 6392–6401, 2019, doi: 10.1109/CVPR.2019.00656.

R. R. Cheela, “A Qualitative Review on Image Processing Algorithms to Detect Early Stage Lung Cancer,” International Journal of Advanced Technology and Engineering Exploration, vol. 6, pp. 83–87, 2017.

P. Dollar, C. Wojek, B. Schiele, and P. Perona, “Pedestrian Detection: An Evaluation of the State of the Art,” IEEE Trans Pattern Anal. Mach. Intell., vol. 34, no. 4, pp. 743–761, Apr. 2012, doi: 10.1109/TPAMI.2011.155.

J. Huang et al., “Speed/Accuracy Trade-Offs for Modern Convolutional Object Detectors,” in 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 3296–3297, 2017, doi: 10.1109/CVPR.2017.351.




DOI: https://doi.org/10.18196/jrc.v5i2.20900

Refbacks

  • There are currently no refbacks.


Copyright (c) 2024 Moulieswaran Elavarasu, Kalpana Govindaraju

Creative Commons License
This work is licensed under a Creative Commons Attribution-ShareAlike 4.0 International License.

 


Journal of Robotics and Control (JRC)

P-ISSN: 2715-5056 || E-ISSN: 2715-5072
Organized by Peneliti Teknologi Teknik Indonesia
Published by Universitas Muhammadiyah Yogyakarta in collaboration with Peneliti Teknologi Teknik Indonesia, Indonesia and the Department of Electrical Engineering
Website: http://journal.umy.ac.id/index.php/jrc
Email: jrcofumy@gmail.com


Kuliah Teknik Elektro Terbaik