[go: up one dir, main page]

Skip to main content
Log in

Accurate detection of fresh and old vertebral compression fractures on CT images using ensemble YOLOR

  • 1237: Advanced Deep Learning for Computer Vision and Multimedia Applications
  • Published:
Multimedia Tools and Applications Aims and scope Submit manuscript

Abstract

Vertebral compression fractures resulting from osteoporosis contribute to pain and disability among older people, necessitating early detection and treatment. While MRI provides effective diagnosis, its higher cost poses a challenge. Conversely, CT imaging offers a more cost-effective alternative. However, CT-based detection of vertebral fractures may lack the accuracy of MRI. This paper presents a novel YOLO-based object detection method for localizing old and fresh fractures on spine CT images to expedite diagnosis and identify the optimal treatment window. We propose replacing the CSPDarknet53 backbone in the native YOLOR model with MobileViT and EfficientNet_NS, training three separate YOLOR models with different backbones. Subsequently, an ensemble approach is employed to leverage the enhanced feature extraction capabilities of the three models. Experimental results show individual model accuracies of 89%, 89.8%, and 89.2%, respectively. Furthermore, by replacing the convolution layer with the Involution layer and integrating the models using the ensemble method, we achieve a remarkable accuracy rate of 93.4%. The proposed Ensemble YOLOR model surpasses other state-of-the-art approaches, offering a fast and precise solution. This advancement provides valuable insights and guidance for physicians in diagnosing compression between old and new fractures.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Subscribe and save

Springer+ Basic
$34.99 /Month
  • Get 10 units per month
  • Download Article/Chapter or eBook
  • 1 Unit = 1 Article or 1 Chapter
  • Cancel anytime
Subscribe now

Buy Now

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Fig. 1
Fig. 2
Fig. 3
Fig. 4
Fig. 5
Fig. 6
Fig. 7
Fig. 8
Fig. 9

Similar content being viewed by others

Data availability

The data that support the findings of this study are not openly available due to IRB restrictions. Data are located in controlled access data storage at Dalin Tzu Chi Hospital.

References

  1. Cheng L-W, Chou H-H, Cai Y-X, Huang K-Y, Hsieh C-C, Chu P-L, Cheng I-S, Hsieh S-Y (2024) Automated detection of vertebral fractures from x-ray images: a novel machine learning model and survey of the field. Neurocomputing 566:126946

    Article  Google Scholar 

  2. Ono Y, Suzuki N, Sakano R, Kikuchi Y, Kimura T, Sutherland K, Kamishima T (2023) A deep learning-based model for classifying osteoporotic lum-bar vertebral fractures on radiographs: A retrospective model development and validation study. J Imaging 9(9):187

    Article  Google Scholar 

  3. Hong N, Cho SW, Shin S, Lee S, Jang SA, Roh S, Lee YH, Rhee Y, Cummings SR, Kim H et al (2020) Deep-learning-based detection of vertebral fracture and osteoporosis using lateral spine x-ray radiography. J Bone Miner Res 38(6):887–895

    Article  Google Scholar 

  4. Bar A, Wolf L, Bergman Amitai O, Toledano E, Elnekave E (2107) Compression fractures detection on ct. Soc Photo-Opt Instrumentation Eng (SPIE) Conf Seri 10134: 1013440. https://doi.org/10.1117/12.2249635

  5. Simonyan K, Zisserman A (2014) Very deep convolutional networks for large-scale image recognition. arXiv preprint arXiv:1409.1556

  6. Hochreiter S, Schmidhuber J (1997) Long short-term memory. Neural Comput 9(8):1735–1780. https://arxiv.org/abs/1409.1556

    Article  Google Scholar 

  7. Tomita N, Cheung YY, Hassanpour S (2018) Deep neural networks for automatic detection of osteoporotic vertebral fractures on ct scans. Comput Biol Med 98:8–15

    Article  Google Scholar 

  8. He K, Zhang X, Ren S, Sun J (2015) Deep residual learning for image recognition. Preprint at https://arxiv.org/abs/1512.03385

  9. Redmon J, Farhadi A (2016) YOLO9000: better, faster, stronger. Preprint at https://arxiv.org/abs/1612.08242

  10. Redmon J, Divvala S, Girshick R, Farhadi A (2015) You only look once: unified, real-time object detection. Preprint at https://arxiv.org/abs/1506.02640

  11. Yabu A, Hoshino M, Tabuchi H, Takahashi S, Masumoto H, Akada M, Morita S, Maeno T, Iwamae M, Inose H, Kato T, Yoshii T, Tsujio T, Terai H, Toyoda H, Suzuki A, Tamai K, Ohyama S, Hori Y, Okawa A, Nakamura H (2021) Using artificial intelligence to diagnose fresh osteoporotic vertebral fractures on magnetic resonance images. Spine J 21(10):1652–1658

    Article  Google Scholar 

  12. Mehta S, Rastegari, M (2021) MobileViT: lightweight, general-purpose, and mobile-friendly vision transformer. Preprint at https://arxiv.org/abs/2110.02178

  13. Xie Q, Luong M-T, Hovy E, Le QV (2019) Self-training with noisy student improves ImageNet classification. Preprint at https://arxiv.org/abs/1911.04252

  14. Wang C-Y, Yeh I-H, Liao H-YL (2021) You only learn one representation: unified network for multiple tasks. Preprint at https://arxiv.org/abs/2105.04206

  15. Wang C-Y, Bochkovskiy A, Liao H-YM (2020) Scaled-YOLOv4: scaling cross stage partial network. Preprint at https://arxiv.org/abs/2011.08036

  16. Wang C-Y, Liao H-YM, Yeh I-H, Wu Y-H, Chen P-Y, Hsieh J-W (2019) CSPNet: a new backbone that can enhance learning capability of CNN. Preprint at https://arxiv.org/abs/1911.11929

  17. Li D, Hu J, Wang C, Li X, She Q, Zhu L, Zhang T, Chen Q (2021) Involution: inverting the inherence of convolution for visual recognition. Preprint at https://arxiv.org/abs/2103.06255

  18. He K, Zhang X, Ren S, Sun J (2015) Spatial pyramid pooling in deep convolutional networks for visual recognition. IEEE Trans Pattern Anal Mach Intell 37(9):1904–1916

    Article  Google Scholar 

  19. Liu S, Qi L, Qin H, Shi J, Jia J (2018) Path aggregation network for instance segmentation. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 8759–8768

  20. Redmon J, Farhadi A (2018) YOLOv3: an incremental improvement. Preprint at https://arxiv.org/abs/1804.02767

  21. Bochkovskiy A, Wang C-Y, Liao H-YM (2020) YOLOv4: optimal speed and accuracy of object detection. Preprint at https://arxiv.org/abs/2004.10934

  22. Howard AG, Zhu M, Chen B, Kalenichenko D, Wang W, Weyand T, Andreetto M, Adam H (2017) Mobilenets: efficient convolutional neural networks for mobile vision applications. Preprint at https://arxiv.org/abs/1704.04861

  23. Sandler M, Howard A, Zhu M, Zhmoginov A, Chen L-C (2018) Mobilenetv2: inverted residuals and linear bottlenecks. Preprint at https://arxiv.org/abs/1801.04381

  24. Howard A, Sandler M, Chu G, Chen L-C, Chen B, Tan M, Wang W, Zhu Y, Pang R, Vasudevan V, Le QV, Adam H (2019) Searching for mobilenetv3. Preprint at https://arxiv.org/abs/1905.02244

  25. Han K, Wang Y, Tian Q, Guo J, Xu C, Xu C (2020) GhostNet: more Features from Cheap Operations. Preprint at https://arxiv.org/abs/1911.11907

  26. Dosovitskiy A, Beyer L, Kolesnikov A, Weissenborn D, Zhai X, Unterthiner T, Dehghani M, Minderer M, Heigold G, Gelly S, Uszkoreit J, Houlsby N (2020) An image is worth 16x16 words: Transformers for image recognition at scale. Preprint at https://arxiv.org/abs/2010.11929

  27. Lee D-H (2013) Pseudo-label: the simple and efficient semi-supervised learning method for deep neural networks. In: Proceedings of the international conference on machine learning, p 896

  28. Cubuk ED, Zoph B, Shlens J, Le QV (2020) Randaugment: practical automated data augmentation with a reduced search space. In: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition workshops, pp 702–703

  29. Tan M, Le QV (2019) Efficientnet: rethinking model scaling for convolutional neural networks. Preprint at https://arxiv.org/abs/1905.11946

  30. Zheng Z, Wang P, Liu W, Li J, Ye R, Ren D (2019) Distance-iou loss: faster and better learning for bounding box regression. Preprint at https://arxiv.org/abs/1911.08287

  31. Tan M, Pang R, Le QV (2019) Efficientdet: scalable and efficient object detection. Preprint at https://arxiv.org/abs/1911.09070

  32. Lin T-Y, Goyal P, Girshick R, He K, Doll'ar P (2017) Focal loss for dense object detection. Preprint at https://arxiv.org/abs/1708.02002

Download references

Acknowledgements

This work was partly supported by the Ministry of Science and Technology of Taiwan under grant 112-2221-E-224-033-MY3.

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Chuan-Yu Chang.

Ethics declarations

Conflict of interests

There is no conflict of interest.

Additional information

Publisher's Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Rights and permissions

Springer Nature or its licensor (e.g. a society or other partner) holds exclusive rights to this article under a publishing agreement with the author(s) or other rightsholder(s); author self-archiving of the accepted manuscript version of this article is solely governed by the terms of such publishing agreement and applicable law.

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Hsieh, MH., Chang, CY. & Hsu, SM. Accurate detection of fresh and old vertebral compression fractures on CT images using ensemble YOLOR. Multimed Tools Appl 83, 89375–89391 (2024). https://doi.org/10.1007/s11042-024-20355-z

Download citation

  • Received:

  • Revised:

  • Accepted:

  • Published:

  • Issue Date:

  • DOI: https://doi.org/10.1007/s11042-024-20355-z

Keywords

Navigation