Abstract
Vertebral compression fractures resulting from osteoporosis contribute to pain and disability among older people, necessitating early detection and treatment. While MRI provides effective diagnosis, its higher cost poses a challenge. Conversely, CT imaging offers a more cost-effective alternative. However, CT-based detection of vertebral fractures may lack the accuracy of MRI. This paper presents a novel YOLO-based object detection method for localizing old and fresh fractures on spine CT images to expedite diagnosis and identify the optimal treatment window. We propose replacing the CSPDarknet53 backbone in the native YOLOR model with MobileViT and EfficientNet_NS, training three separate YOLOR models with different backbones. Subsequently, an ensemble approach is employed to leverage the enhanced feature extraction capabilities of the three models. Experimental results show individual model accuracies of 89%, 89.8%, and 89.2%, respectively. Furthermore, by replacing the convolution layer with the Involution layer and integrating the models using the ensemble method, we achieve a remarkable accuracy rate of 93.4%. The proposed Ensemble YOLOR model surpasses other state-of-the-art approaches, offering a fast and precise solution. This advancement provides valuable insights and guidance for physicians in diagnosing compression between old and new fractures.









Similar content being viewed by others
Data availability
The data that support the findings of this study are not openly available due to IRB restrictions. Data are located in controlled access data storage at Dalin Tzu Chi Hospital.
References
Cheng L-W, Chou H-H, Cai Y-X, Huang K-Y, Hsieh C-C, Chu P-L, Cheng I-S, Hsieh S-Y (2024) Automated detection of vertebral fractures from x-ray images: a novel machine learning model and survey of the field. Neurocomputing 566:126946
Ono Y, Suzuki N, Sakano R, Kikuchi Y, Kimura T, Sutherland K, Kamishima T (2023) A deep learning-based model for classifying osteoporotic lum-bar vertebral fractures on radiographs: A retrospective model development and validation study. J Imaging 9(9):187
Hong N, Cho SW, Shin S, Lee S, Jang SA, Roh S, Lee YH, Rhee Y, Cummings SR, Kim H et al (2020) Deep-learning-based detection of vertebral fracture and osteoporosis using lateral spine x-ray radiography. J Bone Miner Res 38(6):887–895
Bar A, Wolf L, Bergman Amitai O, Toledano E, Elnekave E (2107) Compression fractures detection on ct. Soc Photo-Opt Instrumentation Eng (SPIE) Conf Seri 10134: 1013440. https://doi.org/10.1117/12.2249635
Simonyan K, Zisserman A (2014) Very deep convolutional networks for large-scale image recognition. arXiv preprint arXiv:1409.1556
Hochreiter S, Schmidhuber J (1997) Long short-term memory. Neural Comput 9(8):1735–1780. https://arxiv.org/abs/1409.1556
Tomita N, Cheung YY, Hassanpour S (2018) Deep neural networks for automatic detection of osteoporotic vertebral fractures on ct scans. Comput Biol Med 98:8–15
He K, Zhang X, Ren S, Sun J (2015) Deep residual learning for image recognition. Preprint at https://arxiv.org/abs/1512.03385
Redmon J, Farhadi A (2016) YOLO9000: better, faster, stronger. Preprint at https://arxiv.org/abs/1612.08242
Redmon J, Divvala S, Girshick R, Farhadi A (2015) You only look once: unified, real-time object detection. Preprint at https://arxiv.org/abs/1506.02640
Yabu A, Hoshino M, Tabuchi H, Takahashi S, Masumoto H, Akada M, Morita S, Maeno T, Iwamae M, Inose H, Kato T, Yoshii T, Tsujio T, Terai H, Toyoda H, Suzuki A, Tamai K, Ohyama S, Hori Y, Okawa A, Nakamura H (2021) Using artificial intelligence to diagnose fresh osteoporotic vertebral fractures on magnetic resonance images. Spine J 21(10):1652–1658
Mehta S, Rastegari, M (2021) MobileViT: lightweight, general-purpose, and mobile-friendly vision transformer. Preprint at https://arxiv.org/abs/2110.02178
Xie Q, Luong M-T, Hovy E, Le QV (2019) Self-training with noisy student improves ImageNet classification. Preprint at https://arxiv.org/abs/1911.04252
Wang C-Y, Yeh I-H, Liao H-YL (2021) You only learn one representation: unified network for multiple tasks. Preprint at https://arxiv.org/abs/2105.04206
Wang C-Y, Bochkovskiy A, Liao H-YM (2020) Scaled-YOLOv4: scaling cross stage partial network. Preprint at https://arxiv.org/abs/2011.08036
Wang C-Y, Liao H-YM, Yeh I-H, Wu Y-H, Chen P-Y, Hsieh J-W (2019) CSPNet: a new backbone that can enhance learning capability of CNN. Preprint at https://arxiv.org/abs/1911.11929
Li D, Hu J, Wang C, Li X, She Q, Zhu L, Zhang T, Chen Q (2021) Involution: inverting the inherence of convolution for visual recognition. Preprint at https://arxiv.org/abs/2103.06255
He K, Zhang X, Ren S, Sun J (2015) Spatial pyramid pooling in deep convolutional networks for visual recognition. IEEE Trans Pattern Anal Mach Intell 37(9):1904–1916
Liu S, Qi L, Qin H, Shi J, Jia J (2018) Path aggregation network for instance segmentation. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 8759–8768
Redmon J, Farhadi A (2018) YOLOv3: an incremental improvement. Preprint at https://arxiv.org/abs/1804.02767
Bochkovskiy A, Wang C-Y, Liao H-YM (2020) YOLOv4: optimal speed and accuracy of object detection. Preprint at https://arxiv.org/abs/2004.10934
Howard AG, Zhu M, Chen B, Kalenichenko D, Wang W, Weyand T, Andreetto M, Adam H (2017) Mobilenets: efficient convolutional neural networks for mobile vision applications. Preprint at https://arxiv.org/abs/1704.04861
Sandler M, Howard A, Zhu M, Zhmoginov A, Chen L-C (2018) Mobilenetv2: inverted residuals and linear bottlenecks. Preprint at https://arxiv.org/abs/1801.04381
Howard A, Sandler M, Chu G, Chen L-C, Chen B, Tan M, Wang W, Zhu Y, Pang R, Vasudevan V, Le QV, Adam H (2019) Searching for mobilenetv3. Preprint at https://arxiv.org/abs/1905.02244
Han K, Wang Y, Tian Q, Guo J, Xu C, Xu C (2020) GhostNet: more Features from Cheap Operations. Preprint at https://arxiv.org/abs/1911.11907
Dosovitskiy A, Beyer L, Kolesnikov A, Weissenborn D, Zhai X, Unterthiner T, Dehghani M, Minderer M, Heigold G, Gelly S, Uszkoreit J, Houlsby N (2020) An image is worth 16x16 words: Transformers for image recognition at scale. Preprint at https://arxiv.org/abs/2010.11929
Lee D-H (2013) Pseudo-label: the simple and efficient semi-supervised learning method for deep neural networks. In: Proceedings of the international conference on machine learning, p 896
Cubuk ED, Zoph B, Shlens J, Le QV (2020) Randaugment: practical automated data augmentation with a reduced search space. In: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition workshops, pp 702–703
Tan M, Le QV (2019) Efficientnet: rethinking model scaling for convolutional neural networks. Preprint at https://arxiv.org/abs/1905.11946
Zheng Z, Wang P, Liu W, Li J, Ye R, Ren D (2019) Distance-iou loss: faster and better learning for bounding box regression. Preprint at https://arxiv.org/abs/1911.08287
Tan M, Pang R, Le QV (2019) Efficientdet: scalable and efficient object detection. Preprint at https://arxiv.org/abs/1911.09070
Lin T-Y, Goyal P, Girshick R, He K, Doll'ar P (2017) Focal loss for dense object detection. Preprint at https://arxiv.org/abs/1708.02002
Acknowledgements
This work was partly supported by the Ministry of Science and Technology of Taiwan under grant 112-2221-E-224-033-MY3.
Author information
Authors and Affiliations
Corresponding author
Ethics declarations
Conflict of interests
There is no conflict of interest.
Additional information
Publisher's Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
Springer Nature or its licensor (e.g. a society or other partner) holds exclusive rights to this article under a publishing agreement with the author(s) or other rightsholder(s); author self-archiving of the accepted manuscript version of this article is solely governed by the terms of such publishing agreement and applicable law.
About this article
Cite this article
Hsieh, MH., Chang, CY. & Hsu, SM. Accurate detection of fresh and old vertebral compression fractures on CT images using ensemble YOLOR. Multimed Tools Appl 83, 89375–89391 (2024). https://doi.org/10.1007/s11042-024-20355-z
Received:
Revised:
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s11042-024-20355-z