Carpal bone segmentation using fully convolutional neural network

Background: Bone Age Assessment (BAA) refers to a clinical procedure that aims to identify a discrepancy between biological and chronological age of an individual by assessing the bone age growth. Currently, there are two main methods of executing BAA which are known as Greulich-Pyle and Tanner-Whit...

Full description

Saved in:
Bibliographic Details
Main Authors: Liang, Kim Meng, Khalil, Azira, Ahmad Nizar, Muhamad Hanif, Nisham, Maryam Kamarun, Pingguan Murphy, Belinda, Chai Hum, Yan, Mohamad Salim, Maheza Irna, Khin, Wee Lai
Format: Article
Published: Bentham Science Publishers 2019
Subjects:
Online Access:http://eprints.utm.my/id/eprint/89320/
http://dx.doi.org/10.2174/1573405615666190724101600
Tags: Add Tag
No Tags, Be the first to tag this record!
id my.utm.89320
record_format eprints
spelling my.utm.893202021-02-22T06:04:18Z http://eprints.utm.my/id/eprint/89320/ Carpal bone segmentation using fully convolutional neural network Liang, Kim Meng Khalil, Azira Ahmad Nizar, Muhamad Hanif Nisham, Maryam Kamarun Pingguan Murphy, Belinda Chai Hum, Yan Mohamad Salim, Maheza Irna Khin, Wee Lai QH301 Biology TA Engineering (General). Civil engineering (General) Background: Bone Age Assessment (BAA) refers to a clinical procedure that aims to identify a discrepancy between biological and chronological age of an individual by assessing the bone age growth. Currently, there are two main methods of executing BAA which are known as Greulich-Pyle and Tanner-Whitehouse techniques. Both techniques involve a manual and qualitative assessment of hand and wrist radiographs, resulting in intra and inter-operator variability accuracy and time-consuming. An automatic segmentation can be applied to the radiographs, providing the physician with more accurate delineation of the carpal bone and accurate quantitative analysis. Methods: In this study, we proposed an image feature extraction technique based on image segmentation with the fully convolutional neural network with eight stride pixel (FCN-8). A total of 290 radiographic images including both female and the male subject of age ranging from 0 to 18 were manually segmented and trained using FCN-8. Results and Conclusion: The results exhibit a high training accuracy value of 99.68% and a loss rate of 0.008619 for 50 epochs of training. The experiments compared 58 images against the gold standard ground truth images. The accuracy of our fully automated segmentation technique is 0.78 ± 0.06, 1.56 ±0.30 mm and 98.02% in terms of Dice Coefficient, Hausdorff Distance, and overall qualitative carpal recognition accuracy, respectively. Bentham Science Publishers 2019-07 Article PeerReviewed Liang, Kim Meng and Khalil, Azira and Ahmad Nizar, Muhamad Hanif and Nisham, Maryam Kamarun and Pingguan Murphy, Belinda and Chai Hum, Yan and Mohamad Salim, Maheza Irna and Khin, Wee Lai (2019) Carpal bone segmentation using fully convolutional neural network. Current Medical Imaging, 15 (10). pp. 983-989. ISSN 1573-4056 http://dx.doi.org/10.2174/1573405615666190724101600 DOI:10.2174/1573405615666190724101600
institution Universiti Teknologi Malaysia
building UTM Library
collection Institutional Repository
continent Asia
country Malaysia
content_provider Universiti Teknologi Malaysia
content_source UTM Institutional Repository
url_provider http://eprints.utm.my/
topic QH301 Biology
TA Engineering (General). Civil engineering (General)
spellingShingle QH301 Biology
TA Engineering (General). Civil engineering (General)
Liang, Kim Meng
Khalil, Azira
Ahmad Nizar, Muhamad Hanif
Nisham, Maryam Kamarun
Pingguan Murphy, Belinda
Chai Hum, Yan
Mohamad Salim, Maheza Irna
Khin, Wee Lai
Carpal bone segmentation using fully convolutional neural network
description Background: Bone Age Assessment (BAA) refers to a clinical procedure that aims to identify a discrepancy between biological and chronological age of an individual by assessing the bone age growth. Currently, there are two main methods of executing BAA which are known as Greulich-Pyle and Tanner-Whitehouse techniques. Both techniques involve a manual and qualitative assessment of hand and wrist radiographs, resulting in intra and inter-operator variability accuracy and time-consuming. An automatic segmentation can be applied to the radiographs, providing the physician with more accurate delineation of the carpal bone and accurate quantitative analysis. Methods: In this study, we proposed an image feature extraction technique based on image segmentation with the fully convolutional neural network with eight stride pixel (FCN-8). A total of 290 radiographic images including both female and the male subject of age ranging from 0 to 18 were manually segmented and trained using FCN-8. Results and Conclusion: The results exhibit a high training accuracy value of 99.68% and a loss rate of 0.008619 for 50 epochs of training. The experiments compared 58 images against the gold standard ground truth images. The accuracy of our fully automated segmentation technique is 0.78 ± 0.06, 1.56 ±0.30 mm and 98.02% in terms of Dice Coefficient, Hausdorff Distance, and overall qualitative carpal recognition accuracy, respectively.
format Article
author Liang, Kim Meng
Khalil, Azira
Ahmad Nizar, Muhamad Hanif
Nisham, Maryam Kamarun
Pingguan Murphy, Belinda
Chai Hum, Yan
Mohamad Salim, Maheza Irna
Khin, Wee Lai
author_facet Liang, Kim Meng
Khalil, Azira
Ahmad Nizar, Muhamad Hanif
Nisham, Maryam Kamarun
Pingguan Murphy, Belinda
Chai Hum, Yan
Mohamad Salim, Maheza Irna
Khin, Wee Lai
author_sort Liang, Kim Meng
title Carpal bone segmentation using fully convolutional neural network
title_short Carpal bone segmentation using fully convolutional neural network
title_full Carpal bone segmentation using fully convolutional neural network
title_fullStr Carpal bone segmentation using fully convolutional neural network
title_full_unstemmed Carpal bone segmentation using fully convolutional neural network
title_sort carpal bone segmentation using fully convolutional neural network
publisher Bentham Science Publishers
publishDate 2019
url http://eprints.utm.my/id/eprint/89320/
http://dx.doi.org/10.2174/1573405615666190724101600
_version_ 1692991771521318912
score 13.211869