Back

Novel Autosegmentation Spatial Similarity Metrics Capture the Time Required to Correct Segmentations Better than Traditional Metrics in a Thoracic Cavity Segmentation Workflow

Kiser, K.; Barman, A.; Stieb, S.; Fuller, C. D.; Giancardo, L.

2020-05-18 radiology and imaging
10.1101/2020.05.14.20102103
Show abstract

IntroductionAutomated segmentation templates can save clinicians time compared to de novo segmentation but may still take substantial time to review and correct. It has not been thoroughly investigated which automated segmentation-corrected segmentation similarity metrics best predict clinician correction time. Materials and MethodsBilateral thoracic cavity volumes in 329 CT scans were segmented by a UNet-inspired deep learning segmentation tool and subsequently corrected by a fourth-year medical student. Eight spatial similarity metrics were calculated between the automated and corrected segmentations and associated with correction times using Spearmans rank correlation coefficients. Nine clinical variables were also associated with metrics and correction times using Spearmans rank correlation coefficients or Mann-Whitney U tests. ResultsThe added path length, false negative path length, and surface Dice similarity coefficient correlated better with correction time than traditional metrics, including the popular volumetric Dice similarity coefficient (respectively {rho} = 0.69, {rho} = 0.65, {rho} = -0.48 versus {rho} = -0.25; correlation p values < 0.001). Clinical variables poorly represented in the autosegmentation tools training data were often associated with decreased accuracy but not necessarily with prolonged correction time. DiscussionMetrics used to develop and evaluate autosegmentation tools should correlate with clinical time saved. To our knowledge, this is only the second investigation of which metrics correlate with time saved. Validation of our findings is indicated in other anatomic sites and clinical workflows. ConclusionNovel spatial similarity metrics may be preferable to traditional metrics for developing and evaluating autosegmentation tools that are intended to save clinicians time.

Matching journals

The top 6 journals account for 50% of the predicted probability mass.

1
European Radiology
based on 11 papers
Top 0.1%
18.1%
2
Scientific Reports
based on 701 papers
Top 7%
13.4%
3
PLOS ONE
based on 1737 papers
Top 65%
5.7%
4
Computers in Biology and Medicine
based on 39 papers
Top 0.8%
5.4%
5
npj Digital Medicine
based on 85 papers
Top 4%
4.8%
6
Diagnostics
based on 36 papers
Top 0.9%
3.1%
50% of probability mass above
7
PLOS Digital Health
based on 88 papers
Top 4%
3.0%
8
Journal of the American Medical Informatics Association
based on 53 papers
Top 3%
3.0%
9
Annals of Translational Medicine
based on 14 papers
Top 1%
2.5%
10
Frontiers in Oncology
based on 34 papers
Top 4%
2.4%
11
Radiotherapy and Oncology
based on 11 papers
Top 1%
1.9%
12
Journal of Clinical Medicine
based on 77 papers
Top 8%
1.9%
13
Neuro-Oncology Advances
based on 14 papers
Top 1%
1.7%
14
Journal of Magnetic Resonance Imaging
based on 10 papers
Top 1%
1.7%
15
Archives of Clinical and Biomedical Research
based on 18 papers
Top 0.5%
1.7%
16
Cureus
based on 64 papers
Top 10%
1.7%
17
Scientific Data
based on 30 papers
Top 1%
1.7%
18
Heliyon
based on 57 papers
Top 5%
1.7%
19
eBioMedicine
based on 82 papers
Top 3%
1.4%
20
Informatics in Medicine Unlocked
based on 11 papers
Top 1%
1.4%
21
Medicine
based on 29 papers
Top 5%
1.4%
22
Magnetic Resonance in Medicine
based on 11 papers
Top 1%
0.9%
23
The Lancet Digital Health
based on 25 papers
Top 4%
0.9%
24
Computer Methods and Programs in Biomedicine
based on 12 papers
Top 1%
0.7%
25
Cancers
based on 57 papers
Top 7%
0.7%
26
JCO Clinical Cancer Informatics
based on 14 papers
Top 4%
0.7%
27
Stroke: Vascular and Interventional Neurology
based on 12 papers
Top 2%
0.7%