Which data subset should be augmented for deep learning? a simulation study using urothelial cell carcinoma histopathology images.

Yusra A Ameen, Dalia M Badary, Ahmad Elbadry I Abonnoor, Khaled F Hussain, Adel A Sewisy
Author Information
  1. Yusra A Ameen: Department of Computer Science, Faculty of Computers and Information, Assiut University, Asyut, Egypt. yusra.amin@aun.edu.eg. ORCID
  2. Dalia M Badary: Department of Pathology, Faculty of Medicine, Assiut University, Asyut, Egypt.
  3. Ahmad Elbadry I Abonnoor: Urology and Nephrology Hospital, Faculty of Medicine, Assiut University, Asyut, Egypt.
  4. Khaled F Hussain: Department of Computer Science, Faculty of Computers and Information, Assiut University, Asyut, Egypt.
  5. Adel A Sewisy: Department of Computer Science, Faculty of Computers and Information, Assiut University, Asyut, Egypt.

Abstract

BACKGROUND: Applying deep learning to digital histopathology is hindered by the scarcity of manually annotated datasets. While data augmentation can ameliorate this obstacle, its methods are far from standardized. Our aim was to systematically explore the effects of skipping data augmentation; applying data augmentation to different subsets of the whole dataset (training set, validation set, test set, two of them, or all of them); and applying data augmentation at different time points (before, during, or after dividing the dataset into three subsets). Different combinations of the above possibilities resulted in 11 ways to apply augmentation. The literature contains no such comprehensive systematic comparison of these augmentation ways.
RESULTS: Non-overlapping photographs of all tissues on 90 hematoxylin-and-eosin-stained urinary bladder slides were obtained. Then, they were manually classified as either inflammation (5948 images), urothelial cell carcinoma (5811 images), or invalid (3132 images; excluded). If done, augmentation was eight-fold by flipping and rotation. Four convolutional neural networks (Inception-v3, ResNet-101, GoogLeNet, and SqueezeNet), pre-trained on the ImageNet dataset, were fine-tuned to binary classify images of our dataset. This task was the benchmark for our experiments. Model testing performance was evaluated using accuracy, sensitivity, specificity, and area under the receiver operating characteristic curve. Model validation accuracy was also estimated. The best testing performance was achieved when augmentation was done to the remaining data after test-set separation, but before division into training and validation sets. This leaked information between the training and the validation sets, as evidenced by the optimistic validation accuracy. However, this leakage did not cause the validation set to malfunction. Augmentation before test-set separation led to optimistic results. Test-set augmentation yielded more accurate evaluation metrics with less uncertainty. Inception-v3 had the best overall testing performance.
CONCLUSIONS: In digital histopathology, augmentation should include both the test set (after its allocation), and the remaining combined training/validation set (before being split into separate training and validation sets). Future research should try to generalize our results.

Keywords

References

  1. Cancers (Basel). 2020 Oct 12;12(10): [PMID: 33053723]
  2. Nature. 2015 May 28;521(7553):436-44 [PMID: 26017442]
  3. Cancer Commun (Lond). 2020 Apr;40(4):154-166 [PMID: 32277744]
  4. Sci Rep. 2019 Mar 26;9(1):5174 [PMID: 30914794]
  5. Indian Pediatr. 2011 Apr;48(4):277-87 [PMID: 21532099]
  6. Eur Urol Focus. 2022 Jan;8(1):165-172 [PMID: 33358370]
  7. Sci Rep. 2020 Apr 14;10(1):6423 [PMID: 32286358]
  8. Z Med Phys. 2019 May;29(2):86-101 [PMID: 30686613]
  9. Cell Rep. 2018 Apr 3;23(1):181-193.e7 [PMID: 29617659]
  10. Virchows Arch. 2019 Aug;475(2):131-138 [PMID: 31222375]
  11. Stat Med. 2000 Feb 29;19(4):541-61 [PMID: 10694735]
  12. J Digit Imaging. 2021 Aug;34(4):862-876 [PMID: 34254200]
  13. Comput Med Imaging Graph. 2018 Nov;69:125-133 [PMID: 30243216]
  14. Nat Commun. 2020 Dec 11;11(1):6367 [PMID: 33311458]
  15. Int J Comput Assist Radiol Surg. 2019 Mar;14(3):483-492 [PMID: 30649670]
  16. NPJ Digit Med. 2021 Oct 7;4(1):145 [PMID: 34620993]
  17. J Pathol Inform. 2022 May 21;13:100105 [PMID: 36268064]
  18. BMC Med Inform Decis Mak. 2020 Jul 17;20(1):162 [PMID: 32680493]
  19. Med Image Anal. 2019 Dec;58:101544 [PMID: 31466046]
  20. J Pathol Inform. 2021 Nov 03;12:43 [PMID: 34881098]
  21. EBioMedicine. 2018 Jan;27:317-328 [PMID: 29292031]
  22. Front Med (Lausanne). 2019 Oct 01;6:185 [PMID: 31632973]
  23. Am J Pathol. 2020 Jul;190(7):1483-1490 [PMID: 32283104]
  24. Forensic Sci Int. 2019 Sep;302:109922 [PMID: 31442682]
  25. Front Med. 2020 Aug;14(4):470-487 [PMID: 32728875]
  26. Arch Pathol Lab Med. 2021 Oct 1;145(10):1228-1254 [PMID: 33493264]
  27. Br J Cancer. 2021 Feb;124(4):686-696 [PMID: 33204028]
  28. Comput Biol Med. 2020 Jul;122:103861 [PMID: 32658738]
  29. Front Genet. 2022 Feb 16;12:806386 [PMID: 35251119]
  30. Ann Transl Med. 2019 Nov;7(22):618 [PMID: 31930019]
  31. Korean J Radiol. 2020 Jan;21(1):33-41 [PMID: 31920027]
  32. Med Image Anal. 2021 Jan;67:101813 [PMID: 33049577]
  33. Eur Urol. 2020 Aug;78(2):256-264 [PMID: 32354610]
  34. J Big Data. 2021;8(1):101 [PMID: 34306963]
  35. Front Med (Lausanne). 2019 Oct 15;6:222 [PMID: 31681779]
  36. Comput Biol Med. 2021 Jan;128:104129 [PMID: 33254082]
  37. JCO Clin Cancer Inform. 2020 Sep;4:799-810 [PMID: 32926637]
  38. Comput Methods Programs Biomed. 2021 Oct;210:106375 [PMID: 34500139]
  39. Cancer Med. 2021 Jul;10(14):4805-4813 [PMID: 34114376]
  40. Am J Pathol. 2021 Oct;191(10):1717-1723 [PMID: 33838127]
  41. Med Image Anal. 2017 Dec;42:60-88 [PMID: 28778026]
  42. JCO Clin Cancer Inform. 2020 Apr;4:367-382 [PMID: 32330067]
  43. Diagn Pathol. 2020 Jul 16;15(1):87 [PMID: 32677978]
  44. Urologia. 2021 May;88(2):94-102 [PMID: 33402061]
  45. Comput Struct Biotechnol J. 2018 Feb 09;16:34-42 [PMID: 30275936]

MeSH Term

Humans
Deep Learning
Urinary Bladder Neoplasms
Computer Simulation
Neural Networks, Computer
Carcinoma

Word Cloud

Created with Highcharts 10.0.0augmentationvalidationdatasetimagesdatasettraininghistopathologycellcarcinomatestingperformanceaccuracysetsdeeplearningdigitalmanuallyapplyingdifferentsubsetstestwaysurothelialdoneneuralInception-v3Modelusingbestremainingtest-setseparationoptimisticresultsBACKGROUND:ApplyinghinderedscarcityannotateddatasetscanameliorateobstaclemethodsfarstandardizedaimsystematicallyexploreeffectsskippingwholetwotimepointsdividingthreeDifferentcombinationspossibilitiesresulted11applyliteraturecontainscomprehensivesystematiccomparisonRESULTS:Non-overlappingphotographstissues90hematoxylin-and-eosin-stainedurinarybladderslidesobtainedclassifiedeitherinflammation59485811invalid3132excludedeight-foldflippingrotationFourconvolutionalnetworksResNet-101GoogLeNetSqueezeNetpre-trainedImageNetfine-tunedbinaryclassifytaskbenchmarkexperimentsevaluatedsensitivityspecificityareareceiveroperatingcharacteristiccurvealsoestimatedachieveddivisionleakedinformationevidencedHoweverleakagecausemalfunctionAugmentationledTest-setyieldedaccurateevaluationmetricslessuncertaintyoverallCONCLUSIONS:includeallocationcombinedtraining/validationsplitseparateFutureresearchtrygeneralizesubsetaugmentedlearning?simulationstudyConvolutionalnetworkDataDeepHistopathologyUrothelial

Similar Articles

Cited By

No available data.