- Bukowy, John;
- Foss, Halle;
- McGarry, Sean;
- Lowman, Allison;
- Hurrell, Sarah;
- Iczkowski, Kenneth;
- Banerjee, Anjishnu;
- Bobholz, Samuel;
- Barrington, Alexander;
- Dayton, Alex;
- Unteriner, Jackson;
- Jacobsohn, Kenneth;
- See, William;
- Nevalainen, Marja;
- Nencka, Andrew;
- Ethridge, Tyler;
- Jarrard, David;
- LaViolette, Peter
Purpose: Prostate cancer primarily arises from the glandular epithelium. Histomophometric techniques have been used to assess the glandular epithelium in automated detection and classification pipelines; however, they are often rigid in their implementation, and their performance suffers on large datasets where variation in staining, imaging, and preparation is difficult to control. The purpose of this study is to quantify performance of a pixelwise segmentation algorithm that was trained using different combinations of weak and strong stroma, epithelium, and lumen labels in a prostate histology dataset. Approach: We have combined weakly labeled datasets generated using simple morphometric techniques and high-quality labeled datasets from human observers in prostate biopsy cores to train a convolutional neural network for use in whole mount prostate labeling pipelines. With trained networks, we characterize pixelwise segmentation of stromal, epithelium, and lumen (SEL) regions on both biopsy core and whole-mount H&E-stained tissue. Results: We provide evidence that by simply training a deep learning algorithm on weakly labeled data generated from rigid morphometric methods, we can improve the robustness of classification over the morphometric methods used to train the classifier. Conclusions: We show that not only does our approach of combining weak and strong labels for training the CNN improve qualitative SEL labeling within tissue but also the deep learning generated labels are superior for cancer classification in a higher-order algorithm over the morphometrically derived labels it was trained on.