Path R-CNN for Prostate Cancer Diagnosis and Gleason Grading of Histological Images

PMC 2019 AI 8 Explanations View Original
Original Paper (PDF)

Unable to display PDF. Download it here or view on PMC.

Plain-English Explanations
Pages 1-2
Why Automated Gleason Grading Matters for Prostate Cancer

Prostate cancer is the most common and second deadliest cancer in men in the United States. Once a biopsy is taken, pathologists grade the tissue using the Gleason grading system, which classifies growth patterns on a scale from Gleason 1 (tissue closely resembling normal prostate) to Gleason 5 (poorly differentiated, aggressive tissue). The Gleason score is a cornerstone of risk assessment and treatment planning. However, assigning Gleason grades is done manually, making the process time-consuming and susceptible to significant inter-observer and intra-observer variability.

The G3 vs. G4 distinction: The clinical boundary between Gleason 3 and Gleason 4 is especially critical. Misclassifying a G3 tumor as G4, or vice versa, can meaningfully alter treatment decisions. Studies have shown substantial disagreement among pathologists at this boundary, highlighting the need for a more objective, reproducible method.

Normal vs. cancerous tissue: In healthy prostate tissue, glands are organized structures composed of lumens surrounded by orderly rows of epithelial cells, all held together by fibromuscular stroma. In cancerous tissue, epithelial cells proliferate uncontrollably, disrupting the gland architecture. In high-grade cancer, stroma and lumen are largely replaced by sheets of epithelial cells. A computer-aided diagnosis (CAD) tool that can reliably detect epithelial cells and classify their grade would offer a repeatable, precise complement to pathologist review.

The authors, based at UCLA and Cedars-Sinai Medical Center, propose Path R-CNN, a novel region-based convolutional neural network framework that simultaneously detects epithelial cells and performs Gleason grading on histological whole slide images. Their multi-task approach achieved state-of-the-art performance across both tasks.

TL;DR: Gleason grading is clinically essential but plagued by pathologist variability, especially at the G3/G4 boundary. Path R-CNN is a multi-task deep learning framework that automates both epithelial cell detection (AUC 0.998) and Gleason grading (mIOU 79.56%, pixel accuracy 89.40%) on histological images.
Pages 3-4
Prior Approaches to Automated Gleason Grading

Traditional feature-based methods: Earlier CAD systems for prostate cancer grading relied on handcrafted features combined with classical machine learning classifiers. Stotzka et al. used statistical features from nuclei distributions with a hybrid neural network/Gaussian classifier. Tabesh et al. aggregated color, texture, and morphometric features and compared Gaussian, k-nearest neighbor, and support vector machine classifiers. Gorelick et al. used a two-stage AdaBoost model on sub-images from 50 whole-mount sections across 15 patients. While these approaches achieved reasonable results on their own datasets, they depended heavily on manually designed feature extraction pipelines and required accurate pre-localization of regions of interest, which is itself a non-trivial problem.

Deep learning segmentation models: More recent work leveraged convolutional neural networks that learn features directly from data. U-Net, proposed by Ronneberger et al., introduced a U-shaped encoder-decoder architecture for biomedical segmentation. Li et al. extended this with Multi-scale U-Net, incorporating different input scales without excessive memory overhead. Ing et al. compared FCN-8s, SegNet variants, and Multi-scale U-Net for semantic segmentation of Gleason-graded tissue. Chen et al. proposed DCAN, a multi-task extension to U-Net that won the MICCAI 2015 Gland Segmentation Challenge. Yang et al. built on DCAN with suggestive annotation using active learning to select representative training samples.

The R-CNN lineage: The region-based CNN family began with R-CNN, which trains networks to classify proposed regions of interest. Fast R-CNN improved speed by extracting features on shared feature maps via RoIPool. Faster R-CNN added a Region Proposal Network (RPN) to learn where to look, predicting object bounds and objectness scores at each spatial position. Mask R-CNN extended Faster R-CNN with a third branch for instance segmentation masks and introduced the RoIAlign layer for precise spatial alignment. Path R-CNN builds directly on this Mask R-CNN foundation.

TL;DR: Earlier methods used handcrafted features with classical classifiers (SVM, kNN, AdaBoost) but lacked reproducibility and required manual ROI localization. Deep learning segmentation models (U-Net, DCAN) improved results. Path R-CNN extends the Mask R-CNN object detection framework to histopathology.
Pages 5-7
The Path R-CNN Architecture: Two Heads, One Backbone

Path R-CNN uses a ResNet backbone as its image feature extractor. The backbone produces feature maps that feed into two parallel branches. The left branch follows the standard Mask R-CNN pipeline: a Region Proposal Network (RPN) generates candidate regions of interest, and a Grading Network Head (GNH) then predicts each region's Gleason grade class label, bounding box offset, and binary segmentation mask. The right branch is a novel addition called the Epithelial Network Head (ENH), which outputs a single score indicating whether epithelial cells are present in the image at all.

Multi-task objective function: The total loss combines four components. The GNH contributes three losses: classification loss (L_cls) for Gleason grade accuracy, bounding-box loss (L_box) for spatial localization of epithelial regions, and mask loss (L_mask) for pixel-level segmentation boundaries. The ENH contributes an objectness prediction loss (L_obj), a binary cross-entropy loss that penalizes the model for incorrectly predicting whether epithelial cells exist in a given image. The total loss is L = L_obj + L_cls + L_box + L_mask.

Why the ENH matters: Without the ENH, the system must use a low detection threshold to avoid missing epithelial regions, which causes it to falsely predict cancer areas in images that are entirely stroma. The ENH acts as a gatekeeper, first determining whether any epithelial cells are present before the GNH attempts grading. This simple addition boosts segmentation performance by a large margin.

Transfer learning from MS COCO: Because annotated medical image data is scarce and expensive, the authors initialized the network with weights pre-trained on the MS COCO dataset (over 200,000 images with pixel-level annotations). Lower layers of deep networks learn generic visual features like edges and textures that transfer well across domains, reducing overfitting on the limited medical dataset.

TL;DR: Path R-CNN adds an Epithelial Network Head (ENH) alongside the Mask R-CNN Grading Network Head (GNH) on a shared ResNet backbone. The ENH acts as a gatekeeper to detect whether epithelial cells exist. The model optimizes four losses simultaneously and uses transfer learning from MS COCO to compensate for limited training data.
Pages 5, 7-8
Dataset, Training Strategy, and Post-Processing

Dataset: The study used 513 histological image tiles from 40 patients (20 per set), retrieved from Cedars-Sinai Medical Center pathology archives. Set A (224 images from 20 patients) contained stroma, benign glands, low-grade cancer (Gleason 3), and high-grade cancer (Gleason 4). Set B (289 images from 20 different patients) contained dense high-grade tumors including Gleason 4 (cribriform and non-cribriform) and Gleason 5, plus stroma-only images. All slides were digitized at 20x magnification with 0.5 micron pixel resolution, extracted as 1200x1200 pixel tiles, and hand-annotated by an expert pathologist with consensus cross-evaluation.

Two-stage training: Due to GPU memory constraints, the 1200x1200 tiles were cropped into 16 overlapping patches and downsampled to 512x512 pixels. Training proceeded in two stages. Stage 1 trained the GNH along with the upper layers (stages 4 and 5) of the ResNet-101 backbone, initialized from MS COCO pre-trained weights and optimized using stochastic gradient descent. Stage 2 froze the Stage 1 weights and trained only the ENH, based on the intuition that epithelial cell detection is a simpler task that does not require updating the entire network.

CRF post-processing: After inference, the model's patch-level predictions were stitched back into full 1200x1200 tiles. This stitching process introduced unnatural boundary artifacts at patch edges. The authors applied a fully connected conditional random field (CRF), originally proposed by Krahenbuhl et al. and later incorporated into CNNs by Chen et al. The CRF uses bilateral Gaussian kernels that force nearby pixels with similar color to share the same class label and remove small isolated regions, smoothing out the stitching artifacts.

Data augmentation and normalization: All tiles were stain-normalized to reduce variability from different scanning systems (Leica SCN400F for Set A, Aperio for Set B). Standard augmentation techniques including image flip, mirror, and rotation were applied before feeding tiles into the network.

TL;DR: 513 image tiles from 40 patients across two sets, digitized at 20x (0.5 micron resolution). Two-stage training: Stage 1 trains the GNH with ResNet upper layers, Stage 2 trains only the ENH with frozen weights. CRF post-processing removes stitching artifacts from patch-based inference.
Pages 8-10
Feature Pyramid Networks and Evaluation Metrics

Feature Pyramid Network (FPN): Both the RPN and GNH use a Feature Pyramid Network structure, replacing single-scale feature maps with multi-scale feature pyramids {P2, P3, P4, P5, P6}. The RPN assigns different anchor scales (32x32, 64x64, 128x128, 256x256, 512x512 pixels) to each pyramid level. For the GNH, each region of interest is assigned to a specific pyramid level based on its size, using a formula that maps smaller ROIs to finer-resolution levels and larger ROIs to coarser levels. This multi-scale approach ensures that features are extracted at an appropriate resolution for each region.

Evaluation metrics: The authors used three standard segmentation metrics to enable fair comparison with prior work. Mean Intersection Over Union (mIOU) averages the Jaccard coefficient across all tissue classes, measuring how well predicted segmentation masks overlap with ground truth. Overall Pixel Accuracy (OPA) computes the fraction of correctly classified pixels across the entire image. Standard Mean Accuracy (SMA) averages per-class pixel accuracy, giving equal weight to each tissue category regardless of how much area it occupies.

Cross-validation design: The 513-tile dataset was randomly divided into 5 non-overlapping folds for cross-validation. Instance segmentation outputs from the model were converted to semantic segmentation by selecting the highest-probability class at each pixel location, enabling direct comparison with prior semantic segmentation methods like U-Net.

TL;DR: FPN generates multi-scale feature pyramids with anchor sizes from 32x32 to 512x512 pixels. Evaluation used mIOU, OPA, and SMA metrics across 5-fold cross-validation on 513 tiles. Instance segmentation outputs were converted to semantic segmentation for comparison with baseline models.
Pages 10-11
Performance Results and Model Comparisons

Overall Gleason grading performance: Using 5-fold cross-validation, Path R-CNN achieved a mean IOU of 79.56%, a standard mean accuracy (SMA) of 88.78%, and an overall pixel accuracy (OPA) of 89.40% across four tissue classes (stroma, benign, low-grade, high-grade). The model performed well on stroma, benign, and high-grade classes but achieved a somewhat lower IOU of 79.54% for the low-grade category, which the authors attribute to the large appearance variance of low-grade glands that differ substantially in size and shape.

Epithelial cell detection: The ENH achieved an AUC of 0.9984 (with standard deviation of 1.329e-3) on the receiver operating characteristic curve, translating to an epithelial cell detection accuracy of 99.07%. This confirms that determining whether epithelial cells are present is a relatively straightforward binary classification task that the simple ENH architecture handles robustly.

Comparison with baselines: Path R-CNN outperformed several baseline models including standard U-Net, Multi-scale U-Net, and traditional approaches based on handcrafted features with support vector machine and random forest classifiers. The authors credit the improvement to five key design choices: (1) the two-stage RPN-then-GNH attention mechanism that tells the grading head where to focus, (2) the multi-task framework providing richer training signals (location, shape, aggressiveness), (3) the ENH preventing false-positive cancer predictions in stroma-only images, (4) the large ResNet backbone avoiding the degradation problem, and (5) the GNH's decoupling of segmentation and classification tasks.

ENH ablation: Removing the ENH caused a significant drop in segmentation performance. Without the ENH, the model required a lower detection threshold to maintain high sensitivity, which led to frequent false predictions of epithelial regions in stroma-only images. The ENH boosted mIOU by a large margin, confirming it is a critical component of the system.

TL;DR: Path R-CNN achieved 79.56% mIOU, 88.78% SMA, and 89.40% OPA for Gleason grading, plus 99.07% epithelial detection accuracy (AUC 0.9984). It outperformed U-Net, Multi-scale U-Net, SVM, and random forest baselines. Removing the ENH caused a large performance drop due to false positives in stroma-only images.
Pages 11-12
Limitations of the Study

Non-patient-wise validation: The 5-fold cross-validation was performed at the tile level, not the patient level. Because the authors did not have patient-level information for stratification, tiles from the same patient could appear in both training and testing folds. Since cancer tissue can look similar across spatially adjacent tiles from the same patient, this could introduce a positive bias in the reported performance metrics. The authors argue that relative comparisons between models remain fair because all methods used the exact same train-test splits.

Two-stage training limitation: Training the ENH and GNH in two separate stages, rather than end-to-end, is a practical compromise. The per-image objectness loss (L_obj) from the ENH operates at a different scale than the per-pixel and per-region losses of the GNH. Joint training did not yield substantial improvement over the baseline, likely because the different loss scales caused interference. Careful tuning of the relative loss weights could potentially enable simultaneous end-to-end training in a single stage.

RoIAlign scale information loss: The RoIAlign layer extracts features for each region of interest from a single pyramid level, which means some scale information is lost. In histopathology, the size of glands carries diagnostic meaning within the Gleason system, where different gland sizes correspond to different grades. Incorporating explicit scale information into the GNH could further improve grading accuracy.

Inherent diagnostic ambiguity: When examining the images where the model performed worst, the authors found cases where even expert pathologists might disagree. If the model is treated as another pathologist, some experts would agree with its predictions while others would not. This raises fundamental questions about how to build a "Doctor-AI Ecosystem" and what criteria should determine when a computer system's diagnosis is trustworthy enough to stand alone.

TL;DR: Key limitations include tile-level (not patient-level) cross-validation introducing possible positive bias, two-stage training that prevents true end-to-end optimization, loss of scale information in the RoIAlign layer, and inherent diagnostic ambiguity on difficult cases where pathologists themselves disagree.
Page 12
Conclusion and Future Research Directions

Path R-CNN demonstrated that adapting object detection frameworks from computer vision to digital pathology can achieve strong results on clinically important tasks. The multi-task design, where epithelial cell detection and Gleason grading are performed simultaneously, provides complementary contextual information that boosts both tasks. The CRF post-processing step addresses a practical challenge of patch-based inference by removing stitching artifacts.

Scale-aware architectures: A natural next step is to incorporate explicit gland size information into the grading network. Since the Gleason system inherently considers gland morphology and size, a scale-aware GNH that retains multi-resolution features for each region of interest could improve the model's ability to distinguish between grades, particularly at the G3/G4 boundary where gland architecture matters most.

Patient-level validation: Future work should evaluate the model using patient-wise data splits to produce unbiased performance estimates. Larger, multi-institutional datasets with proper patient stratification would provide a more rigorous assessment of generalization capability and clinical readiness.

Doctor-AI collaboration: The authors raise broader questions about the role of AI in pathology workflows. How should pathologist annotations be used to train AI systems? How do AI predictions influence pathologist decision-making in practice? And what thresholds of accuracy and reliability are needed before an AI system can be trusted to make independent diagnoses? These questions point toward research in human-AI collaborative diagnostics, where the goal is not to replace pathologists but to create a feedback loop that improves both human and machine performance over time.

TL;DR: Future directions include scale-aware architectures that preserve gland size information for better Gleason grading, patient-level validation on larger multi-institutional datasets, and research into Doctor-AI collaboration models where AI complements rather than replaces pathologist expertise.
Citation: Li W, Li J, Sarma KV, et al.. Open Access, 2019. Available at: PMC6497079. DOI: 10.1109/tmi.2018.2875868. License: Open Access.