Week 01 — Q1

Clinical Uses of AI-Based Segmentation in Neuro-Oncology

Before you write a single line of code, you need to understand why this technology matters. This week, you’ll learn how AI-powered brain tumor segmentation is being used in real hospitals to help real patients — from surgical planning to tracking whether treatment is working.

Why Brain Tumor Segmentation Matters

Every year, hundreds of thousands of people worldwide are diagnosed with brain tumors. Treating them requires incredibly precise information: where exactly is the tumor? How big is it? Which parts are actively growing? Is the treatment shrinking it?

Doctors answer these questions using MRI scans. But here’s the problem: a single brain MRI produces a 3D volume made of hundreds of individual image slices, across multiple imaging sequences (T1, T1ce, T2, FLAIR). A neuroradiologist must manually trace the tumor boundaries on every relevant slice, identifying different sub-regions — a process called segmentation. This can take 30 minutes to several hours per patient, is highly subjective, and varies significantly between experts.

This is where AI comes in. Deep learning models — particularly architectures like U-Net and nnU-Net — can perform this segmentation automatically in seconds, with accuracy that matches or exceeds human experts.

0.86–0.93
Dice scores achieved by nnU-Net in multi-center clinical validation
30–45%
Time savings with AI-assisted segmentation across multiple studies
0.74–0.85
Typical human inter-rater agreement (Dice) — AI matches or exceeds this
💡
Key insight: The goal of AI segmentation isn’t to replace radiologists. It’s to handle the tedious, time-consuming volumetric measurement work so doctors can focus on interpretation, clinical judgment, and patient care.

Neuro-Oncology: What You Need to Know

Neuro-oncology is the branch of medicine focused on cancers of the brain and nervous system. As someone building segmentation models, you don’t need to become a doctor, but you do need to understand the basics of what you’re looking at and why it matters clinically.

Common Brain Tumor Types

Gliomas (the BraTS focus)

Gliomas are the most common primary brain tumors, arising from glial cells that support neurons. They’re classified by the World Health Organization (WHO) into grades 1–4, with grade 4 (glioblastoma / GBM) being the most aggressive. Glioblastomas have a median survival of roughly 15 months even with treatment. BraTS has historically focused on glioma segmentation, particularly distinguishing the enhancing tumor core, necrotic regions, and surrounding edema.

Brain Metastases

These are cancers that started elsewhere in the body (lung, breast, melanoma, etc.) and spread to the brain. They’re actually the most commonly diagnosed brain tumors in adults. Patients often have multiple lesions of varying sizes, making manual segmentation extremely tedious. This is where AI tools are especially impactful — and it’s the focus of the BraTS-METS 2025 challenge.

Meningiomas

Tumors arising from the meninges (the membranes surrounding the brain). Usually benign and slow-growing, but they can compress brain tissue and require surgery or radiation. BraTS 2025 includes dedicated meningioma tasks (Tasks 2–3).

Pediatric Brain Tumors

Brain tumors are the leading cause of cancer death in children. They differ significantly from adult tumors in location, biology, and imaging appearance. nnU-Net has been validated on pediatric tumors, achieving Dice scores of 0.90 for whole tumor, though performance varies by sub-region. BraTS-PED (Task 6) addresses this population.

The Treatment Pipeline

Understanding where AI fits requires knowing the clinical workflow:

01
Diagnosis & Imaging

Patient presents with symptoms (headaches, seizures, neurological deficits). MRI scans are acquired with multiple sequences (T1, T1ce, T2, FLAIR). A neuroradiologist interprets the images and identifies the tumor.

02
Tumor Characterization & Grading

The tumor’s size, location, and sub-regions are assessed. WHO grading may be estimated from imaging (later confirmed by biopsy). Segmentation provides the volumetric measurements needed for treatment decisions.

03
Surgical Planning

If surgery is indicated, precise 3D models of the tumor and surrounding anatomy are needed for navigation. AI segmentation can produce these models faster and more consistently than manual methods (mean 1,139 seconds vs 2,851 seconds for traditional approaches).

04
Radiation Therapy Planning

Radiation oncologists need to delineate the gross tumor volume (GTV) and clinical target volume (CTV). AI-assisted contouring improves inter-reader agreement (Dice from 0.86 to 0.90) and is especially helpful for less-experienced physicians.

05
Treatment & Monitoring

After surgery, radiation, and/or chemotherapy, patients undergo regular follow-up MRI scans. AI-based volumetric tracking enables objective, reproducible measurement of whether the tumor is responding, stable, or progressing.

How AI Segmentation Is Used Clinically

Preoperative Planning & Intraoperative Navigation

Before brain surgery, neurosurgeons need detailed 3D models showing the tumor relative to critical structures (motor cortex, language areas, blood vessels). AI segmentation can generate these automatically from MRI in under 20 minutes, compared to nearly an hour with traditional mesh-growing algorithms. During surgery itself, these segmentations feed into neuronavigation systems that guide the surgeon in real-time.

Stereotactic Radiosurgery (SRS)

For brain metastases treatment with focused radiation, every lesion must be precisely contoured. A randomized multi-reader study showed that AI assistance improved lesion detection from 82.6% to 91.3% sensitivity, while providing a median 30.8% time savings. The benefit was greatest for less-experienced physicians, helping close the expertise gap.

Treatment Response Assessment (RANO Criteria)

This is one of the most impactful clinical applications. After treatment, doctors need to determine whether the tumor is responding. The RANO criteria (Response Assessment in Neuro-Oncology) define standardized rules for this assessment.

📚
What are the RANO Criteria? RANO provides standardized definitions for treatment response: Complete Response (tumor disappears), Partial Response (significant shrinkage), Stable Disease (no major change), and Progressive Disease (growth). Traditionally based on 2D diameter measurements, the updated RANO 2.0 (2023) now includes volumetric 3D measurements as an option — which is exactly what AI segmentation provides.

A landmark study in The Lancet Oncology by Kickingereder et al. (2019) demonstrated this powerfully. Analyzing 532 patients from 34 institutions in the EORTC-26101 trial, they showed that AI-based volumetric response assessment was significantly more reliable than manual RANO assessment:

87%
Agreement for AI-based volumetric progression (vs 51% for manual RANO)
HR 2.59
Hazard ratio for predicting survival with AI (vs 2.07 for manual RANO)
~10 min
Median processing time per scan in simulated clinical environment

For brain metastases specifically, AI models have achieved 100% accuracy in predicting RANO-BM criteria compared to expert assessment, demonstrating that automated longitudinal tracking is clinically viable.

PACS Integration: Making It Actually Usable

One of the most exciting developments is embedding AI directly into the tools radiologists already use. A system at Yale New Haven Health integrated a deep learning algorithm directly into their Visage PACS diagnostic workstation, achieving segmentation in 4 seconds with radiomic feature extraction in under 6 seconds. The AI runs before the radiologist even opens the study — they just verify the result.

AI Accuracy vs. Human Experts

A natural question: how does AI actually compare to trained neuroradiologists? The answer is nuanced.

A systematic review and meta-analysis reported an overall Dice score of 0.84 (95% CI: 0.82–0.86) for ML-based glioma segmentation. State-of-the-art models on the BraTS 2021 benchmark achieve Dice scores of 0.936 for whole tumor, 0.921 for tumor core, and 0.872 for enhancing tumor — which is within or above the range of human inter-rater agreement (0.74–0.85).

Where AI Excels

Round, well-demarcated tumors: AI correctly includes necrosis and contrast-enhanced regions in 97–100% of cases (vs 73.68% for manual segmentation). Consistency: AI produces identical results every time — no inter-rater variability. Speed: Seconds vs minutes or hours. Multi-lesion counting: For patients with many brain metastases, AI dramatically reduces the tedium and missed lesions.

Where Humans Still Win

Complex, infiltrative tumors: For tumors with irregular borders and diffuse infiltration, manual segmentation outperforms automated methods (66.67% vs 37.50–50% accuracy). Clinical context: Radiologists integrate patient history, prior scans, and clinical knowledge that current AI models don’t have access to. Post-treatment changes: Distinguishing true progression from pseudoprogression (treatment-related changes that mimic tumor growth) remains challenging for AI.

⚠️
Critical nuance: Technical accuracy (high Dice scores) doesn’t always equal clinical utility. A 2025 study on automated glioblastoma response assessment showed moderate agreement with radiologist assessments (F1 = 0.587–0.755) but slightly worse stratification by overall survival compared to human assessment. The gap between “accurate segmentation” and “clinically useful tool” is still being bridged.

FDA-Cleared & Clinical AI Segmentation Tools

Several AI tools have moved beyond research into clinical practice:

VBrain by Vysioneer (FDA-Cleared)

An FDA-cleared deep learning algorithm for automated detection and contouring of metastatic brain tumors in stereotactic radiosurgery. Validated on 100 patients with 435 brain metastases: 89.3% overall sensitivity, reaching 99% for lesions ≥10mm. Mean false positive rate of just 0.72 tumors per case. Performance was consistent across demographic and clinical subgroups.

BraTS Toolkit (Open Source)

An open-source platform that makes BraTS challenge algorithms available for clinical use. Three components: BraTS Preprocessor (data standardization), BraTS Segmentor (algorithm orchestration), and BraTS Fusionator (consensus segmentation from multiple models using majority voting). This is how research algorithms get translated into usable tools.

Yale PACS-Integrated System

A novel system embedding a UNETR deep learning model directly into the Visage 7 diagnostic workstation at Yale New Haven Health. Achieves median Dice of 86% in 4 seconds, with automatic radiomic feature extraction. Represents the future of seamless clinical integration.

Challenges & Barriers to Clinical Adoption

Despite impressive results, widespread clinical adoption faces significant hurdles. Understanding these challenges is essential for anyone entering this field.

The Generalizability Problem

This is the biggest technical challenge. Models trained at one hospital often perform worse at another due to domain shift — differences in MRI scanners, imaging protocols, patient populations, and image quality. Cross-institutional studies show Dice score drops from 0.72–0.76 to 0.59–0.68 when models are tested at new institutions. BraTS dataset performance often overestimates real-world performance because BraTS data is heavily curated and standardized.

Dataset Bias

A 2026 study found that segmentation models trained on demographically homogeneous datasets exhibited the most bias. Models trained exclusively on one demographic group had lower Dice scores and showed biases related to age and other characteristics. The key finding: greater demographic diversity in training data reduced bias even without increasing dataset size.

Regulatory Hurdles

Getting AI tools approved for clinical use requires navigating FDA (US) and CE marking (Europe) pathways. The AI-RAPNO consensus identified key challenges: variability in imaging protocols, scarce annotated datasets for rare tumor types, lack of standardized validation frameworks, and ethical considerations around transparency and patient consent.

The “Black Box” Problem

Deep learning models are difficult to interpret — they produce results but can’t explain their reasoning. This erodes clinical trust, especially in high-stakes medical decisions. Newer architectures are incorporating attention mechanisms and clinical knowledge to improve interpretability, but this remains an active research area.

Integration Barriers

Even when the AI works, deploying it in a hospital setting requires: IT infrastructure for GPU computing, interoperability with existing PACS and EMR systems, data privacy compliance (HIPAA/GDPR), trained staff to maintain the system, and reimbursement models to justify the cost.

The Role of BraTS & MICCAI in Driving Progress

The BraTS Challenge has been instrumental in everything described above. Since 2012, it has provided standardized datasets, evaluation metrics, and a competitive environment that pushed segmentation performance from research curiosity to clinical viability.

Key contributions: dataset growth from 65 cases (2012) to over 1,251 cases (2021) and expanding further; standardized preprocessing protocols; expansion from gliomas to metastases, meningiomas, pediatric tumors, and sub-Saharan African populations; and establishment of the Dice/Hausdorff/NSD evaluation framework used across the field.

The Federated Tumor Segmentation (FeTS) Challenge extended this further by allowing 32 institutions to collaboratively train models without sharing patient data — addressing privacy concerns while improving generalizability.

The upcoming BraTS 2025 Lighthouse Challenge continues to expand across additional clinical needs, including post-treatment scenarios and integration with genomic data.

This Week’s Learning Resources

Start Here (Beginner-Friendly)

Video lectures and talks from the BraTS challenge organizers covering brain metastases segmentation, dataset annotation, and challenge methodology. Great for seeing the people and context behind the data you’ll be working with.
Free, visual encyclopedia of brain tumor types with real MRI examples and expert annotations. Start here to build visual intuition about what gliomas, metastases, and meningiomas look like on imaging.
Patient-oriented overview of brain tumor types, diagnosis, and treatment options. Useful for understanding the clinical context and why the work you’re doing matters to real people.

Key Papers (Read the Abstracts at Minimum)

The landmark study showing AI-based volumetric assessment outperforms manual RANO for predicting overall survival. 532 patients, 34 institutions. This paper is the strongest evidence for why automated segmentation matters clinically.
Lancet Oncol. 2019;20(5):728–740
Compares DeepMedic, nnU-Net, and NVIDIA-net across 12 hospitals. Shows nnU-Net achieves the highest accuracy with performance within human expert range even on messy real-world clinical data.
Sci Rep. 2023;13:5765
The updated RANO 2.0 criteria that now include volumetric measurements. Essential context for understanding how AI segmentation connects to clinical treatment decisions.
J Clin Oncol. 2023;41(18):3406–3418
Clinical validation of an FDA-cleared AI tool for brain metastasis detection and contouring. Shows what “ready for clinical use” actually looks like in terms of performance metrics.
Radiat Oncol. 2023;18:54
International multi-reader study demonstrating how AI improves reproducibility of tumor response assessment. Key evidence that AI makes clinical decisions more consistent across different readers.
Neuro-Oncology. 2023;25(3):533–543

Deep Dives (Advanced Reading)

Comprehensive review of data gaps, translation gaps, and implementation gaps preventing clinical adoption. Read this to understand the full landscape of challenges.
Consensus statement on translating AI for pediatric neuro-oncology. Outlines regulatory, ethical, and technical requirements for moving research to clinical practice.
The FeTS challenge: 32 institutions collaborating on segmentation without sharing data. Shows how federated learning addresses privacy and generalizability simultaneously.
The foundational BraTS paper that established the benchmark, evaluation metrics, and inter-rater variability baselines. Historical context for everything built since.

Tools to Explore

Open-source platform for running BraTS segmentation algorithms. Preprocessor, Segmentor, and Fusionator components. This is how research algorithms get packaged for clinical use.
Download and install 3D Slicer this week. Load some publicly available brain MRI data and practice navigating through slices, identifying structures, and understanding what you’re looking at before you try to teach a model to do it.