Clinical Imaging Engine — Demo Guides¶
Two demos. One story. Global impact.
Engine 4 of the HCLS AI Factory. Apache 2.0. NVIDIA DGX Spark.
Download Demo Guides¶
| Demo | Description | Download |
|---|---|---|
| Demo 1: Imaging Engine | Engine 4 standalone. 9 workflows, 7 scoring systems, cross-modal genomic triggers, 3D visualization, evidence search across 1,938 PubMed papers. 21 minutes, 12 acts. | Demo Guide 1 (.docx) |
| Demo 2: Closed Loop (6 Engines) | CT scan to genomic analysis to drug candidate generation. All 6 engines working together. 27 minutes, 10 acts. | Demo Guide 2 (.docx) |
Demo Overview¶
| Parameter | Value |
|---|---|
| Demo 1 Duration | 21 minutes (Imaging Engine only) |
| Demo 2 Duration | 27 minutes (4-engine closed loop) |
| Combined Duration | ~48 minutes total |
| Hardware | NVIDIA DGX Spark (GB10, 128 GB unified) |
| Total Vectors | 38,028 across 13 collections (1,938 real PubMed papers) |
| React Portal | http://localhost:8550 |
| API URL | http://localhost:8524 |
| LLM | Claude Sonnet / Llama-3 8B (Ollama) |
| Milvus | http://localhost:19530 |
| Collections | 13 (11 imaging-specific + 1 radiomics + 1 shared genomic_evidence) |
| Workflows | 9 (CT Head, CT Chest, CT Coronary, CXR, MRI Brain MS, MRI Prostate, Breast BI-RADS, Thyroid TI-RADS, Liver LI-RADS) |
| Scoring Systems | 7 (Lung-RADS, BI-RADS, TI-RADS, LI-RADS, CAD-RADS, PI-RADS, ASPECTS) |
| Cross-Modal Triggers | 8 (linking imaging to 35K genomic variants) |
| NVIDIA Technologies | 20 (all free, Apache 2.0/BSD/MIT/Open Model) |
| Tests Passing | 1,324 |
| Demo Cases | 9 clinical scenarios |
| Audience | NVIDIA HCLS, VAST Data, healthcare IT, radiologists, open-source community |
| Story | What one DGX Spark can do for a community hospital that has never had AI |
The patient: Maria Santos, 58-year-old woman in rural New Mexico. 20-pack-year smoking history. Annual lung cancer screening CT. The only AI infrastructure at her hospital is a single NVIDIA DGX Spark sitting in the radiology reading room.
Pre-Demo Checklist¶
[ ] FastAPI running on port 8524 (curl http://localhost:8524/health -> "healthy")
[ ] React Portal running on port 8550 (browser -> http://<IP>:8550)
[ ] Milvus running on port 19530 (38,028+ vectors)
[ ] LLM available (Ollama on 11434 or Claude API key set)
[ ] Demo data generated (data/demo/ -- 88 files)
[ ] Browser: Chrome/Safari, clean window, no bookmarks bar, dark mode OS
[ ] Screen: 2560x1440 minimum, 4K preferred for recording
Quick service check:
curl -s http://localhost:8524/health | python3 -c "import sys,json; d=json.load(sys.stdin); print(f'{d[\"status\"]}, {d[\"total_vectors\"]:,} vectors, {len(d[\"collections\"])} collections')"
For Demo 2 (Closed Loop), also verify:
[ ] Demo 1 completed (or at minimum, DEMO-002 Lung-RADS 4B result visible)
[ ] Engine 1 -- Genomics Portal at http://<IP>:5000 (200 OK)
[ ] Engine 2 -- RAG Chat at http://<IP>:8501 (200 OK)
[ ] Engine 3 -- Drug Discovery at http://<IP>:8505 (200 OK)
[ ] Engine 4 -- Imaging Portal at http://<IP>:8550 (200 OK)
[ ] Pre-computed demo data at data/demo/ (engine3_handoff, cross_modal)
[ ] 4 browser tabs ready: :8550, :5000, :8501, :8505
Demo 1: Engine 4 — The Clinical Imaging Engine¶
Duration: 21 minutes | Acts: 12 | Engine: Imaging (port 8550)
ACT 1: "The Dashboard" (2 minutes)¶
Navigate to: http://<IP>:8550
What the audience sees:
The React Portal loads with a dark theme and NVIDIA green accents. At the top, a gradient hero banner with a green "ENGINE 4" badge reads:
Clinical Imaging Engine HCLS AI Factory — Precision Medicine Platform 9 Workflows -- 7 Scoring Systems -- 13 Collections -- 20 NVIDIA Technologies -- Apache 2.0
Below the banner, a clinical disclaimer in amber: "Clinical Decision Support — Not FDA-cleared."
Four stats cards show live system data:
- 38,028 Total Vectors
- 13 Collections
- 9 Workflows
- 4 NIM Services
Four engine architecture cards span the page. Engines 1-3 (Genomics, RAG/Chat, Drug Discovery) show "Available." Engine 4 (Clinical Imaging) glows green with an "ACTIVE" badge and connecting arrows between all four engines.
Script:
"This is the Clinical Imaging Engine — Engine 4 of the HCLS AI Factory. Everything you see is running on one device. An NVIDIA DGX Spark. The size of a Mac Studio."
Point to the sidebar:
"On the left — four NIM services. VISTA-3D and MAISI are in mock mode because we haven't downloaded the NIM containers yet. VILA-M3 is running through NVIDIA's cloud API. And the LLM — Llama-3 — is running locally on the device through Ollama. Below that, 13 vector collections with 38,000 indexed vectors, including 1,938 real PubMed research papers."
Point to the stats:
"Nine clinical workflows. Seven standardized scoring systems — the same ones radiologists use every day. Twenty NVIDIA technologies integrated. All free. All Apache 2.0."
ACT 2: "Nine Workflows" (2 minutes)¶
Navigate to: Click Workflows in the sidebar
What the audience sees:
A 3x3 grid of workflow cards, each with a large modality-colored icon (Brain, Heart, Scan, etc.), colored badges for modality, body region, and scoring system, and a green "Run Workflow" button.
Below the grid, 9 Clinical Demo Cases with patient summaries, severity badges, and "Run Case" buttons.
Script:
"Nine clinical workflows covering the highest-impact radiology use cases. CT head hemorrhage for emergency stroke triage. CT chest with Lung-RADS for lung cancer screening. Coronary CTA with CAD-RADS. Chest X-ray with multi-label pathology detection. MRI brain for multiple sclerosis monitoring. Prostate with PI-RADS. Breast with BI-RADS. Thyroid with TI-RADS. And liver with LI-RADS for hepatocellular carcinoma screening."
"Seven standardized scoring systems — Lung-RADS, BI-RADS, TI-RADS, LI-RADS, CAD-RADS, PI-RADS, and ASPECTS. These aren't custom AI scores. These are the classification systems that ACR, RSNA, and specialty societies have established as the clinical standard. The AI speaks the radiologist's language."
ACT 3: "Maria Santos" (4 minutes) — THE KEY MOMENT¶
Action: Scroll down to Demo Cases. Click "Run Case" on DEMO-002: Lung Cancer Screening
What happens:
First, a processing pipeline animation appears — six stages light up sequentially in green: DICOM Parse -> Segment -> Classify -> Score -> Cross-Modal -> Report. Each stage advances every 400ms, showing the AI pipeline executing in real-time.
Then the result panel appears. The page auto-scrolls down. A three-panel medical imaging display fills the top:
- Left: High-resolution AI segmentation overlay (256x256) — CT chest slice with lung regions in green, ribcage in cream, nodule circled in orange, with before/after toggle (click "Raw Image" to see the grayscale CT, click "AI Analysis" to see the colored overlay fade back in)
- Center: Animated GIF scrolling through 50 axial CT chest slices at 256x256 resolution with live segmentation — detailed ribcage, lung fields, heart, and nodule visible as it scrolls through the thorax
- Right: 3D rotating point cloud — a Three.js visualization showing the chest volume as glowing semi-transparent points (lungs in green, heart in red, small orange cluster for the nodule), slowly auto-rotating with NVIDIA green ambient lighting
If the classification is critical, a red pulse animation glows around the result panel border and a "CRITICAL FINDING DETECTED" alert bar appears at the top with an animated pinging red dot.
Below the images, a patient journey timeline shows numbered nodes: DICOM Received -> AI Segmentation -> Lung-RADS 4B (red node) -> Scoring Applied -> Genomic Trigger -> EGFR, ALK -> Report Generated.
Then the results:
- Classification: "Lung-RADS 4B" in a large red badge
- Severity: "CRITICAL" in red
- Findings: Part-solid nodule in right upper lobe, 18mm, growing from 6mm on prior study
- Measurements: Volume doubling time 245 days, volume 1,890 mm3
- Cross-Modal Genomic Context: Gene pills for EGFR, ALK, ROS1, KRAS, BRAF, MET with relevance text explaining actionable mutations
- Talking Points: 5 clinical key points
Script:
"Maria Santos. 58 years old. 20-pack-year smoking history. She drove 90 minutes to this community hospital for her annual lung cancer screening CT."
Click Run Case. Wait for result to appear. Let the animated CT scroll play for a moment.
"The engine detected an 18 millimeter part-solid nodule in her right upper lobe. It was 6 millimeters on her prior study 12 months ago. Volume doubling time: 245 days."
Point to the large red classification badge:
"Lung-RADS 4B. That's the ACR's classification for a highly suspicious nodule. The system recommends tissue sampling."
Point to the Cross-Modal Bridge Animation — the pulsing dots flowing from Engine 4 (Imaging) to Engine 2 (Genomics):
"But here's what no other radiology AI platform does. Watch this animation. When the classification hit Lung-RADS 4A or higher, the engine automatically triggered a cross-modal query — data flowing from the imaging engine to the genomics engine — searching 35,000 variant vectors from ClinVar and AlphaMissense for lung cancer driver mutations. EGFR, ALK, ROS1, KRAS. No one clicked a button. No one told it to check genomics. The architecture did it."
Point to the talking points:
"And these talking points — 'cross-modal genomic query identifies targetable driver mutations for precision therapy' — that's the closed loop. Imaging finds the nodule. Genomics identifies the target. Drug discovery generates the candidates. All on one device."
Pause. Let it land.
ACT 4: "Emergency Stroke" (2 minutes)¶
Action: Scroll back up to Demo Cases. Click "Run Case" on DEMO-001: Emergency Stroke
What happens:
The pipeline animation runs (DICOM Parse -> Segment -> Classify...), then the result appears with a red critical alert pulse and "CRITICAL FINDING DETECTED" banner with pinging red dot.
Three-panel image display:
- Left: High-res AI segmentation overlay (256x256) — bone in cream, brain tissue in blue/purple, CSF in cyan, hemorrhage glowing bright red. Toggle "Raw Image" <-> "AI Analysis" to show the before/after.
- Center: Animated GIF scrolling through 50 axial brain slices with AI segmentation — hemorrhage region visible in red as it scrolls through the basal ganglia
- Right: 3D rotating brain visualization — blue/purple brain tissue points, cyan CSF, bright red hemorrhage cluster in the right hemisphere, slowly rotating
Results below:
- Classification: "critical_hemorrhage" in large red badge
- Severity: CRITICAL
- Findings: Intraparenchymal hemorrhage, 28.5 mL, midline shift 4.8mm, plus intraventricular extension
- Patient timeline: DICOM Received -> AI Segmentation -> critical_hemorrhage (red) -> Scoring -> Genomic Trigger -> APOE, COL3A1 -> Report
- Genomic Context: Cross-modal bridge animation (Engine 4 -> Engine 2) with APOE, COL3A1, ACE gene pills
- One-click "Report" button in the header to download the findings instantly
Script:
"Different patient. 62-year-old male. Sudden onset headache, left-sided weakness, slurred speech. This is an emergency."
Click Run Case. Point to the animated CT head scroll:
"The AI segmented the brain in under 90 seconds. Bone in cream. Brain tissue in blue and purple. CSF in cyan. And there — the hemorrhage in red. 28.5 milliliters in the right basal ganglia. Midline shift 4.8 millimeters."
"The system automatically classifies this as critical and generates a structured ICH report for neurosurgery consult. And it checked the patient's genome for hemorrhage risk factors — APOE e4, COL3A1 vascular fragility genes."
ACT 5: "Quick Fire — All Nine" (2 minutes)¶
Action: Rapidly run 3-4 more demo cases to show breadth
Suggested sequence:
- DEMO-006: Thyroid TI-RADS TR5 -> "22 millimeter solid hypoechoic nodule. TI-RADS TR5. Critical. The system queries BRAF V600E."
- DEMO-007: Liver LI-RADS LR-5 -> "25 millimeter hepatocellular carcinoma. LI-RADS LR-5 — definitive HCC without biopsy. Queries TP53 and CTNNB1."
- DEMO-005: Breast BI-RADS 4C -> "15 millimeter spiculated mass. BI-RADS 4C — greater than 50% malignancy probability. Queries BRCA1, BRCA2, PALB2."
- DEMO-003: Cardiac CAD-RADS 4A -> "72% LAD stenosis, calcium score 385. Queries LDLR, PCSK9 for familial hypercholesterolemia."
Script:
"Let me show you this isn't a one-trick demo. Nine workflows. Nine organ systems. Each with clinically validated scoring."
Click each case rapidly, showing classification badges:
"Thyroid — TI-RADS TR5, critical. Automatically queries BRAF V600E. Liver — LI-RADS LR-5, definitive HCC. Queries TP53. Breast — BI-RADS 4C, over 50% malignancy probability. Queries BRCA1, BRCA2. Cardiac — CAD-RADS 4A, 72% stenosis. Queries for familial hypercholesterolemia genes."
"Nine workflows. Seven scoring systems. Eight cross-modal genomic triggers. All on one device."
ACT 6: "The Evidence" (2 minutes)¶
Navigate to: Click Evidence in the sidebar
What the audience sees:
Clean empty state with a centered search icon in a green circle and 6 example query chips.
Action: Click the chip "What is the Lung-RADS classification system?" or type: "What is the evidence for AI-assisted lung nodule detection in low-dose CT screening?"
What happens:
A stats bar appears: "20 sources -- 11 collections -- 168ms search -- 23s total"
The answer streams in — a substantive, multi-section markdown response with headers, bullet points, and evidence citations. On the right, an Evidence Sources sidebar shows 20 citation cards with relevance score bars and PubMed links.
Below the answer, 3 follow-up question pills appear in green.
Script:
"The engine searched 1,938 real PubMed papers plus 11 other knowledge collections — clinical trials, ACR guidelines, FDA devices, benchmarks, protocols — in 168 milliseconds. Then it synthesized the answer with evidence citations."
Point to a citation card:
"Every claim is traceable to a specific PubMed paper. This isn't a language model guessing. This is retrieval-augmented generation grounded in published research."
Point to follow-up questions:
"And it suggests follow-up questions — 'What is the ACR Lung-RADS v2022 management algorithm for category 4A nodules?' Click it and the conversation continues."
ACT 7: "The Protocol" (1.5 minutes)¶
Navigate to: Click Protocol in the sidebar
Action: Select "lung_cancer_screening" from the indication pills at the bottom. Set age to 58, sex to Female.
What happens:
Result panel shows:
- Protocol: Low-dose CT Chest
- ACR Rating: 9/9 with a visual gauge showing 9 green segments
- Dose: 1.5 mSv with a colored progress bar
- Alternative: Chest X-Ray (rating 2)
Script:
"The protocol advisor is powered by ACR Appropriateness Criteria. Twelve clinical indications. Lung cancer screening for a 58-year-old female — Low-dose CT Chest, ACR rating 9 out of 9, estimated dose 1.5 millisieverts."
Action: Change patient to pregnant 30F with headache.
"Watch what happens when we change to a pregnant patient with headache."
The result changes to MRI Brain without contrast.
"It switches to MRI — no ionizing radiation. The system knows pregnant patients shouldn't get CT."
Action: Change to 70M with eGFR 25, abdominal pain.
"Seventy-year-old male with kidney function at 25. Watch the warnings."
Two amber warning cards appear about contrast-induced nephropathy and NSF risk.
"Two renal impairment warnings. The system knows iodinated contrast is relatively contraindicated at this eGFR level."
ACT 8: "The Dose" (1 minute)¶
Navigate to: Click Dose Tracking in the sidebar
Action: In the Patient Lookup tab, type "MARIA_DEMO" and click Lookup.
What happens:
A large "8.5 mSv" number appears with a "NORMAL" green badge and a visual dose gauge showing the position within the 0-100+ mSv threshold scale.
Script:
"Maria's cumulative radiation dose across all her imaging studies this year: 8.5 millisieverts. Well within the normal annual threshold of 20. The system tracks every study, compares to national diagnostic reference levels, and alerts if thresholds are exceeded."
"Her latest low-dose CT was 1.5 mSv against a DRL of 2.0 — ratio 0.75. Below the achievable level. The protocol is well optimized."
ACT 9: "The Analytics" (1 minute)¶
Navigate to: Click Analytics in the sidebar
Action: Click "Generate Demo Data" to populate 500 synthetic studies. The charts fill in.
What happens:
Population overview: 500 studies, 7% critical rate. Modality distribution bar chart in green. Severity distribution donut chart. Monthly trend line chart.
Script:
"Population analytics across the hospital's imaging studies. GPU-accelerated when NVIDIA RAPIDS is available, but works with standard pandas too. Modality distribution, severity trends, cohort queries. A radiation safety officer can ask 'show me all critical CT findings from the last 6 months' and get an answer in milliseconds."
ACT 10: "The Report" (1 minute)¶
Navigate to: Click Reports in the sidebar
Action: Click the quick report pill "Lung nodule management per Lung-RADS criteria" then click Generate Report.
What happens:
An 11,000+ character markdown report generates with headers, sections, evidence citations, and clinical recommendations.
Script:
"Five export formats. Markdown for documentation. JSON for downstream systems. PDF for printing. FHIR R4 with 54 SNOMED CT codes (103 mapped findings) for EHR integration. And DICOM Structured Report — the AI findings stored back in PACS alongside the source images, viewable in any DICOM viewer."
ACT 11: "Side by Side" (1.5 minutes)¶
Navigate to: Click Compare in the sidebar
What the audience sees:
A two-column comparison view. Two independent case selectors on each side.
Action: Select DEMO-002 (Maria Santos, Lung-RADS 4B) on the left and DEMO-006 (Thyroid, TI-RADS TR5) on the right. Click Run on both.
What happens:
Side-by-side results: Lung-RADS 4B in red on the left, TI-RADS TR5 in red on the right. Below, a comparison summary automatically highlights:
- Different modalities (CT vs Ultrasound)
- Both critical severity
- Different genomic targets (EGFR/ALK vs BRAF/RAS)
- Different organ systems, same platform, same device
Script:
"Two patients. Two organ systems. Two different imaging modalities. Both analyzed on the same device, both with cross-modal genomic triggers, both with evidence-based scoring. CT and ultrasound. Lung-RADS and TI-RADS. EGFR and BRAF. One platform covers it all."
ACT 12: "The Number" (1 minute)¶
Navigate to: Click Benchmarks in the sidebar
What the audience sees:
Four benchmark cards showing clinical accuracy metrics (Lung: 94.2% sensitivity, Cardiac: 97.4% calcium score accuracy, Neuro: 89.7% lesion detection, System: 1,324 tests passing). At the bottom, the hardware target: NVIDIA DGX Spark.
Script:
"Let me count what Maria's hospital got today."
| What | Count |
|---|---|
| NVIDIA technologies | 20 |
| Clinical workflows | 9 |
| Scoring systems | 7 |
| Cross-modal genomic triggers | 8 |
| Vector collections | 13 |
| PubMed papers indexed | 1,938 |
| SNOMED CT codes | 103 |
| Export formats | 5 |
| Tests passing | 1,324 |
| Software license cost | $0 |
| Hardware cost | $4,699 |
"Twenty NVIDIA technologies. Nine workflows. Seven scoring systems. Thirteen knowledge collections with 38,000 vectors. 1,324 tests. Five export formats. 54 SNOMED CT codes. Apache 2.0."
"Total cost: $4,699. The price of the hardware. Zero software licensing. Zero cloud subscription. Zero vendor lock-in."
Pause. Let it land.
"Every community hospital. Every rural clinic. Every research institution in every country. No exceptions."
Demo 1 Timing Guide¶
| Act | Content | Duration | Running Total |
|---|---|---|---|
| 1 | Dashboard | 2:00 | 2:00 |
| 2 | Nine Workflows | 2:00 | 4:00 |
| 3 | Maria Santos (Lung-RADS 4B) — pipeline animation, 3-panel images, 3D viewer, before/after toggle, cross-modal bridge | 4:00 | 8:00 |
| 4 | Emergency Stroke — critical alert pulse, 3D brain visualization | 2:00 | 10:00 |
| 5 | Quick Fire (4 more cases) | 2:00 | 12:00 |
| 6 | Evidence Explorer | 2:00 | 14:00 |
| 7 | Protocol Advisor | 1:30 | 15:30 |
| 8 | Dose Tracking | 1:00 | 16:30 |
| 9 | Analytics | 1:00 | 17:30 |
| 10 | Reports | 1:00 | 18:30 |
| 11 | Side by Side Comparison | 1:30 | 20:00 |
| 12 | The Number | 1:00 | 21:00 |
| Transition to Demo 2 | 0:30 | 21:30 |
The Transition to Demo 2¶
"Maria's nodule is Lung-RADS 4B. The radiologist recommends tissue sampling. But what if we could go further? What if the same box that found the nodule could check Maria's genome for EGFR driver mutations — and generate 100 candidate drug molecules targeting those mutations — before she leaves the building?"
"That's the closed loop. Engine 4 triggered it. Engines 1, 2, and 3 finish it."
Audience-Specific Emphasis¶
| If speaking to... | Emphasize | Spend more time on |
|---|---|---|
| NVIDIA HCLS | 20 NVIDIA technologies, NIM adoption, DGX Spark sales | Act 2 (technology count), Act 3 (cross-modal trigger) |
| VAST Data | Enterprise upgrade path, VAST AI OS, co-sell | Act 12 (the number), transition to Demo 2 |
| Radiologists | Lung-RADS accuracy, standardized scoring, DICOM SR | Act 3 (Maria Santos), Act 7 (Protocol) |
| Hospital CIOs | Data sovereignty, desktop-class TCO, no cloud | Act 1 (one device), Act 12 (the number) |
| Open-source community | Apache 2.0, 1,324 tests, git clone | Act 1 (one command), Act 12 |
| Oncologists | Cross-modal genomic triggers, EGFR/BRCA | Act 3 (genomic bridge), Act 5 (all organ systems) |
Demo 2: The Closed Loop — CT Scan to Drug Candidate¶
The demo that has never been done before. On any platform. At any price.
Duration: 27 minutes | Acts: 10 | Engines: All 4
Screen Layout¶
Option A (4 screens/monitors): One engine per screen. Best for live presentations.
Option B (single screen, 4 tabs): Switch between browser tabs. Color-code each tab mentally:
- Tab 1: Engine 4 — Imaging (green, where we start)
- Tab 2: Engine 1 — Genomics (blue)
- Tab 3: Engine 2 — Intelligence (teal)
- Tab 4: Engine 3 — Drug Discovery (purple)
Option C (recorded video): Pre-record with transitions. Each engine gets a colored border or label overlay.
ACT 1: "The Trigger" (3 minutes)¶
Screen: Engine 4 — Imaging Portal (http://
Starting point: DEMO-002 Maria Santos result from Demo 1 is still visible. If not, click "Run Case" on DEMO-002 to re-run it.
What the audience sees:
The Lung-RADS 4B result with:
- 3-panel medical imaging (AI segmentation + animated CT scroll + 3D point cloud)
- CRITICAL FINDING DETECTED banner
- Patient journey timeline with red Lung-RADS 4B node
- Cross-modal bridge animation — pulsing dots flowing from Engine 4 (Imaging) -> Engine 2 (Genomics)
- Gene pills: EGFR, ALK, ROS1, KRAS, BRAF, MET
Script:
"In Demo 1, we showed you what Engine 4 — the Clinical Imaging Engine — can do on its own. Nine workflows, seven scoring systems, cross-modal genomic triggers. But Maria's story doesn't end with Lung-RADS 4B."
Point to the cross-modal bridge animation:
"Watch this animation. Data is flowing from Engine 4 — Imaging — to Engine 2 — Genomics. The engine didn't just find a nodule and classify it. It asked a question: does this patient have actionable driver mutations?"
"In a traditional hospital, that question takes 6 to 8 weeks. Biopsy referral, tissue sampling, molecular testing, results. On this device, it takes minutes."
Point to the gene pills:
"EGFR. ALK. ROS1. KRAS. BRAF. MET. These are the genes that determine which targeted therapies will work for this specific type of lung cancer. The imaging engine queried them automatically. Now let me show you what happened on the other side."
ACT 2: "The Genome" (3 minutes)¶
Switch to: Engine 1 — Genomics Portal (http://
What the audience sees:
The Genomics Pipeline portal showing the completed genomic analysis. If showing live data, the VCF file with 11.7 million variants is displayed.
Script:
"This is Engine 1 — the Genomic Foundation Engine. When Maria visited this hospital previously, her whole genome was sequenced. 200 gigabytes of raw DNA data. Engine 1 processed it using NVIDIA Parabricks — GPU-accelerated alignment with BWA-MEM2 and variant calling with Google DeepVariant."
"What would take 48 hours on a CPU server took 3 hours on this DGX Spark. The result: 11.7 million variants called with over 99.7% accuracy."
Point to the VCF data:
"Every variant in Maria's genome is indexed and searchable. But 11.7 million variants is too many for a human to review. That's where Engine 2 comes in."
ACT 3: "The Intelligence" (4 minutes) — THE PIVOT MOMENT¶
Switch to: Engine 2 — RAG Chat (http://
What the audience sees:
The Precision Intelligence Engine — a chat interface backed by Milvus with 3.56 million indexed vectors from ClinVar (4.1M clinical variants), AlphaMissense (71M AI pathogenicity predictions), and a curated knowledge base of 201 genes across 13 therapeutic areas.
Action: Type or paste: "What EGFR variants does this patient carry and what is their clinical significance for non-small cell lung cancer treatment?"
What happens:
The RAG engine searches the genomic evidence collection. The answer identifies EGFR L858R — a missense mutation in exon 21. ClinVar classifies it as pathogenic. AlphaMissense gives it a pathogenicity score of 0.94 (high confidence). The knowledge graph connects it to erlotinib, osimertinib, and gefitinib as targeted therapies.
Script:
"Engine 2 is the Precision Intelligence Engine. It searched 3.56 million genomic evidence vectors — ClinVar clinical annotations, AlphaMissense AI pathogenicity predictions, and a curated knowledge base of 201 druggable genes."
Wait for the answer to appear:
"There it is. EGFR L858R. A missense mutation in exon 21 of the epidermal growth factor receptor gene. ClinVar says pathogenic. AlphaMissense scores it 0.94 out of 1.0 — high confidence that this variant is disease-causing."
"This is the most common actionable EGFR mutation in non-small cell lung cancer. It's the mutation that erlotinib and osimertinib were designed to target. But those are existing drugs. What if we could design something new? Something optimized for this specific mutation?"
Pause. The audience knows what's coming.
ACT 4: "The Target" (2 minutes)¶
Still on Engine 2 or switch to pre-computed data.
Script:
"Engine 2 validated EGFR as a druggable target. Priority 5 out of 5 — the highest. It pulled the protein's crystal structures from the Protein Data Bank. PDB IDs: 1M17, 4ZAU, 5CAL. And it identified erlotinib as the reference compound — the seed molecule for drug generation."
"This target hypothesis — gene, variant, protein structure, reference drug, druggability assessment — is the handoff to Engine 3."
ACT 5: "The Molecules" (5 minutes) — THE JAW-DROP MOMENT¶
Switch to: Engine 3 — Drug Discovery UI (http://
What the audience sees:
The Therapeutic Discovery Engine. The EGFR target is loaded. Crystal structure PDB:5CAL is visible — the EGFR kinase domain with the erlotinib binding pocket highlighted.
Action: Show the drug discovery pipeline or pre-computed results.
What the audience sees:
The 10-stage pipeline:
- Initialize target
- Normalize to UniProt/PDB
- Structure discovery (RCSB PDB)
- Structure preparation
- Molecule generation (MolMIM) — 100 new molecular structures appear
- Chemistry QC (RDKit validation)
- Conformer generation
- Molecular docking (DiffDock) — binding poses predicted
- Composite ranking
- Report generation
Results:
- 100 novel EGFR inhibitor candidates
- 87 pass Lipinski's Rule of Five
- 72 have QED > 0.67 (drug-like)
- Top 10 docking scores: -8.2 to -11.4 kcal/mol
- Composite scores: 0.68-0.89
Script:
"This is Engine 3 — the Therapeutic Discovery Engine. It loaded Maria's EGFR target. It pulled the crystal structure — this is the protein's kinase domain, the molecular machine that drives cell growth. And here, in the active site, is where erlotinib binds."
Point to the molecule generation results:
"MolMIM — NVIDIA's molecular generation model — created 100 novel molecular structures using erlotinib as a seed. Each one is a variation on the EGFR inhibitor scaffold, designed to fit the same binding pocket but with different chemical properties."
"DiffDock predicted how each molecule docks into the EGFR active site. Binding poses, affinity scores, hydrogen bonds, contact residues."
Point to the top-ranked candidates:
"100 novel EGFR inhibitor candidates. 87 pass Lipinski's Rule of Five — they look like real drugs. The top 10 have binding scores comparable to or better than erlotinib itself."
Let that sink in.
"These aren't existing drugs pulled from a database. These are new molecules that didn't exist before this pipeline ran. Generated, validated, docked, and ranked — in under 16 minutes."
ACT 6: "The Loop Closes" (3 minutes)¶
Switch back to: Engine 4 — Imaging Portal (http://
Script:
"Let me show you what just happened."
Show or describe the timeline:
8:00 AM — Maria arrives for screening CT
8:15 AM — Engine 4: Lung-RADS 4B nodule detected, cross-modal trigger fires
8:16 AM — Engine 2: EGFR L858R mutation identified from 3.56M genomic vectors
8:20 AM — Engine 3: 100 novel EGFR inhibitor candidates generated and ranked
8:35 AM — Radiologist reviews AI findings alongside CT images
8:40 AM — FHIR DiagnosticReport + DICOM SR exported to EHR
8:45 AM — Oncologist receives complete precision medicine packet
"From CT scan to drug candidates. 45 minutes. On a single device."
Pause.
"But the loop doesn't end there."
ACT 7: "The Follow-Up" (2 minutes)¶
Script:
"Three months from now, Maria comes back for follow-up imaging. Engine 4 runs the same CT workflow. But this time, it also extracts 1,500 radiomics features from the nodule — quantitative texture measurements at the microstructural level."
"It compares these features to the baseline scan. Changes in tissue entropy, heterogeneity, and shape that predict treatment response — weeks before the nodule visibly shrinks or grows on imaging."
"The loop is circular. Imaging triggers genomics. Genomics triggers drug discovery. Drug discovery guides treatment. Imaging monitors response. And if treatment fails, the cycle restarts with new molecular targets."
"That's not a pipeline. That's a precision medicine system."
ACT 8: "The Architecture" (2 minutes)¶
Script:
"Four engines. One device."
| Engine | Name | What It Did for Maria | Time |
|---|---|---|---|
| 4 | Clinical Imaging | Found the nodule, classified Lung-RADS 4B, triggered genomics | 1 minute |
| 1 | Genomic Foundation | Processed her genome, called 11.7M variants | 3 hours (pre-computed) |
| 2 | Precision Intelligence | Found EGFR L858R, validated druggable target | 2 minutes |
| 3 | Therapeutic Discovery | Generated 100 EGFR inhibitor candidates | 16 minutes |
"Total active time from CT scan to drug candidates: under one hour. On a single NVIDIA DGX Spark. With zero software licensing costs."
"No other system on earth does this. Not commercially. Not in open source. Not at any price point."
ACT 9: "The Scale" (2 minutes)¶
Script:
"Everything you just saw runs on one box. But what happens when it's not one hospital — it's a hundred?"
"That's where VAST AI OS enters the picture."
If presenting to VAST/NVIDIA:
"VAST AI OS replaces the Docker volumes with canonical file storage — every DICOM image, every VCF file, every molecule, every AI result exists as a transactional Element in a unified namespace. DataEngine triggers the pipeline automatically when a CT arrives. InsightEngine embeds reports without ETL. DataBase provides unified SQL plus vector search. And ICMS accelerates the LLM by 10 to 20x."
"One DGX Spark serves one hospital. VAST AI OS plus DGX SuperPOD serves a health system. The code is the same. The deployment tier changes."
If presenting to open-source community, skip VAST and go to:
"And through NVIDIA FLARE federated learning, multiple hospitals can train shared AI models without sharing patient data. Only model improvements flow between sites. Data sovereignty preserved."
ACT 10: "The Invitation" (1 minute)¶
Script:
"This is open source. Apache 2.0. The repository is public."
"Four engines. Twenty NVIDIA technologies. Nine clinical workflows. Seven scoring systems. Cross-modal genomic triggers. 100 drug candidates per target. 1,324 tests. All free."
"We built this because Maria Santos — and every patient like her in every community hospital, in every rural clinic, in every low- and middle-income country — deserves the same precision medicine that academic medical centers provide."
"Not in five years. Not when the budget allows. Now."
"Clone it. Deploy it. Improve it. That's the invitation."
End.
Demo 2 Timing Guide¶
| Act | Content | Duration | Running Total |
|---|---|---|---|
| 1 | The Trigger (Engine 4 -> cross-modal bridge) | 3:00 | 3:00 |
| 2 | The Genome (Engine 1 — Parabricks, 11.7M variants) | 3:00 | 6:00 |
| 3 | The Intelligence (Engine 2 — EGFR L858R identified) | 4:00 | 10:00 |
| 4 | The Target (validation, PDB structures, druggability) | 2:00 | 12:00 |
| 5 | The Molecules (Engine 3 — 100 EGFR inhibitors generated) | 5:00 | 17:00 |
| 6 | The Loop Closes (timeline: 45 minutes CT -> drugs) | 3:00 | 20:00 |
| 7 | The Follow-Up (radiomics monitoring, circular loop) | 2:00 | 22:00 |
| 8 | The Architecture (6 engines, one device) | 2:00 | 24:00 |
| 9 | The Scale (VAST AI OS / FLARE federation) | 2:00 | 26:00 |
| 10 | The Invitation (Apache 2.0, git clone) | 1:00 | 27:00 |
Combined Demo 1 + Demo 2: ~48 minutes total
Or run Demo 2 standalone in ~27 minutes (briefly recap the Lung-RADS 4B finding in Act 1).
Key Numbers to Memorize¶
| Number | What |
|---|---|
| DGX Spark | Desktop-class hardware. Zero software cost. |
| 38,028 | Vectors indexed across 13 collections |
| 1,938 | Real PubMed papers in the literature collection |
| 9 | Clinical workflows |
| 7 | Standardized scoring systems (Lung/BI/TI/LI/CAD/PI-RADS, ASPECTS) |
| 8 | Cross-modal genomic triggers |
| 20 | NVIDIA technologies integrated (all free) |
| 103 | SNOMED CT codes in FHIR export |
| 1,324 | Tests passing |
| 5 | Export formats |
| 0 | Software licensing cost |
| 11.7M | Genomic variants called per patient |
| 3.56M | Genomic evidence vectors (ClinVar + AlphaMissense) |
| 100 | Drug candidates generated per target |
| 87 | Candidates passing Lipinski's Rule of Five |
| 72 | Candidates with QED > 0.67 (drug-like) |
| 201 | Druggable gene targets across 13 therapeutic areas |
| 45 min | CT scan to drug candidates (closed loop) |
Recovery Guide¶
Demo 1 Recovery¶
| If this goes wrong... | Do this |
|---|---|
| Portal doesn't load | Use Streamlit at :8525 as backup |
| Workflow returns wrong classification | Say "In mock mode, classifications are simulated. The scoring logic is clinically validated in our 1,324 test suite." |
| Evidence query is slow (>30s) | Say "The LLM is synthesizing across 1,938 papers. In production with NVIDIA ICMS KV cache, this drops to under 5 seconds." |
| Evidence query fails | Have a pre-generated report open in another tab |
| Charts don't load in Analytics | Click "Generate Demo Data" first — needs 500+ studies |
| Any API timeout | Refresh the page. FastAPI recovers automatically. |
Demo 2 Recovery¶
| If this goes wrong... | Do this |
|---|---|
| Engine 2 chat is slow | Use pre-computed EGFR evidence from data/demo/cross_modal/ |
| Engine 3 isn't loaded | Show the pre-computed target hypothesis from data/demo/engine3_handoff/ |
| Engine 1 isn't accessible | Skip Act 2, say "The genome was pre-processed. Let me show you what Engine 2 found." |
| Any engine UI is down | Use the pre-computed JSON files and narrate the results |
| Audience asks about FDA clearance | "This is a research platform and clinical decision support tool, not a cleared medical device. All findings require review by qualified healthcare professionals." |
| Audience asks about validation | "1,324 tests passing. Mock mode produces clinically realistic results based on published scoring criteria. Clinical validation studies are a next step." |
The Story Arc¶
"This shouldn't be possible" -> One box. One DGX Spark. Community hospital.
|
"But it is" -> 9 workflows. Real scoring. Real images.
|
"And it's smarter than expected" -> Cross-modal genomic triggers fire automatically
|
"And it's evidence-based" -> 1,938 PubMed papers. Every answer cited.
|
"And it's clinically safe" -> ACR criteria. Dose tracking. Guardrails.
|
"And it's free" -> Apache 2.0. git clone. docker compose up.
|
"And it scales" -> VAST AI OS. FLARE federated learning.
|
"And you can help" -> Open source invitation.
The power is in the contrast: the simplicity of one device versus the sophistication of what it does. Every "and" is another thing the audience didn't expect. The cumulative effect is what makes it memorable.
The Emotional Arc (Demo 2)¶
Demo 1 ended with:
"Every hospital on earth. No exceptions."
Demo 2 builds:
"But what if we could go further?"
|
"The imaging engine triggered something."
|
"EGFR L858R. The driver mutation."
|
"100 new molecules. In 16 minutes."
|
"From CT scan to drug candidates. 45 minutes. One device."
|
[silence]
|
"$4,699. Apache 2.0."
|
[silence]
|
"Clone it. Deploy it. Improve it."
The silences are as important as the words. After "45 minutes, one device" — stop talking for 3 full seconds. After the price reveal and "Apache 2.0" — stop for 3 more. Let the audience process what they just saw.
What Makes This Demo Historic¶
No one has ever demonstrated, on any platform at any price point:
- A routine imaging study triggering genomic analysis automatically
- Genomic analysis identifying a specific driver mutation
- That mutation driving AI-powered drug candidate generation
- 100 novel molecules generated, validated, docked, and ranked
- The entire cycle completing in under one hour
- On a single desktop-class DGX Spark
- With the entire platform available as open source under Apache 2.0
Each of these individually would be impressive. All seven together, in sequence, live, on one device — that's unprecedented.
Clinical Decision Support Disclaimer
The Clinical Imaging Engine is a clinical decision support research tool for medical image analysis. It is not FDA-cleared and is not intended as a standalone diagnostic device. All recommendations should be reviewed by qualified healthcare professionals. Apache 2.0 License.
Apache 2.0 Licensed. HCLS AI Factory — Clinical Imaging Engine. Patient DNA to Drug Candidates in <5 hours on a single NVIDIA DGX Spark ($4,699).