PhaseFolio
Methodology · JHTV Portfolio Scoring

JHTV Public-Data Portfolio Scoring

This page documents the methodology behind the risk-adjusted scoring of all ~602 publicly-listed JHU Tech Ventures Therapeutic Modalities inventions, representing the published rubric for dataset version jhtv-portfolio@2026-Q2. Each invention is scored on a three-dimensional rubric — clinical relevance, modality fit, and competitive whitespace — with locked weights of 40/30/30 respectively. The top 10 scoring inventions receive asset-specific deep-dive treatment: comparator-anchored stage profiles, evidence-register citations, Monte Carlo distributions, and per-asset thesis narratives. The 592 lower-ranked inventions are rubric-scored only.

Cohort size~602 Therapeutic Modalities inventions
Data sourceJHU Technology Ventures public listing (Algolia extract, 2026-05-06)
Rubric weights40 / 30 / 30 (clinical relevance / modality fit / whitespace — locked)
Top tier10 inventions with comparator-anchored deep-dive (rNPV + Monte Carlo + tornado + evidence)
Mid tier100 inventions (composite score + dimension breakdown)
Tail tier~492 inventions (aggregate summary only)
Dataset versionjhtv-portfolio@2026-Q2
Engine version1.0.0
01

Cohort and provenance

~602 publicly-listed Therapeutic Modalities inventions, Algolia-extracted 2026-05-06.

The cohort comprises approximately 602 inventions listed under the Therapeutic Modalities category on the JHU Technology Ventures public technology publisher at jhu.technologypublisher.com. Data were extracted via the Algolia search API on 2026-05-06, capturing only publicly disclosed listing fields: title, description, technology category, development stage (where listed), and JHU reference number. No PII, no internal IP filings, and no non-public data were persisted. Inventions without a Therapeutic Modalities category tag were excluded from the scored cohort.

The extracted listing data represent the public-facing invention disclosure only — not the full patent text, internal research status, or licensing activity. The development stage field is often absent or listed at preclinical; where not specified, the scoring pipeline defaults to Preclinical as the assumed stage for rNPV envelope construction. This is a conservative default: actual stage may be more advanced, which would reduce time-to-market and raise the rNPV envelope.

02

Rubric weights

Locked at 40/30/30 for the methodology@2026-05-07 rubric.

The composite score for each invention is computed as a weighted sum across three dimensions. Weights are locked at the values below for the methodology@2026-05-07 rubric. Future revisions will be documented in Section 10 with a version bump.

DimensionWeightWhat it captures
Clinical relevance40%Unmet-need severity, market opportunity, treatment gap vs. current standard of care
Modality fit30%Alignment between modality and indication LoA from PoS calibration table
Competitive whitespace30%Inverse of competing CT.gov programs and FDA approvals in the indication-modality space

What is not in the rubric, and why. Earlier methodology versions (through methodology@2026-06) included an “IRA exposure” dimension at 20% — a modality-derived proxy for Medicare Drug Price Negotiation risk under the Inflation Reduction Act. As of methodology@2026-05-07 this dimension was removed because it double-counts: IRA cliff risk is already modeled inside the rNPV engine via the Year 9 / Year 13 MFP discount on terminal-value cash flows, so weighting it again in a scouting-stage rubric counts the same effect twice (once in the composite score, once in the dollar number). The dimension also correlated with the modality-fit dimension already in the rubric, and provided ~zero discrimination across the JHTV cohort (every top-10 asset clustered at 0.85–0.90). The 20% weight was redistributed across the three retained dimensions; see Section 10 for changelog and substrate-guarantee details.

03

Subscore methodology

How each of the three dimensions is computed, normalized, and combined.

Clinical relevance

Assessed using LLM-based enrichment via Claude (Anthropic) with a structured prompt producing a 0–10 numerical score. The scoring prompt is published at backend/scripts/jhtv/clinical_relevance_prompt.md and covers: disease burden (mortality, morbidity, prevalence), current standard-of-care adequacy, regulatory and reimbursement pathway clarity, and patient population size. The LLM output is normalized to [0,1] for weighting.

Modality fit

Computed from the cumulative launch-of-approval (LoA) for the (indication, modality) pair using PhaseFolio’s PoS calibration table (BIO/QLS 2021 base rates with modality-specific adjustments per Tufts NEWDIGS 2023 and Parker 2015). As of methodology@2026-06 the raw LoA is normalized to [0,1] using the actual ceiling of the modality-baseline cumulative-LoA table (CAR-T at ~0.0088) as the divisor — the highest-LoA modality maps to 1.0 and lower-LoA modalities scale linearly down. Pre-@2026-06 the divisor was anchored at 0.15 (the Phase-1-onward LoA, not the preclinical-through-NDA LoA), which compressed all modalities into a 0.03–0.06 band with no discriminative work. See the version notes below.

Competitive whitespace

Scored as a soft-capped count of competing programs: the count of active or recently completed CT.gov studies and FDA-approved drugs in the same indication-modality space, pulled from the PhaseFolio data pipeline (ctgov.studies + ctgov.fda_applications). The raw count is transformed via a soft cap function so that the marginal penalty for the 50th competitor is smaller than for the 5th. Fewer competing programs yield a score closer to 1.0.

Where IRA risk is modeled instead

Inflation Reduction Act terminal-value risk is applied inside the rNPV engine, not in the rubric. The engine discounts post-cliff cash flows starting at Year 9 (small molecules) or Year 13 (biologics, gene therapies, cell therapies) per the modality-specific Maximum Fair Price schedule. This affects the dollar rNPV envelope shown on each top-10 deep-dive sub-page directly. See the rNPV engine and IRA-framework methodology pages for the schedule and the substrate-discipline rationale for keeping rubric and engine concerns separated.

04

Top-10 asset-specific deep-dive treatment

Comparator-anchored stage profiles, evidence-register citations, Monte Carlo, tornado, narrative thesis.

The 10 highest-scoring inventions receive asset-specific engine treatment. Each top-10 deep-dive includes named real-world comparator programs anchoring stage costs and PoS, a per-asset stage profile with each value cited to comparator evidence, an evidence-register table linking every load-bearing assumption to a verifiable public source, a deterministic rNPV envelope (low / base / high) plus a 1,000-trial Monte Carlo distribution, tornado sensitivity ranking the top six drivers of rNPV variance, revenue trajectories of comparable launched programs where applicable, and a thesis narrative plus key-risks list specific to that asset. The deep-dive payload is exposed at /api/v1/case-studies/jhtv-2026/deep-dive.

Comparator selection criteria, in priority order:

  • Same target / mechanism class — mechanism identity beats unrelated launched comparators for stage calibration. POLθ inhibition is anchored to Artios ART558/ART4215 and Repare RP-3467 even when those programs are still preclinical or Phase 1.
  • Same indication + modality combination with a launched program — AAV gene therapy for ultra-rare neurological disease is anchored to Zolgensma economics regardless of the specific target.
  • Same regulatory pathway (orphan, fast-track, RMAT, accelerated approval) — used when target and indication-modality comparators don’t exist.
  • Same modality with comparable patient population size — last-resort comparator class.

Engine input perturbation for the rNPV envelope. Low case applies costs ×1.25 and peak revenue ×0.75; high case applies costs ×0.75 and peak revenue ×1.25; base case uses the comparator-anchored values directly. WACC is 12% base case unless an asset-specific WACC is justified in the deep-dive narrative.

05

Funding-path archetype classification

Honest framing for assets where standalone rNPV doesn’t apply.

Each top-10 deep-dive carries a funding-path archetype validated against the engine output:

  • VC-fundable — base rNPV is positive at industry-median WACC and standalone-equity assumptions. Standard biotech VC structures apply.
  • Partnership candidate — base rNPV is negative or near-zero under standalone-equity assumptions but the asset is a real strategic prospect under co-development or pharma-deal economics. The narrative explains what deal structure or platform context would clear the asset.
  • Grant / non-commercial — non-traditional revenue model (Gates, GAVI, NIH, product-development partnership). rNPV doesn’t apply meaningfully; valuation is a different exercise (cost-effectiveness, public-health impact).

The archetype is shown alongside the rNPV envelope on every deep-dive sub-page so that a negative envelope can be honestly classified as “needs partnership” or “needs grant funding” rather than “not investable.” This is intentional: a sophisticated reader should be able to distinguish a structurally non-VC-fundable asset (e.g., a malaria vaccine) from a VC-fundable one with execution risk.

06

Multiplier handling — non-double-counting

Why eligible PoS multipliers are tracked but not stacked for this case study.

The PhaseFolio methodology defines seven evidence-based PoS multipliers (genetic validation 2.6×, biomarker 1.7×, orphan 1.4×, CAR-T 1.73×, gene therapy 1.41×, fast-track / RMAT 1.3×, pediatric voucher 1.2×). For the JHTV deep-dive, the comparator research pass already calibrates per-stage PoS values against asset-specific comparator programs that themselves carry the relevant evidence (orphan-validated genetic medicines, biomarker-stratified oncology, etc.). Re-applying the formal multiplier stack on top of the comparator-calibrated PoS would double-count that evidence and produce implausibly high cumulative LoA for preclinical assets.

Decision: the engine consumes Day-1 comparator-calibrated PoS values as final. The list of multipliers eligible for each asset is recorded in the deep-dive payload (multipliers_eligible) and surfaced on the sub-page for audit, but the values are not re-applied via log-odds stacking. This is a deliberate methodology choice for the JHTV case-study build; the formal multiplier stack remains active in the standard PhaseFolio rNPV engine for scenarios where a user provides raw BIO/QLS baseline PoS as input.

Empirical sanity check. Post-deep-dive cumulative LoA across the 10 assets ranges from 6.3% to 10.7%, consistent with published cumulative success rates for orphan-validated genetic medicines (~10–15%) and biomarker-stratified oncology (~5–10%). Pre-fix values (with multipliers stacked) ranged from 59% to 73%, which is implausible for any preclinical asset and would have been correctly challenged by any academic-rigor reviewer.

07

Rubric versus rNPV — what each measures

Composite score and rNPV envelope are correlated, not equivalent.

  • The composite score ranks each invention by scientific and strategic merit — clinical unmet need, modality-indication fit, and competitive whitespace. It does not depend on market size, cost structure, or commercial-policy risk (the latter is captured in the engine, not the rubric).
  • The rNPV envelope measures expected financial value under standard pharma economics — comparator-anchored development costs, indication-specific peak revenue, cumulative PoS, WACC discounting, and IRA terminal-value adjustment per the modality cliff schedule.

An invention can rubric-rank high but rNPV-rank low (e.g., an ultra-rare-disease platform with strong unmet need but small market) and vice versa. Within the top 10 under methodology@2026-05-07, three show positive base rNPV; six are tagged VC-fundable (the gap reflects three assets where the high-input scenario is positive but the base case is not); the remaining four are honestly classified as partnership candidates. The 592 lower-ranked inventions are rubric-scored only — full-portfolio engine treatment is a separate engagement, not part of the public artifact.

Cohort selection vs display order. Top-10 membership is locked to the highest composite-score inventions — the rubric, not rNPV, decides who enters the deep-dive cohort. Display order on the index page is then sorted by base rNPV (descending) for investor readability, so the asset with the strongest standalone economics appears first. Composite-score rank is preserved on every deep-dive sub-page header (a footer line reads “composite-score rank #X of 10 top-tier inventions · page header uses rNPV rank #Y”), so a reader who wants to follow the rubric’s strict ordering can do so. Under methodology@2026-05-07, Personalized Allogeneic Neoantigen CAR-T (objectID 24614) is rank #1 by composite at 0.770 but rank #5 by rNPV (−$16M) — its CAR-T modality_pos = 1.00 lifts the composite while platform-breadth assumptions hold the standalone-equity rNPV down. The override condition for dropping an asset from the top 10 is narrow: high-case rNPV below −$50M and no fitting funding-path archetype. In the 2026-Q2 build this condition triggered zero times — every asset has a defensible archetype framing even when the rNPV envelope is fully negative.

08

Limitations

What the public-data scoring can and cannot tell you.

  • Public-data-only scope. Each listing is described in approximately 350 characters of public disclosure text. No internal IP filing details, preclinical data packages, or licensing activity are available. Clinical relevance scores based on this text are coarser than scores derived from a full data room.
  • Comparator-anchored cost calibration. Top-10 stage profiles are anchored to named real-world comparator programs, but those comparators may differ from the JHU asset on details not captured in public listings (specific patient population definitions, manufacturing scale-up risk, IP overlap). Treat each stage profile as the closest defensible reference, not an asset-specific forecast.
  • Top-10 only. Asset-specific deep-dive treatment is applied to the 10 highest-composite-score inventions only. The remaining 592 are rubric-scored without engine output. Full-portfolio engine treatment is a separate engagement and out of scope for this public artifact.
  • Preclinical-stage default. When the development stage is not specified in the public listing, the pipeline defaults to Preclinical. If an invention is actually in Phase 1 or later, the envelope understates the rNPV by embedding additional preclinical cost and time that has already been incurred.
  • LLM classification noise. Clinical relevance subscores are generated by an LLM and subject to classification noise — particularly for novel modalities or indications with limited training data representation. The structured prompt reduces but does not eliminate variability across runs.
  • IRA risk lives in the engine, not the rubric. The composite score deliberately does not weight Inflation Reduction Act exposure (see Section 2). The IRA effect is captured in the rNPV envelope through a Year 9 / Year 13 MFP discount on terminal-value cash flows. The cliff schedule is binary by modality class and does not model partial-lifecycle exposure, indication-specific negotiation probability, or potential exemptions (orphan, pediatric). Treat the engine’s IRA adjustment as a proxy, not a legal or regulatory determination.
09

Bucketing

Top 10 / Mid 100 / Tail ~492 — three tiers, three levels of detail.

Inventions are sorted by composite score and assigned to one of three tiers:

TierCountTreatment
Top 1010Asset-specific deep-dive: comparators, stage profile, rNPV envelope, Monte Carlo, tornado, evidence register, thesis. Funding-path archetype on every asset.
Mid 100100Secondary table: composite score, top two dimensions, indication summary. No rNPV envelopes.
Tail~492Aggregate only: count by modality category, score distribution quartiles. No individual detail.

Total cohort: ~602 inventions. All counts are approximate — the Algolia extract may include or exclude listings on the boundary of the Therapeutic Modalities category depending on tagging at time of extraction.

§

References

01JHU Technology Ventures — Technology Publisher Johns Hopkins University, 2026

02Clinical Development Success Rates 2011–2020 BIO, QLS Advisors, Informa Pharma Intelligence, 2021

03Tufts Impact Reports Tufts CSDD

04Inflation Reduction Act of 2022 — Drug Pricing Provisions U.S. Congress, 2022

05Probability of Success Calibration — PhaseFolio Methodology

Section 10

Methodology version notes

methodology@2026-05-07 · 2026-05-07

IRA exposure was removed from the rubric. Pre-bump, IRA contributed 20% of the composite via a modality-derived risk score; post-bump, it contributes 0% in the rubric. The rationale is anti-double-counting: IRA cliff risk is already modeled inside the rNPV engine via the Year 9 / Year 13 MFP discount on terminal-value cash flows (per the IRA-framework methodology page), so weighting it again in a scouting-stage rubric counts the same effect twice — once in the composite score, once in the dollar number. The dimension also correlated with modality fit (which has its own 30% weight) and provided ~zero discrimination in the JHTV cohort — every top-10 asset clustered at 0.85–0.90.

The 20% weight was redistributed across the three retained dimensions: Clinical relevance 30% → 40%, Modality fit 25% → 30%, Competitive whitespace 25% → 30%. RUBRIC_VERSION bumps from rubric@2026-Q2-v2 to rubric@2026-Q2-v3; the jhtv-portfolio@2026-Q2 dataset version is unchanged. The top 10 reranks under the new weights — three assets exit (POLθ, mRNA Intensifier, malaria VIP) and three enter (Velocity Receptor CAR-T, retinal Muller-glia reprogramming, polymer-NA CFTR base-editing) — but the new top 10 stays within the previously-researched fixture inventory, so every deep-dive page renders as before with its full thesis prose, comparators, evidence register, Monte Carlo, and tornado.

Format note (this entry only). The version label was promoted from monthly granularity (@YYYY-MM) to daily granularity (@YYYY-MM-DD) for this release per the changelog format header in docs/methodology-changelog.md (“promote to daily only if multiple ships in one month introduce confusion”). Three methodology ships in May 2026 — @2026-05 additive, @2026-06, and this one — had reached that bar. Going forward, all bumps use the daily form. Old monthly-form labels (@2026-04, @2026-05, @2026-06) remain valid forever as the strings stamped on already-issued signed exports; the public /verify endpoint accepts both forms indefinitely. Substrate guarantee preserved per docs/versioning.md.

methodology@2026-06 · 2026-05-06

The modality fit dimension’s normalization denominator was corrected from 0.15 (Phase-1-onward LoA) to the actual ceiling of the modality-baseline cumulative-LoA table (~0.0088, set by CAR-T per Tufts NEWDIGS 2023 1.73× per-stage multipliers on Phases I/II). Under the prior denominator the dimension produced 0.03–0.06 across all modalities, providing no discriminative work; under the corrected denominator modalities span [0.27, 1.0].

Composite scores under methodology@2026-05 remain valid as historical artifacts and continue to verify via the public /verify endpoint — substrate guarantee preserved per docs/versioning.md. The jhtv-portfolio@2026-Q2 dataset version is unchanged. RUBRIC_VERSION bumps from rubric@2026-Q2-v1 to rubric@2026-Q2-v2.

The deep-dive sub-pages additionally fixed a Monte Carlo “success tail” display bug introduced in commit c9349871 where bin-midpoint counting mislabeled failure-cluster trials as “success” on assets where the histogram bin straddling $0 had a positive midpoint. Display-layer only; no engine math change.

methodology@2026-05 · initial release

Public-data scoring of all ~602 JHU Tech Ventures Therapeutic Modalities inventions, four-dimensional rubric with locked weights 30/25/25/20, top-10 deep-dive treatment with comparator-anchored stage profiles, evidence-register citations, Monte Carlo distributions, and per-asset thesis narratives. The 592 lower-ranked inventions are rubric-scored only; full-portfolio engine treatment is a separate engagement.

Engine 1.0.0 · methodology@2026-05-07 · jhtv-portfolio@2026-Q2 · Built 2026-05-07