<?xml version="1.0" encoding="utf-8"?><feed xmlns="http://www.w3.org/2005/Atom" xml:lang="en"><generator uri="https://jekyllrb.com/" version="4.4.1">Jekyll</generator><link href="https://noahrflynn.com/feed.xml" rel="self" type="application/atom+xml"/><link href="https://noahrflynn.com/" rel="alternate" type="text/html" hreflang="en"/><updated>2026-05-12T07:29:12+00:00</updated><id>https://noahrflynn.com/feed.xml</id><title type="html">Noah Flynn</title><subtitle>Senior Research Scientist at Google Cloud AI, author of Machine Learning for Drug Discovery, and UC Berkeley adjunct faculty. </subtitle><author><name>Noah Flynn</name><email>noahrflynn@gmail.com</email></author><entry><title type="html">Druggability, ligandability, and modality choice in the AlphaFold 3 era</title><link href="https://noahrflynn.com/blog/2026/druggability-assessment-alphafold-3/" rel="alternate" type="text/html" title="Druggability, ligandability, and modality choice in the AlphaFold 3 era"/><published>2026-05-08T16:20:00+00:00</published><updated>2026-05-08T16:20:00+00:00</updated><id>https://noahrflynn.com/blog/2026/druggability-assessment-alphafold-3</id><content type="html" xml:base="https://noahrflynn.com/blog/2026/druggability-assessment-alphafold-3/"><![CDATA[<p>In 2002, a foundational analysis estimated that roughly 10–15% of the human proteome - the full set of proteins a genome can produce - could be hit by small-molecule drugs. Restated in current terms, <strong>80–90% of the proteome is “undruggable” by conventional small-molecule inhibitors</strong>: proteins without the deep, hydrophobic pockets that support high-affinity ligand binding, including flat protein–protein interaction surfaces, where proteins contact each other over broad shallow areas, transcription factors, and intrinsically disordered regions, which do not settle into a stable folded structure. That estimate shaped two decades of target selection. If a protein fell outside the druggable fraction, teams usually moved on to another project. By 2026, the picture is different. AlphaFold 3 can jointly predict a protein’s structure with its ligand, DNA, RNA, and post-translational modifications in one pass. The AlphaFold Database now covers roughly 200 million predicted structures. One PROTAC is awaiting an FDA June 5, 2026 PDUFA decision - the agency’s target date for acting on a New Drug Application - and the first regulatory test of a modality class that barely existed in the clinic a decade ago. The operative question in 2026 is <em>which modality fits this target best?</em></p> <blockquote> <p><strong>Concept Translation:</strong> “Undruggable” is an artifact of the tool - it means the <em>current</em> inhibitor toolkit (small molecules that fit deep pockets) can’t act on this protein. It does not mean the protein is intrinsically resistant to intervention. The ML analogy: a feature you cannot perturb with your current intervention API. Once the API expands (PROTACs, ASOs, ADCs), the same protein moves from undruggable to druggable without any biological change.</p> </blockquote> <h2 id="why-druggability-assessment-matters-in-2026">Why druggability assessment matters in 2026</h2> <picture> <source srcset="/assets/img/blog/target-discovery/APPH_FLYNN_UN09.webp" type="image/webp"/> <img src="/assets/img/blog/target-discovery/APPH_FLYNN_UN09.png" alt="Two-track timeline from 2018 to 2026 - AlphaFold generations on top, modality milestones below, both inflecting in 2020–2024." loading="lazy"/> </picture> <p>Two shifts have moved the frontier of druggability at roughly the same time, and together they change the analysis.</p> <p>The first is structural. AlphaFold 2 turned single-chain protein structure prediction from a research problem into routine infrastructure in 2020. AlphaFold 3 in May 2024 extended that capability to biomolecular complexes - proteins with small-molecule ligands, with DNA and RNA, with ions, and with post-translational modifications, all in a single diffusion-based model (Abramson et al., <em>Nature</em> 630:493–500, 2024). The AlphaFold 2 work earned Demis Hassabis and John Jumper a share of the 2024 Nobel Prize in Chemistry, with David Baker taking the other half for computational protein design. AF3 now sits inside a broader field that includes Boltz-1 and Boltz-2 (MIT Jameel Clinic), Chai-1 and Chai-2 (Chai Discovery), Protenix (ByteDance), HelixFold-3 (Baidu), and OpenFold3. Predicted structures are now a default input to virtual screening - the computational ranking of large compound libraries against a target - rather than a special case.</p> <p>The second shift is modality. Antisense oligonucleotides (ASOs), small interfering RNAs (siRNAs), targeted protein degraders (PROTACs and molecular glues), antibody-drug conjugates (ADCs), and cyclic peptides all have approved products in the market or late-stage assets in development. Each of these expands the “druggable” envelope along a different axis. An RNA-binding ASO can modulate a protein that has no structured binding pocket at all. A PROTAC can take down a scaffolding protein - one that organizes other proteins rather than catalyzing a reaction - even when there is no catalytic activity to inhibit. An ADC can deliver a cytotoxic payload selectively to tumor cells whose surface antigen (a protein displayed on the cell surface) is otherwise non-actionable.</p> <p>A 2015-era druggability analysis might ask <em>does this target have a small-molecule binding pocket, and is its structure known?</em> A 2026 analysis has to ask that question across at least five modality classes. It also has to answer the harder follow-up: <em>if several modalities are tractable, which one fits this target’s biology, safety profile, and commercial context best?</em></p> <h2 id="what-druggable-actually-means">What “druggable” actually means</h2> <picture> <source srcset="/assets/img/blog/target-discovery/APPH_FLYNN_UN10.webp" type="image/webp"/> <img src="/assets/img/blog/target-discovery/APPH_FLYNN_UN10.png" alt="Druggability assessment decomposed into three resolvable questions - target quality, ligandability, and modality fit." loading="lazy"/> </picture> <p>The vocabulary needs care because the field uses these terms loosely.</p> <p><strong>Druggability</strong> refers to the ability of a protein, peptide, or nucleic acid to be modulated by a drug. The 2002 analysis that anchored the field estimated that about 3,000 of the ~20,000 human protein-coding genes - roughly 15% - are druggable in the small-molecule sense, and that only around 27% of those druggable genes actually have an approved drug. Small-molecule drugs target roughly 2–5% of the human genome in total. That figure marks the reach of current chemistry.</p> <p>Druggability decomposes into two more specific questions.</p> <p><strong>Target quality</strong> asks whether modulating the target will actually change the disease, and whether doing so is safe. The relevant variables are the target’s role in essential cellular processes, its network centrality (how many pathways and interactions depend on it), its tissue-expression pattern (a target essential for heart-muscle function is a bad target, almost regardless of its affinity profile), its approval-adjacent precedent (whether closely related targets or modalities have already cleared regulatory review), and its cellular localization (an intracellular target requires permeability that a cell-surface target does not). Target quality is what [target–disease evidence] → <em>How to tell a drug target matters: evidence frameworks for target–disease linkage</em> (C1) gives you the machinery to evaluate.</p> <p><strong>Ligandability</strong> asks whether a drug molecule can bind the target in a way that modulates it. Ligandability depends on three things: the availability of the target’s structure (experimental or predicted), the presence of closely related proteins with known ligands (if a close cousin of your target has been drugged, the odds are better), and the accessibility of a binding site (a small-molecule pocket for small molecules, a regulatory region for nucleic-acid-targeted drugs, or a surface-exposed patch for antibodies).</p> <blockquote> <p><strong>Concept Translation:</strong> Target quality and ligandability are independent axes. A target can be highly ligandable - it has a deep, well-defined pocket, a structure, and a binder - while being a terrible drug target if hitting it doesn’t change disease or causes serious toxicity. Many of the 80–90% of proteins called “undruggable” in older analyses are really <em>not yet ligandable</em> by small molecules; their target quality may be excellent. New modalities are mostly ligandability-extenders, not target-quality-extenders.</p> </blockquote> <p>Classical small-molecule druggability tools such as DoGSiteScorer (Volkamer et al., 2012), fpocket (Le Guilloux et al., 2009), and SiteMap (Halgren, 2009) turn that third question into a quantitative pocket score. They compute geometric and chemical descriptors of candidate cavities. These remain widely cited workhorses. Newer ML-based approaches (PockDrug, PocketMiner) have emerged and are worth tracking, but the original three are still common reference points in 2026 ligandability analyses. Structure prediction enters the picture at both stages of ligandability: it provides a structure for targets that have never been crystallized, and it increasingly provides one with a candidate ligand already docked - computationally placed into a proposed binding pose.</p> <h2 id="the-classical-druggability-landscape-by-protein-family">The classical druggability landscape by protein family</h2> <p>Before structure prediction started moving the boundary, the druggable genome was shaped largely by which protein families had accessible binding sites and tractable chemistry. A handful of families dominate approved drugs.</p> <p><strong>G protein-coupled receptors (GPCRs)</strong> are the single most targeted class. They sit in the cell membrane and relay outside signals inward. Their extracellular ligand-binding domains can be reached by small molecules, peptide-like molecules, antibodies, and allosteric modulators - drugs that bind away from the main ligand site and tune receptor activity.</p> <p><strong>Ion channels</strong> provide a good contrast. The sequenced human genome identified more than 400 putative ion channels, though only a fraction have been cloned and functionally characterized. Their widespread tissue distribution and the physiological consequences of opening or closing them make them both compelling and treacherous targets.</p> <p><strong>Kinases</strong> are the third-most-targeted group. Kinases transfer phosphate groups from ATP to substrate proteins, and the ATP-binding pocket is the canonical small-molecule site. Five inhibitor strategies recur across the literature: ATP mimics (the default), covalent inhibitors (which form a bond with a cysteine in or near the pocket and are irreversible), bivalent inhibitors (binding the catalytic site and a second stability or localization site), allosteric inhibitors that bind distal regulatory sites (type III and IV), and degraders or molecular glues that hijack the ubiquitin–proteasome system to remove the kinase rather than inhibit it.</p> <blockquote> <p><strong>Concept Translation:</strong> Kinases are useful to keep in mind because they’re the canonical “well-modeled” target family - thousands of crystal structures, deep mutagenesis data, and a consistent pocket geometry across the family. They’re roughly the ImageNet of structure-based drug design: the family where everything was tuned first and where most ML benchmarks were established. The frontier targets are unlike kinases - flat surfaces, no obvious pocket, sparse training data - and that’s why the modality expansion matters.</p> </blockquote> <p>Kinase degraders blur the classical taxonomy. A “kinase inhibitor” and a “kinase degrader” act on the same target through different mechanisms, so the druggability criteria differ as well. That overlap sets up the rest of this post.</p> <h2 id="structure-prediction-and-the-20202024-inflections">Structure prediction and the 2020/2024 inflections</h2> <p>For most of drug discovery’s history, structure-based drug design was available only for targets that had been successfully crystallized. The c-Met/crizotinib program is a useful illustration of that earlier workflow. c-Met is a receptor tyrosine kinase with elevated levels and abnormal activity in several cancers. Pfizer’s program started from a series of pyrrole-substituted 2-indolinones (oxindoles), ATP-competitive tyrosine kinase inhibitors first explored in the mid-1990s, that showed activity against c-Met but poor drug-like properties. Iterative co-crystallography with PHA-665752 (an early lead) revealed the c-Met ATP-pocket environment in detail; that structure motivated a switch to a novel 2-amino-5-aryl-3-benzyloxypyridine series, which yielded the 3,5-disubstituted 2-aminopyridine clinical candidate crizotinib (PF-02341066, Cui et al., <em>J. Med. Chem.</em> 54:6342–6363, 2011). Each iteration of the design cycle was gated on a new crystal. Crizotinib, a c-Met/ALK dual inhibitor, was approved by the FDA in 2011.</p> <p>That workflow still works, and for targets that crystallize well it remains the standard. Crystallography, however, is slow and expensive. For many targets - especially membrane proteins, intrinsically disordered regions, and transient complexes - it either fails or produces low-resolution artefacts that are hard to use for design.</p> <p>Two inflection points changed the practical calculus.</p> <p><strong>2020: AlphaFold 2.</strong> At CASP14, the community benchmark competition for protein-structure prediction, DeepMind’s AlphaFold 2 produced predictions indistinguishable from experimental structures for a majority of single-chain proteins. By 2022 the AlphaFold Database had released ~800,000 predicted structures. By 2025 it was around <strong>200 million predicted structures</strong>, covering the bulk of UniProt, the standard protein-sequence database. Single-chain structure prediction stopped being a bottleneck.</p> <p><strong>2024: AlphaFold 3.</strong> AF3 (Abramson et al., <em>Nature</em> 630:493–500, 2024) moved from single-chain prediction to <em>complex</em> prediction using a diffusion-based architecture built around a Pairformer block - which updates pairwise residue relationships - plus a diffusion module. It jointly predicts proteins with small-molecule ligands, DNA, RNA, ions, and post-translational modifications in a single forward pass. On the PoseBusters ligand-pose benchmark, a test of whether models place ligands in the right binding pose, DeepMind reports AF3 as roughly 50% more accurate than the best traditional physics-based docking tools. That pushes protein–ligand pose-prediction success rates to about 76% versus roughly 50% for the docking programs AutoDock Vina and Gold on the same benchmark, and makes AF3 the first AI system to outperform physics-based tools for biomolecular structure prediction. Model weights were released for non-commercial academic use in November 2024, with broader (still non-commercial) availability in February 2025. Commercial use remains gated behind Isomorphic Labs partnerships.</p> <blockquote> <p><strong>Concept Translation:</strong> The AF2-to-AF3 jump is conceptually familiar to anyone who has watched generative-model architectures evolve. AF2 was a transformer-style model that emitted a single deterministic structure. AF3 swaps the structure-emission head for a <em>diffusion</em> module - the same general framework as image diffusion models - that samples from a distribution over plausible structures. The Pairformer is a transformer-style block that operates on pairwise residue relations rather than tokens. The architectural pattern (encoder over pairwise relations + diffusion decoder) is now the template for every AF3-class model in the field.</p> </blockquote> <p>The competitive field matters because AF3’s non-commercial licensing makes commercially usable alternatives important for industrial drug discovery:</p> <ul> <li><strong>Boltz-1</strong> (MIT, Nov 2024): MIT license, fully open source, matches AF3 accuracy on standard benchmarks. Boltz-1x added physics-based steering to reduce steric clashes; Boltz-2 (2025) added binding-affinity prediction, claiming roughly 2× the precision of standard ML and docking baselines on hit discovery.</li> <li><strong>Chai-1</strong> and <strong>Chai-2</strong> (Chai Discovery): Apache 2.0 license, inference-only weights, can run with or without a multiple-sequence alignment. Chai-2 (2025) introduced zero-shot de novo antibody design with reported experimental hit rates around 16–20% on novel targets.</li> <li><strong>Protenix</strong> (ByteDance, 2025), <strong>HelixFold-3</strong> (Baidu), and <strong>OpenFold3</strong> (in preview as of late 2025) round out the field.</li> </ul> <p>By 2026, teams treat predicted structures as routine inputs to virtual screening and structure-based design. AF3-class models have substantially reduced the need for experimental crystallographic structures in early-discovery programs, particularly for targets where co-crystal structures are slow or expensive to generate. Virtual screening campaigns at 10K-compound scale now run in days on cloud GPUs.</p> <h3 id="a-worked-example-alphafold-enabled-design-for-an-undrugged-target">A worked example: AlphaFold-enabled design for an undrugged target</h3> <p>A 2022 hepatocellular carcinoma (HCC) program provides the first published demonstration of using AlphaFold-predicted structures for hit identification in practice (Ren, Ding, Zheng et al., <em>Chemical Science</em> 14:1443–1452, 2023). The target, CDK20, also known as cell cycle-related kinase (CCRK), had no experimental structure and was effectively undrugged. Traditional structure-based design on CDK20 would have required first solving its crystal structure, a multi-year project in itself.</p> <p>Instead, the team used the AlphaFold-predicted structure (AF-Q8IZL9-F1-model_v1) of CDK20 directly. The prediction was high-confidence across most of the protein except the C-terminus, which they removed because it occluded the solvent-exposed region of the ATP pocket. Pocket analysis identified a shallow ATP binding pocket of approximately 150 Å³ in the DFG-in conformation - one common active-state arrangement of kinase active-site residues - with <strong>MET84 as the hinge residue</strong> and <strong>PHE81 occupying the gatekeeper position</strong> (the hinge is where many kinase inhibitors make key hydrogen bonds; the gatekeeper residue helps control access to the back pocket). They fed the AlphaFold structure into Insilico Medicine’s generative chemistry platform (Chemistry42), which generated <strong>8,918 candidate molecules</strong>. After docking, clustering, and pose inspection, seven were synthesized for the first round of testing. The first hit (ISM042-2-001) bound CDK20 with a Kd of 9.2 ± 0.5 µM (lower Kd means tighter binding) and was identified <strong>30 days from target selection</strong>. A second optimization round produced ISM042-2-048, with Kd ≈ 567 nM and a CDK20 kinase-inhibition IC50 of <strong>33.4 ± 22.6 nM</strong> - the concentration required to cut enzyme activity by half. ISM042-2-048 also showed selective antiproliferation against the CDK20-overexpressing HCC line Huh7 (cellular IC50 ≈ 209 nM) versus the HEK293 counter-screen (~1707 nM), a comparison cell line used to check selectivity.</p> <p>AlphaFold converted a target that would previously have been deferred until crystallography succeeded into a target that could be designed against immediately. In programs that use this workflow, the structural-modeling portion of the pipeline has moved from years to weeks. Hit-to-lead optimization, preclinical pharmacology, and everything downstream still take the time they always did.</p> <p>Compared with the crizotinib workflow described above, the difference is sharp. Crizotinib needed a crystal of each intermediate compound; CDK20’s lead compound was designed against an AI-predicted structure before any experimental structure existed. Structure-based drug design used to be gated on crystallography. In 2026 it is gated on whether the structure-prediction model is accurate enough at the specific pocket you care about. That is a softer gate, and the tooling keeps improving.</p> <h2 id="beyond-small-molecules-modality-choice-is-now-a-first-class-question">Beyond small molecules: modality choice is now a first-class question</h2> <picture> <source srcset="/assets/img/blog/target-discovery/APPH_FLYNN_UN11.webp" type="image/webp"/> <img src="/assets/img/blog/target-discovery/APPH_FLYNN_UN11.png" alt="Six-by-six matrix showing which modalities address which target challenges; no single modality dominates all cells." loading="lazy"/> </picture> <p>Categories of therapeutic molecule that did not have approved products a decade ago now have market-stage drugs and late-stage candidates, each reaching targets that small molecules cannot and expanding the druggable envelope.</p> <h3 id="targeted-protein-degradation-protacs-and-molecular-glues">Targeted protein degradation (PROTACs and molecular glues)</h3> <p>Targeted protein degradation (TPD) shifts pharmacology from <em>occupancy-driven</em> - where you block a protein’s function by occupying its binding site - to <em>event-driven</em>, where you induce the cell’s endogenous waste-disposal machinery to destroy the protein outright. The relevant machinery is the ubiquitin–proteasome system (UPS), the main pathway cells use to tag and remove unwanted intracellular proteins. A cascade of three enzymes - E1 (activating), E2 (conjugating), and E3 (ligase) - tags target proteins with polyubiquitin chains and marks them for destruction by the 26S proteasome. TPD drugs hijack this cascade.</p> <blockquote> <p><strong>Concept Translation:</strong> The occupancy-vs-event shift is closer to a control-flow change than a chemistry change. An occupancy-driven inhibitor is a function call that holds an object in a blocked state for as long as the call is on the stack - release the call, the object returns to normal. An event-driven degrader is a single instruction that triggers garbage collection on the object. The first scales with drug concentration; the second scales with the rate at which the cell remakes the protein. That difference is why PROTACs can be effective at much lower concentrations than equivalent inhibitors and why their pharmacology is often dominated by protein resynthesis rates rather than drug residence time.</p> </blockquote> <p>A <strong>PROTAC (PROteolysis TArgeting Chimera)</strong> is a heterobifunctional molecule - it has one binding end for the target and another for an E3 ligase, joined by a linker. By forming a ternary complex (a three-part assembly of target, PROTAC, and ligase), it brings the target into proximity with the ubiquitination machinery. Because of this tripartite structure, PROTACs are large molecules, typically <strong>700–1,000 Da</strong>. A PROTAC also acts <em>catalytically</em>. Once the target is ubiquitinated and released into the proteasome, the PROTAC dissociates and recruits another target copy, so one degrader molecule can remove many target molecules. It can therefore hit targets that have no functional small-molecule binding site at all, as long as <em>some</em> binder exists for the target.</p> <p><strong>Molecular glues</strong> are the close cousin: small molecules typically <strong>&lt;500 Da, with no linker</strong>, that bind directly to an E3 ligase and induce a conformational change in the ligase surface, creating a novel interaction interface that recruits “neosubstrate” proteins (proteins the ligase would not normally bind) for degradation. The original IMiDs (thalidomide, lenalidomide, pomalidomide) are retrospectively understood as molecular glues. In 2026 the line between PROTACs and molecular glues is blurring; newer-generation degraders combine features of both.</p> <p>The opportunity space is large. The human genome encodes <strong>over 600 E3 ligases, of which fewer than 2% have been successfully exploited for TPD to date</strong>. Most current PROTACs recruit just a few E3 ligases, especially CRBN or VHL; DCAF15 and MDM2 are emerging; the rest of the ligase repertoire is largely untouched. Tumor-specific E3 ligases are an active frontier for tissue-selective degradation. The motivation is direct: a CRBN-recruiting PROTAC can degrade its target in healthy tissue as well as diseased tissue.</p> <p>The state of TPD in 2026 is a useful test case for “modality choice as a first-class question”:</p> <ul> <li><strong>Bavdegalutamide (ARV-110)</strong>, Arvinas’s androgen-receptor PROTAC, was the first PROTAC to enter human clinical trials (Phase 1, 2019). In patients with AR T878X/H875Y mutations in ARDENT, it produced 46% PSA50 response rates, where PSA50 is a prostate-cancer endpoint based on prostate-specific antigen decline. Development was <strong>discontinued in 2024</strong> in favor of ARV-766, a next-generation AR degrader with broader mutant coverage. The first clinical entrant did not become the lasting lead program.</li> <li><strong>Vepdegestrant (ARV-471)</strong>, Arvinas/Pfizer’s estrogen-receptor PROTAC, reached Phase 3 (VERITAC-2) in estrogen-receptor-positive, HER2-negative breast cancer after prior CDK4/6 inhibitor failure. Topline results in March 2025 met the primary endpoint in the ESR1-mutant subpopulation, the subgroup whose tumors carried ESR1 mutations (median PFS 5.0 vs 2.1 months, HR 0.57, p&lt;0.001). PFS is progression-free survival, and HR is the hazard ratio comparing event rates over time. The study <strong>did not</strong> reach statistical significance in the intent-to-treat population, meaning all enrolled patients. NDA, the New Drug Application, filed June 2025. <strong>PDUFA action date: June 5, 2026</strong>. If approved, vepdegestrant will be the first FDA-approved PROTAC.</li> <li>In September 2025, Arvinas and Pfizer publicly sought a third-party partner to commercialize vepdegestrant. That move suggests cooling internal commitment despite the positive result in the ESR1-mutant subgroup. The modality has clinical signal, but the commercial case remains unsettled.</li> </ul> <p>The broader TPD space has expanded well beyond ER and AR. CC-94676 (BMS, AR PROTAC), the CELMoDs mezigdomide and iberdomide - cereblon E3-ligase modulators now in Phase 3 in multiple myeloma - and a growing number of Phase 1 molecular-glue degraders for BCL6, BTK, and other targets are all active programs.</p> <p>For a target-assessment document, this changes the screen. For a scaffolding protein or transcription factor without an inhibitable catalytic site, the relevant question becomes “does it have a surface patch we can bind <em>at all</em>?” If yes, a PROTAC may be tractable.</p> <h3 id="rna-targeting-drugs-asos-and-sirnas">RNA-targeting drugs (ASOs and siRNAs)</h3> <p><strong>Antisense oligonucleotides (ASOs)</strong> are 15–30 nucleotide sequences that bind target RNA via Watson–Crick base pairing - that is, ordinary sequence complementarity. Two mechanisms dominate:</p> <ul> <li><strong>RNase H1 cleavage.</strong> The ASO–RNA duplex is a substrate for endogenous RNase H1, an enzyme that cleaves the RNA strand. Most approved ASOs work this way.</li> <li><strong>Steric blockage / splice switching.</strong> The ASO binds pre-mRNA (the RNA transcript before splicing) and prevents ribosomal assembly or alters splicing. Exon-skipping ASOs (Duchenne muscular dystrophy) and exon-inclusion ASOs (spinal muscular atrophy) are the canonical examples.</li> </ul> <p><strong>siRNAs</strong> act through the RNA-induced silencing complex (RISC), a cellular complex that uses the guide RNA to find and silence matching transcripts. Approved examples include patisiran (hereditary transthyretin amyloidosis), givosiran (acute hepatic porphyria), lumasiran (primary hyperoxaluria type 1), and inclisiran/Leqvio - Novartis’s siRNA targeting PCSK9, originally approved by the FDA in December 2021 for adults with ASCVD (atherosclerotic cardiovascular disease) or heterozygous familial hypercholesterolemia, expanded in 2023 to broader primary hyperlipidemia, and updated again in July 2025 to permit first-line monotherapy use in hypercholesterolemia.</p> <blockquote> <p><strong>Concept Translation:</strong> ASOs and siRNAs target <em>messenger RNA</em> rather than the protein itself. The matching is sequence-based - Watson–Crick base pairing is exact-match string lookup over a 4-letter alphabet (A, U, G, C). For target assessment this is a different design space entirely: the target is a string, the drug is a string, and the binding rule is a known function of the two strings. Compare this to small-molecule binding, where the binding rule is a learned function of two 3D shapes. Sequence-matching is much easier to design but introduces a new failure mode - off-target hits anywhere in the transcriptome that match closely enough.</p> </blockquote> <p>For target assessment, ASOs and siRNAs require an accessible sequence rather than a structured <em>protein</em> binding pocket. A target with no deep pocket, no covalent warhead opportunity, and no PROTAC handle can still be tractable if its RNA is accessible. That rule has caveats: small-molecule splice modulators like risdiplam for SMA exploit RNA <em>secondary</em> structure - the local folded shape of the RNA strand - so structure does come back into play at the RNA level for some programs.</p> <h3 id="antibody-drug-conjugates-adcs">Antibody-drug conjugates (ADCs)</h3> <p>An ADC is a monoclonal antibody covalently linked to a cytotoxic payload via a chemical linker. The antibody provides selectivity; the payload provides potency. An ideal ADC stays stable in circulation, internalizes on binding to the target (the antibody–target complex is pulled into the cell), and releases the payload intracellularly in the vicinity of the target.</p> <blockquote> <p><strong>Concept Translation:</strong> An ADC is a delivery system. The antibody is a routing label that says “deliver to cells expressing this surface marker”; the payload is the actual cytotoxic drug, often too toxic to administer freely. The linker controls when the payload separates - ideally only after the package has been internalized. The pattern is roughly the same as targeted delivery in any system: route on a label, release on arrival. The engineering challenges are also analogous (label specificity, in-transit stability, controlled release).</p> </blockquote> <p>2026 status: <strong>15 ADCs are FDA-approved</strong>, with 19 ADCs approved globally across FDA, EMA, NMPA, and PMDA. Pipeline scale is over <strong>400 ADCs in development</strong>, <strong>200+ in clinical trials</strong>, and at least <strong>24 in Phase 3</strong>. Recent FDA approvals include Datroway (datopotamab deruxtecan, January 2025) and Emrelis (telisotuzumab vedotin-tllv, May 2025, targeting c-Met-overexpressing NSCLC, or non-small cell lung cancer). Topoisomerase I inhibitor payloads (DXd, SN-38) now dominate late-stage development due to strong bystander effects, where the released drug can also kill neighboring cells.</p> <p>For target assessment, ADCs open up targets where <em>selective expression</em> matters more than <em>pocket tractability</em>. A tumor-surface antigen that cannot be usefully modulated by a naked antibody (an antibody without an attached payload) can still be a good ADC target if it internalizes on binding.</p> <h3 id="therapeutic-peptides-and-cyclic-peptides">Therapeutic peptides and cyclic peptides</h3> <p>Therapeutic peptides occupy the middle ground between small molecules and biologics. They typically fall in the 500–5,000 Dalton range, often offer high target specificity, are less likely to trigger an immune response than antibodies, and have better membrane permeability than most proteins. Linear peptides suffer from poor stability and short half-life. <strong>Cyclic peptides</strong> address both by constraining the peptide into a ring, which improves rigidity, stability, and permeability. Cyclic peptides sit within the broader <strong>“beyond-Rule-of-Five” (bRo5)</strong> chemical space - shorthand for compounds that break the usual small-molecule size and polarity heuristics. Macrocycles (large rigid rings of 12 or more atoms), PROTACs, peptides, and metallodrugs all populate this space. By 2024, the field broadly accepted that strict adherence to Lipinski’s Rule of Five was not a prerequisite for intracellular engagement, and that bRo5 chemistry was central to engaging the flat protein–protein interaction surfaces that dominate the “undruggable” set. Cyclic peptides have shown activity in oncology, antiviral, antibacterial, and antimalarial contexts.</p> <h3 id="car-t-gene-therapy-cell-therapy">CAR-T, gene therapy, cell therapy</h3> <p>CAR-T and TCR-T therapies genetically modify a patient’s T cells to express a chimeric antigen receptor or T-cell receptor, then infuse the modified cells back. Kymriah and Yescarta were the first two FDA approvals in 2017; Tecartus, Breyanzi, and Abecma followed for mantle cell lymphoma, large B-cell lymphoma, and multiple myeloma respectively. By early 2026, the FDA reported having approved <strong>close to 50 cell and gene therapy products</strong> over the prior decade across the Center for Biologics Evaluation and Research (CBER) Office of Therapeutic Products, the cumulative result of approvals such as Casgevy and Lyfgenia (sickle cell disease), Beqvez (hemophilia B), Elevidys (Duchenne muscular dystrophy), Vyjuvek (dystrophic epidermolysis bullosa), Aucatzyl (B-ALL, or B-cell acute lymphoblastic leukemia), and an ongoing run of CAR-T approvals.</p> <p>For target assessment, CAR-T targets are a different class of object from small-molecule targets. A CAR-T target is a surface antigen expressed, ideally exclusively, on disease cells - the “druggability” question is about on-tumor versus off-tumor expression (whether the antigen appears on cancer cells, healthy cells, or both), not pocket geometry. That leads to the other half of the 2026 druggability question.</p> <h2 id="the-druggable-genome-is-expanding-and-the-ignorome-remains-large">The druggable genome is expanding, and the ignorome remains large</h2> <picture> <source srcset="/assets/img/blog/target-discovery/APPH_FLYNN_UN12.webp" type="image/webp"/> <img src="/assets/img/blog/target-discovery/APPH_FLYNN_UN12.png" alt="TDL pyramid (Tclin / Tchem / Tbio / Tdark) showing the ignorome as the unaddressed majority of the human proteome." loading="lazy"/> </picture> <p>There are more than 10,000 known human diseases. The original 2002 estimate put 3,000–10,000 disease-related genes in the genome, with roughly 10% of those being disease-modifying on knockout (disabling the gene changed a disease-relevant phenotype in model systems). That yields the 600–1,500 small-molecule-druggable target estimate cited at the top of this post.</p> <p>The <strong>Rule of Five</strong> (Lipinski et al., 1997) shaped small-molecule design for a generation: no more than 5 hydrogen-bond donors, 10 acceptors, molecular weight under 500, and LogP under 5, with LogP being a measure of lipophilicity. By 2026, approved drugs increasingly exceed Rule-of-Five boundaries. The main alternative is the <strong>Rule of Three</strong> for fragment-based discovery, a strategy that starts from very small weak-binding molecules (Congreve et al., <em>Drug Discov. Today</em> 8:876–877, 2003: MW &lt; 300, ≤3 H-bond donors, ≤3 H-bond acceptors, cLogP ≤ 3, where cLogP is the computed version of LogP). The empirical boundaries of “drug-like” are shifting as chemistry expands to PROTACs, macrocycles, ADCs, and covalent warheads. These categories routinely break Lipinski’s original rules and still get approved.</p> <blockquote> <p><strong>Concept Translation:</strong> Rule of Five was a hand-crafted feature filter - a set of property thresholds that empirically separated absorbable drugs from non-absorbable ones, fitted on roughly 2,200 oral drugs in the late 1990s. Like any hand-crafted filter, it was tuned to its training distribution. As the chemistry expanded into larger and more polar molecules, the filter started rejecting things that worked. The 2026 stance is to use Lipinski thresholds as a soft prior for the small-molecule region of chemical space and to ignore them entirely outside it.</p> </blockquote> <p>Meanwhile, <strong>over 75% of protein research still focuses on the 10% of proteins known before the human genome was mapped</strong>. The NIH’s Illuminating the Druggable Genome (IDG) initiative, launched in 2014, created a four-level Target Development Level classification to make this visible:</p> <ul> <li><strong>Tclin</strong>: targets linked to at least one approved drug by mechanism of action.</li> <li><strong>Tchem</strong>: proteins known to bind small molecules with high potency, but without approved-drug links.</li> <li><strong>Tbio</strong>: proteins with a confirmed Mendelian disease phenotype or meeting certain experimental criteria.</li> <li><strong>Tdark</strong>: proteins meeting none of the above.</li> </ul> <p>IDG found that Tdark proteins receive less research funding than other categories, which perpetuates the knowledge gap. The <strong>ignorome</strong> - the Tdark set - is both the largest opportunity space for novel target discovery and the hardest place to work, because almost every downstream tool (antibodies, assays, knockout models) has to be built from scratch. The computational tractability gains from AF3-class structure prediction matter most here: a target with no crystal structure and minimal biological characterization now at least comes with a predicted structure to work from.</p> <blockquote> <p><strong>Concept Translation:</strong> The TDL classification is a label hierarchy on the human proteome - closer to a knowledge-graph annotation than a clean training-data split. Tclin proteins have the most data on them (drugs, structures, papers, assays) and Tdark have the least. ML methods trained on protein-level features inherit the same skew as the labels: they perform best on Tclin and worst on Tdark, the long tail. The “ignorome” framing emphasizes that the tail is where novel drug targets actually live.</p> </blockquote> <p>A 2026 assessment of a novel target should include, at minimum, an AF3-class structure prediction with pocket analysis, a TDL classification, a scan across modality options (small molecule, PROTAC, ASO, ADC, peptide), and an explicit note on whether Rule-of-Five chemistry is even the right design envelope for this target. The question [novelty scoring and the ignorome] → <em>Novel vs repurposed targets: quantifying novelty and extending drug-repurposing methods</em> (C5)<sup id="fnref:c5"><a href="#fn:c5" class="footnote" rel="footnote" role="doc-noteref">1</a></sup> takes up how novelty itself is now quantified across the TDL spectrum.</p> <h2 id="whats-next-in-druggability-assessment">What’s next in druggability assessment</h2> <p>By 2026, the central druggability question is “which of five or six modalities fits this target best, and what is the safety profile of hitting it that way?” Three follow-ups stand out.</p> <p><strong>Safety.</strong> A PROTAC that catalytically degrades its target in tumor tissue may also degrade the same target in healthy tissue where it is essential. An ASO that knocks down a transcript in the liver may behave differently in the kidney. Modality-specific on-target, off-tissue toxicity - where the intended target is hit in the wrong tissue - is the next frontier, and it’s directly continuous with the tissue-specificity question the next post in this series picks up: [tissue specificity] → <em>Tissue specificity as a safety filter: GTEx, Human Protein Atlas, and scRNA-seq for target prioritization</em> (C3).<sup id="fnref:c3"><a href="#fn:c3" class="footnote" rel="footnote" role="doc-noteref">2</a></sup></p> <p><strong>Two targets, not one.</strong> The entire framing of druggability assumes you’re choosing a single target and asking what modality hits it. For some diseases, most notably cancers with defined synthetic-lethal vulnerabilities (where a cell tolerates either perturbation alone but not both together), the right answer is that you need to hit two targets together, and each of those targets individually is a poor drug candidate. That shifts the druggability question to a two-body problem, with its own computational and experimental methods: [synthetic lethality] → <em>Synthetic lethality and combination targets: ML methods for finding drug pairs that work together</em> (C7).<sup id="fnref:c7"><a href="#fn:c7" class="footnote" rel="footnote" role="doc-noteref">3</a></sup></p> <p><strong>Benchmarks for modality-choice models.</strong> We don’t yet have a standard benchmark for “given this target, which modality will work best?” The PoseBusters benchmark answers a narrower question (given this target and this ligand, is the predicted pose right?). Benchmarks for PROTAC ternary-complex geometry, ASO off-target profiles, and ADC internalization kinetics exist in pieces but do not yet form a coherent evaluation suite. That gap is a natural target for the next round of infrastructure investment. For the pillar-post view of where this fits in the overall workflow: [the pillar post on drug target discovery] → <em>Target discovery: the front-of-funnel decision behind most Phase II failures</em> (P0).</p> <p>AlphaFold 3 was an inflection point. The remaining work now sits in modality selection, safety, and benchmark design.</p> <hr/> <h2 id="further-reading">Further reading</h2> <p>This Further Reading section leans on primary papers rather than review-heavy framing because the topic dates quickly. AF3 is two years old at the post’s publication, PROTAC/TPD clinical status moves every few months, and the ADC field adds approvals on a rolling basis. Primary papers and canonical databases age more gracefully.</p> <ul> <li> <p>Abramson, Adler, Dunger, Evans, et al. (2024). <em>Accurate structure prediction of biomolecular interactions with AlphaFold 3.</em> <em>Nature</em> 630, 493–500. DOI: 10.1038/s41586-024-07487-w. The primary source for every AF3-specific claim in the post, including the PoseBusters benchmark, the Pairformer-plus-diffusion architecture, and the joint protein–ligand–DNA–RNA prediction capability. Cited rather than a more recent review because the benchmark numbers in reviews inherit from this paper anyway. The competing AF3-class models (Boltz-2, Chai-2, Protenix, OpenFold3) have released benchmark head-to-heads since 2024; readers building a current-day pipeline should consult the relevant tool’s release notes for the most up-to-date head-to-head numbers.</p> </li> <li> <p>Hopkins, A. L., and Groom, C. R. (2002). <em>The druggable genome.</em> <em>Nature Reviews Drug Discovery</em> 1, 727–730. The foundational paper that anchors the “10–15% of the proteome is druggable” framing used throughout the post’s definitional section. Oprea, Bologa, Brunak et al. (2018), <em>Unexplored therapeutic opportunities in the human genome</em>, <em>Nature Reviews Drug Discovery</em> 17, 317–332, is the closest-to-current methodological update and provides the Tclin/Tchem/Tbio/Tdark classification that the ignorome section discusses - consider citing it alongside Hopkins &amp; Groom if the post needs a more current primary reference.</p> </li> <li> <p>Ren, F., Ding, X., Zheng, M., et al. (2023). <em>AlphaFold accelerates artificial intelligence powered drug discovery: efficient discovery of a novel CDK20 small molecule inhibitor.</em> <em>Chemical Science</em> 14, 1443–1452. DOI: 10.1039/d2sc05709c. The primary source for the worked CDK20/HCC example used in the post - the first published demonstration of AlphaFold-predicted structures driving hit identification end-to-end. Verifies the 30-day target-to-hit timeline, the 8,918-molecule generative pool, the MET84/PHE81 hinge–gatekeeper geometry, and the ISM042-2-048 33 nM IC50 figure.</p> </li> <li> <p>Cui, J. J., Tran-Dubé, M., Shen, H., et al. (2011). <em>Structure based drug design of crizotinib (PF-02341066), a potent and selective dual inhibitor of mesenchymal–epithelial transition factor (c-MET) kinase and anaplastic lymphoma kinase (ALK).</em> <em>Journal of Medicinal Chemistry</em> 54, 6342–6363. DOI: 10.1021/jm2007613. The medicinal-chemistry case history behind the crizotinib comparison - documents the path from pyrrole-substituted oxindoles through PHA-665752 to the 2-aminopyridine clinical candidate, including the iterative crystallography that the AF3-era CDK20 workflow makes optional.</p> </li> <li> <p>NIH Illuminating the Druggable Genome Program. <em>Pharos: the IDG Knowledge Management Center’s public resource for target development-level classification.</em> pharos.nih.gov. The canonical open-access source for the TDL classification (Tclin/Tchem/Tbio/Tdark) the post uses in its ignorome section. Cited as the database home page rather than a specific IDG paper because the TDL categories are continuously updated against new drug approvals and omics evidence - a release-notes or current-database link ages better than a one-time publication link.</p> </li> </ul> <div class="footnotes" role="doc-endnotes"> <ol> <li id="fn:c5"> <p>Coming in a later post - novelty scoring across the TDL spectrum. <a href="#fnref:c5" class="reversefootnote" role="doc-backlink">&#8617;</a></p> </li> <li id="fn:c3"> <p>Coming in the next post - tissue specificity as a safety filter. <a href="#fnref:c3" class="reversefootnote" role="doc-backlink">&#8617;</a></p> </li> <li id="fn:c7"> <p>Coming in a later post - synthetic lethality and combination targets. <a href="#fnref:c7" class="reversefootnote" role="doc-backlink">&#8617;</a></p> </li> </ol> </div>]]></content><author><name>Noah Flynn</name><email>noahrflynn@gmail.com</email></author><category term="target discovery"/><category term="machine learning"/><category term="drug discovery"/><category term="target discovery"/><category term="protein structure"/><summary type="html"><![CDATA[Druggability assessment used to ask "does this protein have a pocket?" In 2026 it asks "which of six modalities fits this target best?" A field guide.]]></summary></entry><entry><title type="html">How to tell a drug target matters: evidence frameworks for target–disease linkage</title><link href="https://noahrflynn.com/blog/2026/target-disease-association-evidence/" rel="alternate" type="text/html" title="How to tell a drug target matters: evidence frameworks for target–disease linkage"/><published>2026-05-08T16:10:00+00:00</published><updated>2026-05-08T16:10:00+00:00</updated><id>https://noahrflynn.com/blog/2026/target-disease-association-evidence</id><content type="html" xml:base="https://noahrflynn.com/blog/2026/target-disease-association-evidence/"><![CDATA[<p>A drug candidate can hit its intended protein with favorable potency, selectivity, and pharmacokinetics and still fail in Phase II because the <em>target</em> was never causally linked to the disease in humans. Nelson et al. (2015, <em>Nature Genetics</em>), followed by a revised analysis from King, Davis, and Degner (2019, <em>PLOS Genetics</em>), established that drug programs backed by human genetic evidence are roughly twice as likely to succeed in Phase II and Phase III as programs without it. BIO/Informa industry data over 2011–2020 put biomarker-stratified programs - programs that enroll or analyze patient subgroups defined by a measurable biomarker - at roughly 2× the likelihood of approval versus unstratified ones, a finding that has replicated in subsequent analyses. So evidence for <em>why this target, for this disease, in this patient population</em> carries much of the Phase II burden long before the drug reaches humans.</p> <blockquote> <p><strong>Concept Translation:</strong> Think of a drug target as a feature in a model that you’ve decided to intervene on. Picking a feature that strongly correlates with disease tells you nothing about whether modulating it will change the outcome - that’s the difference between a predictive feature and a causal one. Most of this post is about the evidence techniques the field uses to tell these apart for biological “features” (genes and proteins).</p> </blockquote> <h2 id="why-targetdisease-evidence-matters-in-2026">Why target–disease evidence matters in 2026</h2> <p>Industry-wide Phase I likelihood of approval has fallen from about 10.4% to 6.7%. The drivers include a shift toward first-in-class and riskier targets, greater use of biomarkers as surrogate efficacy endpoints (early readouts used as stand-ins for clinical benefit), and tougher regulatory scrutiny. As easy targets are mostly drugged, what remains rests on thinner evidence bases. Biomarker-stratified programs run at roughly double the likelihood of approval of unstratified ones, and targets backed by human genetic evidence run at roughly double the success rate of targets without that support.</p> <p>Two industry-side facts motivate greater evidence rigor. First, the pipeline is crowded around a small number of targets. About 25% of the global R&amp;D pipeline (roughly 13,600 drug-target pairs) concentrates on just 38 unique biological targets, and in oncology specifically the number of assets per target has grown from 1.8 to 9 over the last two decades (Plenge, 2026). When nine companies are pursuing the same protein, evidence becomes the differentiator - which program has the clearest human-genetic case linking target to disease, the most defensible biomarker, and the cleanest safety picture after accounting for off-target effects (unwanted activity at proteins other than the intended target). Second, human genetic evidence now accompanies approximately two-thirds of recently FDA-approved therapeutics (Ochoa et al., 2022, <em>Nat Rev Drug Discov</em>; Trajanoska et al., 2023, <em>Nature</em>). What was a competitive advantage a decade ago is now table stakes for serious target programs.</p> <h2 id="what-targetdisease-association-evidence-actually-means">What target–disease association evidence actually means</h2> <picture> <source srcset="/assets/img/blog/target-discovery/APPH_FLYNN_UN04.webp" type="image/webp"/> <img src="/assets/img/blog/target-discovery/APPH_FLYNN_UN04.png" alt="Four-step target–disease evidence argument: mechanistic hypothesis, human evidence, experimental perturbation, competing hypotheses." loading="lazy"/> </picture> <p>Target-disease evidence is a structured argument built from heterogeneous data sources. A credible argument has four parts:</p> <ol> <li><strong>A mechanistic hypothesis.</strong> A story about how modulating this protein would alter the disease phenotype. We seek a plausible causal chain, going beyond naive correlation with the disease state.</li> <li><strong>Human evidence that the hypothesis is correct.</strong> Genetic when possible, since human genetic variation imitates the design of a randomized controlled trial without requiring a drug intervention. Multi-omics, patient-tissue, and literature evidence all count, with known caveats.</li> <li><strong>Experimental evidence that the modulation produces the therapeutic effect.</strong> Knockout, knockdown, or pharmacological perturbation - that is, genetic removal, genetic reduction, or chemical modulation of the target - in disease-relevant models, ideally more than one, that reverses the disease phenotype rather than only altering a biomarker.</li> <li><strong>A defensible position on the competing hypothesis.</strong> Why this mechanism rather than the three adjacent ones the literature also supports?</li> </ol> <blockquote> <p><strong>Concept Translation:</strong> Knockouts and knockdowns are the biology field’s version of ablation studies. Knockout = the gene is removed entirely (full ablation). Knockdown = expression is reduced but not eliminated (partial ablation, more like dropout). Pharmacological perturbation = a chemical inhibits the protein, which is closer to “intervening at inference time” than “retraining without the feature.” All three answer the same question: <em>what does the system do when this component is missing or weakened?</em></p> </blockquote> <p>Teams that skip one of these four risk failure. The most common failure is #4 - the team has strong data on its favored mechanism and has never seriously engaged with the alternatives.</p> <p>Two frameworks for organizing this argument are the AstraZeneca 5R framework and the GOT-IT recommendations.</p> <h3 id="the-astrazeneca-5r-framework-right-target-as-the-first-r">The AstraZeneca 5R framework: “right target” as the first R</h3> <picture> <source srcset="/assets/img/blog/target-discovery/APPH_FLYNN_UN05.webp" type="image/webp"/> <img src="/assets/img/blog/target-discovery/APPH_FLYNN_UN05.png" alt="AstraZeneca 5R framework with &quot;Right Target&quot; highlighted; the four other R&#39;s map to companion posts in the series." loading="lazy"/> </picture> <p>The 5R framework defines five criteria a drug program has to satisfy: <strong>right target</strong>, <strong>right tissue</strong>, <strong>right safety profile</strong>, <strong>right patient</strong>, <strong>right commercial potential</strong>. The “right target” criterion is defined narrowly as a strong link between target and disease, predictive biomarkers that identify likely responders, and demonstrated differentiated efficacy relative to alternatives. Lack of efficacy has been the most important cause of project failure in clinical trials. The 5R framework forces teams to separate evidence of <em>target-disease linkage</em> from evidence of <em>tissue availability</em>, <em>safety</em>, and <em>patient stratification</em> (the choice of which patient subgroup to treat). A single piece of data often speaks to only one of these categories, and conflating them makes the evidence base look stronger than it is.</p> <p>The framework originated at AstraZeneca in the mid-2010s with Cook et al. (2014) <em>Nature Reviews Drug Discovery</em>, followed by Morgan et al. (2018) in <em>Drug Discovery Today</em>, which documented an approximately five-fold improvement in the company’s small-molecule project survival from candidate-selection to Phase III decision after the framework was institutionalized.</p> <h3 id="got-it-four-assessment-blocks">GOT-IT: four assessment blocks</h3> <p>The GOT-IT (Good Target Identification and Validation) recommendations, published by Emmerich et al. (2021) in <em>Nature Reviews Drug Discovery</em>, organize target assessment into four blocks. The first is the focus of this post:</p> <ul> <li><strong>Disease linkage:</strong> evaluating the relationship between the target and the disease of interest and determining whether the target is involved in the underlying disease biology.</li> <li><strong>Target-related safety:</strong> covered in a later post on tissue-specificity.</li> <li><strong>Strategic issues:</strong> commercial potential, portfolio fit, competition.</li> <li><strong>Technical feasibility:</strong> covered in a later post on druggability.</li> </ul> <p>5R and GOT-IT overlap, but they aren’t redundant. 5R is a pipeline-level framework, like a checklist for whether a program should advance. GOT-IT is a target-level framework, a checklist for whether the target-assessment work has been done thoroughly. Many organizations use both, sometimes customizing one internally and treating the other as the external benchmark, or using them in tandem with additional frameworks.</p> <h2 id="four-important-distinctions-within-targetdisease-linkage">Four important distinctions within target–disease linkage</h2> <h3 id="driver-vs-passenger-mutations">Driver vs passenger mutations</h3> <p>In cancer genomics, <strong>driver mutations</strong> provide selective advantages for cancer initiation and progression. <strong>Passenger mutations</strong> accumulate alongside them without contributing to disease fitness. A single gene can host both classes. The APC gene, a highly mutated tumor suppressor in colorectal cancer, carries both driver and passenger mutations across its coding region. Distinguishing them is a computational problem: which variants observed in a tumor caused the malignant phenotype, and which reflect background mutation noise?</p> <blockquote> <p><strong>Concept Translation:</strong> Drivers and passengers are the cancer-genomics version of signal vs. correlated noise. A tumor genome contains hundreds to thousands of mutations, only a handful of which actually move the phenotype. The other mutations co-occur with the drivers because they accumulated in the same lineage. Picking drivers out of the noise is the same kind of feature-selection problem as identifying which input variables actually drive a model’s predictions when the inputs are highly correlated.</p> </blockquote> <p>A list of “genes mutated in disease X” is raw material. A target-discovery workflow has to rank those genes by probability of being a driver, using statistical tests on mutation frequency against background, functional impact scoring, recurrence across patients, pathway enrichment, and so on. Tools like MutSigCV, OncodriveCLUST, and later deep-learning classifiers exist for this. The output of a cancer-genomics pipeline is a ranked driver list. Moving from that list to a target list requires additional evidence.</p> <p>A related wrinkle is that some passengers are clinically informative. Weak drivers that fall below detection thresholds, and accumulated “passenger” burden that shifts cell phenotype cumulatively, can matter clinically. Some passengers also provide evolutionary evidence of tumor lineage and can be used for timing inferences.</p> <h3 id="causative-supportive-and-symptomatic-treatment">Causative, supportive, and symptomatic treatment</h3> <p>A second distinction concerns the kind of therapeutic effect a target is meant to produce:</p> <ul> <li><strong>Causative treatment:</strong> therapy directed against the cause of disease. Antiviral agents against SARS-CoV-2. Ritonavir inhibiting HIV protease. A PARP inhibitor in a BRCA-mutated tumor. Targets for causative treatment live upstream in the disease mechanism.</li> <li><strong>Symptomatic treatment:</strong> therapy that eases symptoms without addressing the underlying cause. Pain relievers. Cough suppressants. Targets for symptomatic treatment live downstream of the disease cause, typically in host response pathways.</li> <li><strong>Supportive treatment:</strong> care that addresses broader patient needs (physical, existential, palliative) rather than targeting disease biology directly.</li> </ul> <p>For target discovery, this distinction clarifies what “linkage to the disease” has to establish. A target for causative treatment has to be causally upstream of the disease process. A target for symptomatic treatment has to be causally upstream of a symptom, which is a weaker claim. A program’s target-evidence burden depends on which kind of drug it’s trying to develop.</p> <p>Alzheimer’s provides a standard example of a disease where causative target discovery has resisted easy success. Acetylcholinesterase inhibitors (donepezil, rivastigmine, galantamine) and memantine provide symptomatic therapies that target downstream neurotransmitter biology rather than disease etiology. The amyloid-targeting antibodies aducanumab (accelerated approval 2021, voluntarily withdrawn in 2024), lecanemab (Leqembi, FDA traditional approval July 2023, EMA approval late 2024), and donanemab (Kisunla, FDA approval July 2024) aim at a putative causal mechanism. Phase III data for both lecanemab and donanemab show statistically significant slowing of cognitive decline relative to placebo over 18 months. Debate remains over whether those effect sizes are clinically meaningful. The safety burden of amyloid-related imaging abnormalities (ARIA) has also been a major clinical concern, and the EMA has had a mixed reception of the two drugs. The target-disease linkage evidence for amyloid is therefore stronger than it was a decade ago - there is now Phase III evidence that targeting amyloid produces a measurable clinical effect - but this does not clarify whether amyloid is the causative driver of cognitive decline, a downstream marker, or one of several parallel drivers.<sup id="fnref:amyloid"><a href="#fn:amyloid" class="footnote" rel="footnote" role="doc-noteref">1</a></sup> The earlier generation of amyloid-targeting programs provides a cautionary tale, though continuing studies test whether the hypothesis was right but underpowered, or only partly right with diminishing returns.</p> <h3 id="oncogene-addiction">Oncogene addiction</h3> <p>Oncogene addiction describes cancer cells’ dependence on individual oncogenes - genes whose altered activity drives tumor growth - to sustain the malignant phenotype. A tumor that depends on an activated oncogene is selectively vulnerable to inhibition of that oncogene. Tumor cells need the oncogenic signal to survive; healthy cells do not. This has been a successful organizing concept in targeted oncology. EGFR mutations in a subset of non-small cell lung cancer, BCR-ABL fusion in chronic myeloid leukemia, and BRAF V600E mutations in melanoma all produce tumors with oncogene-addicted phenotypes, and targeted inhibitors in each context have produced durable clinical responses.</p> <p>For target-discovery methodology, oncogene addiction is useful because it predicts a specific clinical-trial readout: patients whose tumors harbor the addiction should respond, and others should not. That prediction is testable in Phase II with biomarker-stratified cohorts, and it helps explain why biomarker-stratified oncology programs show higher LOA than unstratified ones (we’ll quantify this in a later post on LOAs by therapeutic area). An evidence package for an oncogene-addiction target should therefore include the candidate biomarker - the measurable feature that marks the likely responder subgroup.</p> <h3 id="target-disease-correlation-vs-causal-target-disease-linkage">Target-disease correlation vs causal target-disease linkage</h3> <p>The most consequential distinction is also the most abstract. A gene whose expression differs between disease tissue and healthy tissue is <em>associated</em> with the disease. A gene whose loss-of-function variants - variants that reduce or abolish a gene’s activity - segregate with disease risk in human populations at a genetic locus that survives multiple-testing correction, with a plausible protein-level mechanism and Mendelian-randomization-style evidence that inherited variation in the gene affects disease rather than the reverse, is a <em>causal candidate</em>. Curating such evidence is hard. Transcriptomic association is relatively cheap; causal genetic evidence is expensive and sparse.</p> <blockquote> <p><strong>Concept Translation:</strong> Mendelian randomization is the genetics field’s instrumental-variable method. Because alleles are randomly assorted at conception, an inherited variant acts like a randomized treatment assignment - it isn’t influenced by lifestyle, environment, or downstream disease state. If carriers of a loss-of-function variant in gene X are healthier on average, that’s strong evidence that <em>reducing X’s activity</em> reduces disease risk, not just that X is correlated with it. This is closer to an A/B test than to an observational dataset, which is why genetic evidence carries more causal weight than transcriptomic evidence.</p> </blockquote> <p>Referring back to the 2× headline from earlier, the Nelson et al. (2015) result and the King et al. (2019) revised analysis both confirmed the population-average ~2× clinical-success multiplier for genetically supported drug programs. The same general 2× factor also enriches for labeled side effects, giving it predictive value for toxicology programs (Minikel &amp; Nelson, 2024; Carss et al., 2023). The multiplier is not constant across evidence types. King et al. found that when the causal gene is unambiguous - Mendelian traits, single-gene disorders with clear inheritance patterns, or GWAS associations driven by coding variants where the variant-to-gene mapping is direct - the approval probability multiplier rises above 2× and into the 3× range. The 2026 update from Minikel and Nelson refined this, showing that the multiplier scales with confidence in the causal gene assignment and is largely independent of genetic effect size, minor allele frequency, or year of discovery. In practice, a low-effect-size GWAS hit with confident causal-gene mapping is worth more than a high-effect-size hit at a locus where the variant-to-gene call is ambiguous. Trajanoska et al. (2023, <em>Nature</em>) place this in longer historical context: they identified 40 germline genetic observations that translated directly into approved therapies for 36 rare and 4 common conditions, with a median 25-year interval between target discovery and drug approval. The genetic-anchor strategy works, but it compounds slowly.</p> <picture> <source srcset="/assets/img/blog/target-discovery/APPH_FLYNN_UN08.webp" type="image/webp"/> <img src="/assets/img/blog/target-discovery/APPH_FLYNN_UN08.png" alt="Genetic-evidence confidence and target-disease clinical-success multiplier." loading="lazy"/> </picture> <p>When assembling a target’s evidence package, the <em>quality</em> of the genetic anchor matters more than the <em>quantity</em> of associated data. A coding variant with established protein consequences is worth more than a half-dozen non-coding GWAS signals at adjacent loci with ambiguous fine-mapping.</p> <h2 id="the-omics-stack-as-targetdisease-evidence">The omics stack as target–disease evidence</h2> <picture> <source srcset="/assets/img/blog/target-discovery/APPH_FLYNN_UN06.webp" type="image/webp"/> <img src="/assets/img/blog/target-discovery/APPH_FLYNN_UN06.png" alt="The omics evidence stack - genomics, transcriptomics, proteomics, metabolomics, epigenomics - each annotated with its characteristic failure mode." loading="lazy"/> </picture> <p>Target-disease linkage evidence comes from an omics stack, and each layer has a characteristic failure mode. Knowing when each layer can mislead you is the core methodological skill.</p> <h3 id="genomics-and-gwas">Genomics and GWAS</h3> <p>Genomics is usually the strongest layer for target-disease linkage. Human genetic variation provides a natural experiment: people who carry a loss-of-function variant in gene X are, on average, similar to people who would be pharmacologically treated with an inhibitor of gene X. If carriers of the variant have lower disease risk, that is a strong argument that inhibiting gene X would lower disease risk too. This is the Mendelian-randomization logic that gives genetic evidence special weight.</p> <p>Genome-wide association studies (GWAS) are the dominant method for finding these signals. GWAS identifies statistical associations between genetic variants and traits or diseases across large population cohorts. The method has identified thousands of risk loci across hundreds of diseases, now aggregated in resources like the NHGRI-EBI GWAS Catalog.</p> <blockquote> <p><strong>Concept Translation:</strong> GWAS is in some ways a giant feature-importance scan over the genome - for each of millions of variants, test whether allele frequency differs between cases and controls. The genome-wide significance threshold of <em>p</em> &lt; 5×10⁻⁸ is a Bonferroni-style correction for testing roughly a million effective independent variants. The catch is that statistical association picks the <em>neighborhood</em> the causal variant lives in, not the variant itself. Most GWAS hits sit in non-coding DNA - regulatory regions outside protein-coding sequence - which is why follow-up <em>fine-mapping</em> is needed to narrow the signal. Fine-mapping is a feature-attribution problem for a region of the genome: which variant in the correlated cluster is actually doing the work?</p> </blockquote> <p>The practical limitations matter. GWAS has to apply heavy multiple-testing correction (the classic <em>p</em> &lt; 5×10⁻⁸ genome-wide significance threshold), which means it misses real signals at smaller effect sizes, especially in underpowered cohorts. Most GWAS hits are in non-coding regions, so the protein-level mechanism is often implicit. Identifying the <em>causal gene</em> from a genome-wide-significant locus often requires additional work - fine-mapping (narrowing the locus to the most plausible causal variants), eQTL analysis (testing whether a variant shifts gene expression), and experimental follow-up.</p> <p>The populations represented in GWAS are also overwhelmingly European-ancestry, which creates a real equity problem when target discovery built on GWAS evidence is applied clinically to more diverse populations. The June 2021 GWAS Catalog inventory put European-ancestry participants at approximately 86.3% of all GWAS subjects, with East Asian at ~5.9% and African at ~1.1%. By 2023 the GWAS Catalog European share had only modestly declined, to around 86.5%, while the GWAS Diversity Monitor, which tracks the running participant counts more broadly, recorded European representation of ~94.5% of total participants by September 2024 (Mills &amp; Rahal, <em>Nat Genet</em>, 2019; Corpas et al., <em>Cell Genomics</em>, 2025). The two metrics differ because the GWAS Diversity Monitor weights large biobank cohorts heavily, but the conclusion is the same: target discovery built on this evidence base will systematically under-serve non-European populations unless deliberately corrected with fine-mapping in diverse cohorts and downstream validation across ancestries. This is a target-evidence problem as much as a deployment problem. A target whose causal gene is well mapped only in European cohorts may not support a defensible target-disease linkage claim for patients of other ancestries until cross-ancestry replication is done.</p> <blockquote> <p><strong>Concept Translation:</strong> This is the same dataset-bias problem that ML practitioners deal with constantly. A model trained on a non-representative dataset can be highly accurate on the in-distribution group while failing on out-of-distribution groups. The fix is the same in principle: collect more representative data, evaluate per-subgroup, and don’t ship a model whose validation doesn’t cover the deployment population.</p> </blockquote> <h3 id="transcriptomics">Transcriptomics</h3> <p>Transcriptomics measures RNA levels, sitting between genotype and protein as a sensitive readout of cellular state. Bulk and single-cell RNA sequencing, along with older microarray data, make up the largest public omics archives (GEO, ArrayExpress) and provide differential-expression evidence: which genes change in disease tissue versus healthy tissue?</p> <p>The main failure mode of transcriptomics for target discovery is simple. Many transcriptomic changes are <em>bystanders</em>. A gene that changes expression in disease is not necessarily causing the disease; it may be responding to the disease, or changing in parallel with it because of upstream regulation. A transcriptomic hit generates a hypothesis; it does not establish causal linkage. Evidence becomes much stronger when transcriptomics and genetics agree, and the gene is both differentially expressed <em>and</em> under genetic selection in disease.</p> <blockquote> <p><strong>Concept Translation:</strong> Bystander vs. causal in transcriptomics is the same trap as confounded correlation in observational ML. A feature that strongly correlates with the label can still be downstream of the label rather than upstream - predictive of it without driving it. RNA-level changes pick up both kinds and can’t separate them on their own. This is why GWAS-supported transcriptomic hits are weighted more heavily than transcriptomic hits in isolation.</p> </blockquote> <h3 id="proteomics-and-metabolomics">Proteomics and metabolomics</h3> <p>Proteomics measures protein abundance directly. Since most drugs target proteins, proteomic evidence sits closer to the mechanism-of-action question than transcriptomic evidence. The limitations are technical: mass-spec proteomics is expensive, less sensitive than RNA-seq, and struggles with membrane proteins, which include many of the cell-surface receptor and transporter classes that matter most in drug discovery.</p> <p>Metabolomics measures small-molecule metabolites. It often points to disease-associated metabolic changes that can be traced back to protein targets - which enzymes are producing or consuming these altered metabolites. Metabolomics is particularly useful for metabolic diseases and for target discovery in contexts where disease biology is about pathway flux more than signaling.</p> <h3 id="epigenomics">Epigenomics</h3> <p>Epigenomics (DNA methylation, histone modifications, chromatin accessibility) provides a reversible regulatory layer above the genome. Epigenetic enzymes are themselves drug targets (HDAC inhibitors, DNMT inhibitors, both with approved drugs), and epigenetic data can identify non-genetic mechanisms driving disease. The characteristic limitation is interpretability: it is often unclear what functional impact a given epigenetic change has on gene expression in the specific disease context.</p> <blockquote> <p><strong>Concept Translation:</strong> A useful mental model: if the genome is the source code, epigenomics is the runtime configuration. Methylation marks and chromatin states change which parts of the code get executed in which cell types, without altering the code itself. Same source, different behavior - and the configuration is reversible, which is why epigenetic enzymes are drug targets.</p> </blockquote> <h3 id="multi-omics-integration-and-its-limits">Multi-omics integration and its limits</h3> <p>Combining these layers sounds straightforward. In practice, it’s a difficult technical problem. Batch effects across studies, confounding by disease severity or patient demographics, differential dropout in single-cell data, and different noise models across modalities all complicate naive integration. The best-case design, when available, analyzes multi-omics data from <em>the same set of patients</em> rather than integrating across cohorts, so sample-level confounders are controlled. That is rarely possible at GWAS-cohort scales, where multi-omics profiling of every participant is prohibitively expensive. Target-discovery workflows that build evidence from public multi-omics resources need to state these limitations explicitly. Workflows that report “our multi-omics integration identified target X” without addressing them are not making a credible argument.</p> <p>A useful target-disease evidence package usually includes (a) genetic evidence where it exists, as the anchor, (b) transcriptomic and proteomic evidence in disease tissue that is consistent with the genetic signal, (c) tissue-specificity data (see [tissue-specificity evidence] → <em>Tissue specificity: the safety half of target selection</em><sup id="fnref:tissue"><a href="#fn:tissue" class="footnote" rel="footnote" role="doc-noteref">2</a></sup>), and (d) network or pathway context that situates the target in plausible biology. No single layer is enough - the layers have to agree.</p> <h2 id="networks-and-knowledge-graphs-evidence-beyond-single-gene-associations">Networks and knowledge graphs: evidence beyond single-gene associations</h2> <p>A gene-at-a-time view of the omics stack misses <em>mechanistic context</em>. A gene with moderate genetic support and moderate transcriptomic support may look like a weak target in isolation, yet sit at a critical node in a disease-implicated pathway. In that setting, network position is itself evidence. Target-discovery methodology increasingly builds evidence at the network level as well as the gene level.</p> <p>The method family is <strong>biological graphs and knowledge graphs</strong>. Protein-protein interaction networks, gene regulatory networks, gene co-expression networks, metabolic networks, and signaling networks each represent a different axis of biological relationship, and each carries evidence relevant to target linkage. Knowledge graphs integrate these with literature-derived associations, drug-target relationships, disease-gene associations, and related data, producing heterogeneous graphs with multiple types of nodes and relationships that can be queried for evidence patterns around a candidate target.</p> <blockquote> <p><strong>Concept Translation:</strong> A biomedical knowledge graph is a heterogeneous graph in the standard ML sense - multiple node types (genes, proteins, diseases, drugs, pathways) and multiple edge types (binds, expressed-in, treats, associated-with). The same graph-ML toolkit applies: node classification (is this gene a likely target?), link prediction (does this gene-disease edge exist but is undiscovered?), graph neural network embeddings, and metapath-based random walks. The biology adds structure but doesn’t change the modeling problem.</p> </blockquote> <p>Two network features are especially useful for target-linkage arguments. <strong>Hubs</strong> are highly connected nodes. They are enriched for essential genes and are often promising targets in diseases driven by loss of function, but they are also more likely to be toxic to modulate pharmacologically. <strong>Bridging nodes</strong> are high-betweenness nodes - nodes that sit on many shortest paths between modules. They are often more attractive targets because their inhibition tends to be non-lethal while still disrupting specific disease-relevant pathways.</p> <blockquote> <p><strong>Concept Translation:</strong> Hubs are nodes with high degree centrality (think high-PageRank pages); knock them out and a lot of the graph collapses, which is why they’re both attractive and dangerous. Bridging nodes are nodes with high betweenness centrality - they sit on many shortest paths, like a bottleneck router connecting subnets. Inhibiting a bridge can disconnect a specific module without crashing the whole network, which is roughly what a clean drug effect looks like.</p> </blockquote> <p>A knowledge graph over disease-gene-drug-pathway data also lets teams <em>aggregate and score</em> evidence across heterogeneous sources. The Open Targets Platform is the canonical public example: an open-source platform that integrates evidence from genetics (GWAS, ClinVar, ClinGen), somatic mutation data (Cancer Gene Census, COSMIC), known drug-target relationships, expression data, text mining, and animal models, producing a quantitative association score between each gene and each disease. As of the most recent platform release, the Open Targets evidence layer comprises over 27.8 million timestamped evidence assertions across the genes-and-diseases space.</p> <p>The platform’s scoring methodology matters in practice. The overall target–disease association score is computed as a <em>harmonic sum</em> of source-weighted evidence scores - a scheme in which each additional piece of similar evidence adds less than the previous one, rather than a raw sum or arithmetic mean. The harmonic-sum construction keeps heavily studied targets from accumulating inflated scores through redundant literature citations. Adding the 50th literature reference does not move the score the way the 1st reference does. Within data types that contain multiple sources, the scaling factor uses an inverse-squares construction so that source contributions saturate gracefully rather than letting any single high-throughput source dominate.</p> <blockquote> <p><strong>Concept Translation:</strong> The harmonic-sum aggregator is a regularizer against redundant evidence. Each additional similar source contributes less than the previous one, much like a saturating activation function. Without it, well-studied targets - those with thousands of papers - would dominate the rankings purely by literature volume, and the score would be a popularity ranking instead of an evidence ranking. This is the same problem that plagues citation-count metrics in academia and is solved with the same general technique: diminishing returns on duplicate signal.</p> </blockquote> <p>Source weights vary by evidence type. ClinGen clinical-validity curations at the “definitive” tier receive an absolute weight of 1.0, the strongest possible single-source evidence. Somatic mutation evidence from the Cancer Gene Census uses a tiered modifying scheme that adjusts the base score by mutational frequency and disease-specificity context: a mutation observed in only 1 sample in the dataset receives a −0.25 modifier (penalizing weak recurrence), while a mutation that occurs more frequently in a particular disease relative to others receives a +0.25 modifier (rewarding disease-specificity). The platform’s “Associations on the Fly” (AOTF) feature, introduced in recent releases, lets users dynamically reweight the evidence contributions to formulate custom therapeutic hypotheses without forking the underlying scoring schema.</p> <p>The consortium itself is worth noting for field-guide purposes. Open Targets is a pre-competitive public-private partnership - direct competitors contribute to a shared evidence resource before they diverge into proprietary drug programs. As of 2026 the active partner roster includes EMBL-EBI, the Wellcome Sanger Institute, Genentech (Roche Group), GSK, MSD (Merck &amp; Co.), Pfizer, Sanofi, and Bristol Myers Squibb (which joined in late 2022). Biogen and Takeda were earlier members who exited in 2020. The platform aggregates signals that no single company would assemble alone.</p> <p>Formalizing evidence aggregation as a graph problem also opens the door to graph machine learning. Target-disease association can be formulated as a link-prediction problem on a heterogeneous biomedical knowledge graph: given a graph where some gene-disease edges are known, predict which missing edges are real. Graph neural networks, network-embedding methods, and random-walk-based algorithms have all been applied to this problem. The [knowledge graphs and the rentosertib case study] → <em>Knowledge graphs, network medicine, and the first end-to-end AI-discovered drug</em> picks up this thread in detail.<sup id="fnref:kg"><a href="#fn:kg" class="footnote" rel="footnote" role="doc-noteref">3</a></sup></p> <h2 id="a-worked-example-senolytic-target-identification-in-idiopathic-pulmonary-fibrosis">A worked example: senolytic target identification in idiopathic pulmonary fibrosis</h2> <picture> <source srcset="/assets/img/blog/target-discovery/APPH_FLYNN_UN07.webp" type="image/webp"/> <img src="/assets/img/blog/target-discovery/APPH_FLYNN_UN07.png" alt="Four evidence sources - pathology, mechanism, computational prioritization, dasatinib convergence - converge on PTK2/FAK as a senolytic IPF candidate." loading="lazy"/> </picture> <p>To make the frameworks concrete, consider a target-identification workflow for a specific disease-hypothesis pair: <strong>cellular senescence as a target class in idiopathic pulmonary fibrosis (IPF)</strong>.</p> <p><strong>The mechanistic hypothesis.</strong> Cellular senescence is a permanent state of cell-cycle arrest with a secretory phenotype (SASP) that promotes inflammation and disrupts tissue homeostasis. Senescent cell accumulation is associated with many age-related diseases, and in IPF specifically, senescent cell programs have been documented in disease tissue. The working hypothesis: compounds that selectively kill senescent cells (<strong>senolytics</strong>) should slow IPF progression by removing the SASP-producing source of inflammation and fibrotic signaling. This is a <em>causative</em> hypothesis in the sense of the earlier framework - senescent cells are positioned as upstream drivers rather than symptoms.</p> <blockquote> <p><strong>Concept Translation:</strong> Senescent cells are sometimes called “zombie cells” - they’ve stopped dividing but won’t die, and they leak inflammatory signals (the SASP) into surrounding tissue. The senolytic strategy is to selectively kill them. From a target-discovery angle, the question is which proteins keep senescent cells alive that are <em>not</em> equally important for keeping healthy cells alive. Those proteins are the senolytic targets.</p> </blockquote> <p><strong>Evidence layer 1: tissue pathology.</strong> Senescent cells are identifiable in IPF lung tissue by senescence-associated markers; their burden correlates with disease severity. This is differential-expression-style evidence - the disease tissue differs from healthy tissue in a specific, characterizable way. It is necessary but insufficient on its own. Senescence could still be a bystander response to lung injury rather than a driver of ongoing fibrosis.</p> <p><strong>Evidence layer 2: mechanism.</strong> Senescent cells are apoptosis-resistant by design; that is part of what makes them persistent. The resistance depends on anti-apoptotic pathways collectively called <strong>SCAPs</strong> (senescent cell anti-apoptotic pathways). That yields a specific target-class hypothesis: <em>proteins involved in SCAPs are candidate senolytic targets</em>, because inhibiting them should remove the apoptosis resistance and allow senescent cells to die. This narrows the search from “any gene differentially expressed in IPF” to “genes in a specific functional class whose inhibition would have a mechanistically predicted effect.”</p> <p><strong>Evidence layer 3: computational target prioritization.</strong> A knowledge-graph-based target discovery platform can operationalize the combined evidence. The workflow:</p> <ol> <li>Start from a disease-specific dataset (IPF patient tissue vs healthy controls) and score candidate targets by disease-association strength.</li> <li>Apply a <strong>gene-set filter</strong> restricting candidates to pre-specified functional classes (apoptosis-related pathways, SCAPs).</li> <li>Apply a <strong>small-molecule druggability filter</strong>. This asks whether a target can plausibly be modulated by a conventional small-molecule drug. For senolytic drug development, the delivered modality is small molecule, so candidates that are only tractable by antibody or biologic are de-prioritized.</li> <li>Apply an <strong>expression filter</strong> that focuses on protein classes suitable as drug targets (e.g., kinases, enzymes).</li> <li>Apply an optional <strong>novelty filter</strong> that preferentially surfaces targets with fewer published references, for teams pursuing first-in-class programs against new mechanisms rather than fast-follower strategies against established mechanisms.</li> </ol> <p><strong>The output, checked against known biology.</strong> A representative output of this workflow is <strong>PTK2 (also known as FAK, focal adhesion kinase)</strong> as a high-ranked senolytic candidate. PTK2 is a tyrosine kinase, druggable by small molecules, and expressed in relevant contexts. The biological cross-check comes from independent experimental literature. <strong>Dasatinib</strong>, a multi-kinase inhibitor originally approved for chronic myeloid leukemia, has been identified as a senolytic in human dermal fibroblasts, and its senolytic activity has been attributed in part to inhibition of PTK2/FAK among other kinases. This independent wet-lab evidence predates the computational prioritization and converges on the same target class.</p> <p><strong>What this example illustrates.</strong> The evidence package for PTK2/FAK as a senolytic target is an aggregation of four sources: disease-tissue pathology (senescent cell burden in IPF), a mechanistic hypothesis (SCAP dependence), computational target prioritization weighted by disease association, pathway membership, and druggability, and convergent experimental evidence from a repurposed clinical-stage compound. No individual layer would be sufficient; together they make a credible argument.</p> <p>The claim is narrower than it might sound. PTK2 remains a candidate rather than a validated senolytic target in IPF patients; validation would require dosing studies, biomarker responses, and clinical evidence. On the evidence presented here, PTK2 clears the 5R “right target” and GOT-IT “disease linkage” bars for advancing into target validation. Moving from candidate to validated target requires [druggability assessment] → <em>Druggability, ligandability, and modality choice in the AlphaFold 3 era</em><sup id="fnref:drug"><a href="#fn:drug" class="footnote" rel="footnote" role="doc-noteref">4</a></sup>, and the downstream posts in this series.</p> <h2 id="open-questions-in-targetdisease-evidence-work">Open questions in target–disease evidence work</h2> <p><strong>The benchmarking problem.</strong> Target-discovery ML is unusually hard to benchmark. The ground-truth signal - did this target hold up in Phase III? - takes a decade and nine figures per data point. Surrogate benchmarks exist (does the method rediscover known targets for known diseases when given only pre-discovery literature? does it recover genetic signals already in Open Targets? does it predict held-out drug-target relationships?) but all share the same weakness: the evaluation corpus overlaps with the corpus the ML methods were trained on. Truly unbiased benchmarking for target-discovery ML, where the test set is future-clinical rather than past-published, remains a live methodological problem.</p> <p><strong>The ignorome problem.</strong> Research attention across the ~20,000 human protein-coding genes is highly skewed: a small fraction of genes accounts for most of the literature, the annotations, the available assays, and the training data for ML methods. A target-discovery method that only surfaces candidates adjacent to already-well-studied proteins is solving a weaker version of the problem than the one the industry actually needs. The [novelty vs repurposing] → <em>Novelty vs repurposing: when to invent a target and when to reuse one</em> post picks up this thread.<sup id="fnref:nov"><a href="#fn:nov" class="footnote" rel="footnote" role="doc-noteref">5</a></sup></p> <p><strong>The causality problem.</strong> Most of the evidence stack is associational. Genetic evidence - especially Mendelian-randomization-style evidence - carries more causal weight because it approaches causality in a way other omics layers do not. Methods that can extract more causal structure from non-genetic data (perturbation screens, single-cell perturbation readouts like Perturb-seq, large-scale CRISPR screens) are one of the most active areas of methodological development, and they feed directly into target-disease evidence work. The [synthetic lethality] → <em>Synthetic lethality and combination targets: ML methods for finding drug pairs that work together</em> post gets into CRISPR-screen analysis in depth.<sup id="fnref:synleth"><a href="#fn:synleth" class="footnote" rel="footnote" role="doc-noteref">6</a></sup></p> <p>Evidence frameworks for target-disease linkage separate disciplined target discovery from assertion. Teams that internalize them are less likely to skip steps, and skipped steps are rarely recoverable later. The framework languages (5R, GOT-IT) and the evidence-stack discipline - genomics as anchor, corroborated by transcriptomics and network context, challenged by alternative hypotheses - are what distinguish a defensible target list from one that merely sounds plausible.</p> <hr/> <h2 id="further-reading">Further reading</h2> <p><strong>Foundational framework papers</strong></p> <ul> <li> <p>Cook, D., Brown, D., Alexander, R., March, R., Morgan, P., Satterthwaite, G., &amp; Pangalos, M. N. (2014). Lessons learned from the fate of AstraZeneca’s drug pipeline: a five-dimensional framework. <em>Nature Reviews Drug Discovery</em>, 13(6), 419–431. The canonical 5R paper, published from inside AstraZeneca’s R&amp;D leadership at the moment the framework was institutionalized.</p> </li> <li> <p>Morgan, P., Brown, D. G., Lennard, S., Anantharaman, M. J., Gilbert, J., Cook, D., et al. (2018). Impact of a five-dimensional framework on R&amp;D productivity at AstraZeneca. <em>Drug Discovery Today</em>, 23(8), 1395–1399. The quantitative follow-up documenting an approximately five-fold improvement in small-molecule project survival from candidate-selection to Phase III decision after the 5R was institutionalized - the empirical basis for the framework’s enduring methodological status.</p> </li> <li> <p>Emmerich, C. H., Gamboa, L. M., Hofmann, M. C. J., Bonin-Andresen, M., Arbach, O., Schendel, P., et al. (2021). Improving target assessment in biomedical research: the GOT-IT recommendations. <em>Nature Reviews Drug Discovery</em>, 20(1), 64–81. The definitive methodological reference for the four-block target-assessment framework (disease linkage, target-related safety, strategic issues, technical feasibility) that this post’s structure mirrors.</p> </li> </ul> <p><strong>Genetic evidence and the success multiplier</strong></p> <ul> <li> <p>Nelson, M. R., Tipney, H., Painter, J. L., Shen, J., Nicoletti, P., Shen, Y., et al. (2015). The support of human genetic evidence for approved drug indications. <em>Nature Genetics</em>, 47(8), 856–860. The original analysis establishing that drug programs with human genetic support are roughly twice as likely to succeed in clinical development. This is the source for the 2× headline that anchors most subsequent literature.</p> </li> <li> <p>King, E. A., Davis, J. W., &amp; Degner, J. F. (2019). Are drug targets with genetic support twice as likely to be approved? Revised estimates of the impact of genetic support for drug mechanisms on the probability of drug approval. <em>PLOS Genetics</em>, 15(12), e1008489. The revised estimate using updated pipeline data, confirming the ~2× general result but showing that high-confidence causal evidence (Mendelian traits, coding-variant GWAS hits) yields a larger &gt;2× multiplier - the source for the 2× / 3× distinction in the post’s hook.</p> </li> <li> <p>Trajanoska, K., Bhérer, C., Taliun, D., Zhou, S., Richards, J. B., &amp; Mooser, V. (2023). From target discovery to clinical drug development with human genetics. <em>Nature</em>, 620(7975), 737–745. Identified 40 germline genetic observations that translated directly into approved therapies for 36 rare and 4 common conditions, with a median 25-year interval between target discovery and drug approval - the long-horizon companion to the multiplier finding.</p> </li> <li> <p>Minikel, E. V., Painter, J. L., Dong, C. C., &amp; Nelson, M. R. (2024). Refining the impact of genetic evidence on clinical success. <em>Nature</em>, 629, 624–629. The 2024 update showing that the genetic-support multiplier scales with confidence in the causal-gene assignment but is largely independent of effect size, allele frequency, or year of discovery.</p> </li> <li> <p>Ochoa, D., Karim, M., Ghoussaini, M., Hulcoop, D. G., McDonagh, E. M., &amp; Dunham, I. (2022). Human genetics evidence supports two-thirds of the 2021 FDA-approved drugs. <em>Nature Reviews Drug Discovery</em>, 21(8), 551. The source for the “two-thirds of recent FDA approvals are underpinned by human genetic evidence” statistic in the introduction.</p> </li> <li> <p>Carss, K. J., Deaton, A. M., Del Rio-Espinola, A., Diogo, D., Fielden, M., Kulkarni, D. A., et al. (2023). Using human genetics to improve safety assessment of therapeutics. <em>Nature Reviews Drug Discovery</em>, 22(2), 145–162. Companion reference for the side-effect-prediction half of the genetic-evidence story.</p> </li> </ul> <p><strong>GWAS diversity</strong></p> <ul> <li> <p>Mills, M. C., &amp; Rahal, C. (2019). A scientometric review of genome-wide association studies. <em>Communications Biology</em>, 2, 9. Foundational quantitative analysis of European-ancestry overrepresentation in the GWAS literature; ongoing version available at the GWAS Diversity Monitor (gwasdiversitymonitor.com).</p> </li> <li> <p>Corpas, M., Pius, M., Poburennaya, M., Guio, H., Dwek, M., Nagaraj, S., et al. (2025). Bridging genomics’ greatest challenge: the diversity gap. <em>Cell Genomics</em>, 5(1), 100724. The most recent quantitative update on GWAS diversity, with the September 2024 GWAS Diversity Monitor numbers (~94.5% European participants) and the 2023 GWAS Catalog comparison (~86.5%) used in this post.</p> </li> </ul> <p><strong>Open Targets</strong></p> <ul> <li> <p>Open Targets Platform Documentation: target–disease associations and target–disease evidence. platform-docs.opentargets.org. The canonical reference for the harmonic-sum scoring construction, the source-weight schema (including the ClinGen 1.0 absolute weight and the Cancer Gene Census ±0.25 modifiers), and the Associations on the Fly (AOTF) feature. The release-notes page is the right link because evidence-category weighting changes across releases - cite release notes rather than a static platform URL so readers land on the version-of-record for whatever’s current at your publish date.</p> </li> <li> <p>Open Targets news and partner roster: opentargets.org/partners and opentargets.org/news. For the partner-consortium roster cited in the body (EMBL-EBI, Wellcome Sanger, Genentech, GSK, MSD, Pfizer, Sanofi, BMS) and the historical exits (Biogen and Takeda, both 2020).</p> </li> </ul> <p><strong>Anti-amyloid antibodies in Alzheimer’s (for the causative-vs-symptomatic example)</strong></p> <ul> <li> <p>van Dyck, C. H., Swanson, C. J., Aisen, P., Bateman, R. J., Chen, C., Gee, M., et al. (2023). Lecanemab in early Alzheimer’s disease. <em>New England Journal of Medicine</em>, 388(1), 9–21. The pivotal Phase III result for lecanemab; the source for the “statistically significant slowing of cognitive decline” claim with the precise CDR-SB effect size.</p> </li> <li> <p>Sims, J. R., Zimmer, J. A., Evans, C. D., Lu, M., Ardayfio, P., Sparks, J., et al. (2023). Donanemab in early symptomatic Alzheimer disease: the TRAILBLAZER-ALZ 2 randomized clinical trial. <em>JAMA</em>, 330(6), 512–527. The pivotal Phase III result for donanemab. The post deliberately frames both drugs as test cases of the amyloid hypothesis rather than as resolved validation of it.</p> </li> </ul> <hr/> <div class="footnotes" role="doc-endnotes"> <ol> <li id="fn:amyloid"> <p>For readers who want to follow up, recent controversies around the amyloid hypothesis and the fallout for Alzheimer’s research as a whole are worth a separate dive. We don’t cover the controversy or its causes here. <a href="#fnref:amyloid" class="reversefootnote" role="doc-backlink">&#8617;</a></p> </li> <li id="fn:tissue"> <p>Coming in the next post - how GTEx, Human Protein Atlas, and scRNA-seq sharpen target safety decisions. <a href="#fnref:tissue" class="reversefootnote" role="doc-backlink">&#8617;</a></p> </li> <li id="fn:kg"> <p>Coming in a later post - how knowledge graphs get used in end-to-end target-discovery case studies. <a href="#fnref:kg" class="reversefootnote" role="doc-backlink">&#8617;</a></p> </li> <li id="fn:drug"> <p>Coming in the next post - how target candidates get screened for tractability and modality fit. <a href="#fnref:drug" class="reversefootnote" role="doc-backlink">&#8617;</a></p> </li> <li id="fn:nov"> <p>Coming in a later post - how target novelty gets quantified without over-penalizing under-studied biology. <a href="#fnref:nov" class="reversefootnote" role="doc-backlink">&#8617;</a></p> </li> <li id="fn:synleth"> <p>Coming in a later post - how CRISPR-screen evidence extends from single targets to target pairs. <a href="#fnref:synleth" class="reversefootnote" role="doc-backlink">&#8617;</a></p> </li> </ol> </div>]]></content><author><name>Noah Flynn</name><email>noahrflynn@gmail.com</email></author><category term="target discovery"/><category term="machine learning"/><category term="drug discovery"/><category term="target discovery"/><category term="genomics"/><summary type="html"><![CDATA[Target-disease association evidence is what Phase II efficacy hinges on. A practitioner's guide to 5R, GOT-IT, the omics stack, and knowledge graphs.]]></summary></entry><entry><title type="html">Drug target discovery: the front-of-funnel decision behind most Phase II failures</title><link href="https://noahrflynn.com/blog/2026/drug-target-discovery-phase-ii-failures/" rel="alternate" type="text/html" title="Drug target discovery: the front-of-funnel decision behind most Phase II failures"/><published>2026-05-08T16:00:00+00:00</published><updated>2026-05-08T16:00:00+00:00</updated><id>https://noahrflynn.com/blog/2026/drug-target-discovery-phase-ii-failures</id><content type="html" xml:base="https://noahrflynn.com/blog/2026/drug-target-discovery-phase-ii-failures/"><![CDATA[<p>A drug candidate can enter Phase II clinical trials with a nine-figure budget behind it, years of preclinical work completed, and an IND (Investigational New Drug application) cleared by the FDA, yet still carry the wrong answer to a question asked five to ten years earlier: <em>is this the right biological target?</em> For readers new to the pipeline, Phase I is mainly about safety and dosing. Phase II is where efficacy first gets a serious test in patients. Phase III is the larger confirmatory stage. Roughly 72% of drugs entering Phase II do not transition to Phase III, and approximately 90% of investigational drugs fail somewhere in clinical development overall (source: GlobalData clinical analytics, 2024). Lack of efficacy against the intended disease accounts for roughly half of all clinical trial failures; another ~30% are halted for unmanageable safety and toxicity findings. In many cases, the problem starts before dosing or formulation matter. The target was wrong, or the evidence for its role in the disease was too weak. Target discovery is one of the most consequential stages of the drug development pipeline, and this series intends to teach why.</p> <h2 id="why-drug-target-discovery-matters-in-2026">Why drug target discovery matters in 2026</h2> <p>When it comes to understanding chemistry and biology, the models are getting better. Generative models can design and optimize novel small molecules with greater success. AlphaFold 3 and the open-weights field around it, meaning models released with their parameters, have made routine structure prediction a solved problem for many classes of proteins. Docking, virtual screening, and ADMET prediction (absorption, distribution, metabolism, excretion, and toxicity) continue to mature as well.</p> <p>Yet clinical success rates have worsened. Industry-wide likelihood of approval (LOA) for a Phase I compound fell from about 10.4% for the 2014 single-year cohort to around 6.7% (average LOA for 10-year period from 2014-2023). Phase II remains the dominant attrition step, with only about 28% of drugs that enter Phase II making it to Phase III (phase-by-phase transitions: Phase I = 47%, Phase II = 28%, Phase III = 55%, Filing-to-approval = 92%).</p> <p>To assist instruction, front-of-funnel ML now has a concrete clinical example. In June 2025, <em>Nature Medicine</em> published Phase IIa results for rentosertib, a TNIK inhibitor for idiopathic pulmonary fibrosis (IPF), whose target (identified from multi-omics analysis of IPF patient tissue, meaning joint analysis of several molecular data types) and molecule (generative chemistry) were both AI-derived end-to-end. The study was small and short: 71 patients over 12 weeks (it was hypothesis-generating rather than a study designed to support approval) with reported improvements in standard lung-function measures relative to a placebo. Still, it marked the first time a fully AI-discovered drug, target included, produced a direction-of-effect efficacy signal in humans. Industry trackers now count more than 173 AI-discovered drug programs in clinical testing as of early 2026.</p> <p>This post lays out the series: what target discovery is, why it matters, what makes it hard, and where ML helps. Each subsequent post will take one axis of target discovery and examine it in detail:</p> <ul> <li><em>How to tell a drug target matters using evidence frameworks</em></li> <li><em>Druggability, ligandability, and modality choice in the AlphaFold 3 era</em></li> <li><em>Tissue specificity: the safety half of target selection</em></li> <li><em>The most crowded and abandoned therapeutic areas</em></li> <li><em>Novelty vs repurposing: when to invent a target and when to reuse one</em></li> <li><em>Synthetic lethality and combination targets: ML methods for finding drug pairs that work together</em></li> <li><em>Knowledge graphs and case studies in AI-driven target discovery</em></li> <li><em>Virtual cells for target discovery, perturbation models, and benchmarks</em></li> </ul> <h2 id="what-a-drug-target-is">What a drug target is</h2> <p>A drug target is a biomolecule, usually a protein and sometimes a nucleic acid, whose activity we want to modulate to produce a therapeutic effect. “Modulate” covers a range: inhibit an overactive enzyme, block a receptor from binding its ligand, degrade a disease-driving protein, replace a missing transcript. Binding the target has to produce the therapeutic effect. A molecule that binds to the target without changing disease outcomes is hitting a decoy.</p> <p>Nearly half of oral drugs on the market target enzymes, with kinases as the single most productive sub-family. About a third target cell-surface receptors, especially G-protein-coupled receptors. Ion channels, transporters, and nuclear hormone receptors round out the major protein classes. Nucleic acids form a smaller but fast-growing non-protein class. If you’ve read chapter 1 of <em>Machine Learning for Drug Discovery</em>, we derive a ~10⁵ protein “biological search space” covering on the order of 100,000 potential human protein targets once splice variants and post-translational modifications, that is, alternative versions of a protein and chemical changes added after it is made, are counted. A drug-like molecule has to find, bind, and modulate the right one.</p> <p>Target discovery, the front-of-funnel activity this series focuses on, is the work of deciding <em>which</em> of those proteins to go after for a given disease. It splits into two phases that casual usage often folds together:</p> <ul> <li><strong>Target identification</strong>: generating candidate target-disease links. “Here are ten proteins that seem to be involved in pancreatic fibrosis; let’s prioritize the top three.”</li> <li><strong>Target validation</strong>: building the evidence that one of those candidates is actually causal. Does knocking it out in a disease model reverse the phenotype? Does a human genetic signal support it? Is there pharmacological precedent?</li> </ul> <p>Both phases happen before a single compound is screened. Get either wrong and the downstream pipeline is solving the wrong problem.</p> <h2 id="funnel-economics-the-cost-of-being-wrong-at-target-selection">Funnel Economics: The Cost of Being Wrong at Target Selection</h2> <picture> <source srcset="/assets/img/blog/target-discovery/APPH_FLYNN_UN01.webp" type="image/webp"/> <img src="/assets/img/blog/target-discovery/APPH_FLYNN_UN01.png" alt="Drug development funnel showing six pipeline stages with approximate duration and cost; target identification is the cheapest stage." loading="lazy"/> </picture> <p>The drug-development funnel runs roughly like this:</p> <ul> <li>Disease hypothesis and early target identification: years to decades, often academic.</li> <li>Target validation: up to ~2.5 years, averaging around $353M when done end-to-end including follow-on studies.</li> <li>Lead discovery and optimization: up to ~2 years, over half a billion dollars.</li> <li>Preclinical development: about a year, ~$340M.</li> <li>Clinical development: Phase I about 1.5 years, Phases II/III about 2.5 years combined, roughly half a billion dollars in total.</li> <li>Regulatory approval: ~1.5 years, ~$3M, the cheapest transition once you’re there.</li> </ul> <p>Aggregate “cost to develop a new drug” figures vary widely because analysts treat time costs, failure allocation, and capitalization differently. The Tufts Center for the Study of Drug Development’s widely cited $2.8 billion mean per-approved-drug cost includes roughly $1.16 billion of foregone-investor returns over the development window plus $312 million in post-approval R&amp;D. More recent cost estimates from RAND (2025), JAMA (Wouters et al., 2020), and Deloitte (2024) report different distributions, with medians closer to $0.7 billion to $1.0 billion and means weighted by high-cost outliers in the $0.95 billion to $2.23 billion range. Though reported headline figures depend heavily on which costs are included, the point is consistent. This is an expensive process, and being wrong costs a lot of cash!</p> <p>That makes the target-selection decision unusually consequential. It spends comparatively less money, yet determines whether everything underneath it is wasted. Successful or unsuccessful completion of Phase 2 for an individual drug costs pharmaceutical companies around 20% of the sum spent on the drug discovery pipeline recalculated for each individual drug. Phase III is a larger absolute commitment, but Phase II is where we first get an answer to “is this target right?” and where many disappointments occur.</p> <p>In a 2024 report, GlobalData attributes roughly half of clinical trial failures across phases to lack of efficacy and another ~30% to safety or toxicity. If efficacy collapses because the chosen target never had enough evidence behind it, then better target selection compounds downstream. Prevent one Phase II failure and the savings can exceed what you spent on target-assessment work by an order of magnitude. Catch the same mistake at target selection instead of in Phase II and the savings are larger still.</p> <picture> <source srcset="/assets/img/blog/target-discovery/APPH_FLYNN_UN02.webp" type="image/webp"/> <img src="/assets/img/blog/target-discovery/APPH_FLYNN_UN02.png" alt="Phase-transition and likelihood-of-approval chart for clinical drug development." loading="lazy"/> </picture> <p>The 2014–2023 industry data also points to a second problem. Phase I likelihood of approval fell from 10.4% for the 2014 cohort to 6.7% for the most recent ten-year cohort, partly driven by an industry shift toward first-in-class, riskier targets where the evidence base is thinner. Many easier targets have already been drugged. What’s left is harder, and harder problems put more weight on front-of-funnel methods. Biomarker-stratified programs, which enroll patient subgroups defined by a measurable biomarker, run at roughly double the LOA of unstratified ones. This result follows from stronger target-selection evidence, as the biomarker identifies the patient subgroup in whom the target matters. For this reason, target and biomarker development often move together.</p> <h2 id="why-drug-target-discovery-is-hard">Why drug target discovery is hard</h2> <picture> <source srcset="/assets/img/blog/target-discovery/APPH_FLYNN_UN03.webp" type="image/webp"/> <img src="/assets/img/blog/target-discovery/APPH_FLYNN_UN03.png" alt="Five target-assessment criteria - linkage, safety, commercial, feasibility, data - arranged in a pentagon around a central go/no-go decision." loading="lazy"/> </picture> <p>There is no single solution or approach as every target hunting strategy is context dependent. Useful target-assessment work has to satisfy at least five criteria simultaneously.</p> <h3 id="targetdisease-linkage">Target–disease linkage</h3> <p>Does the target play a causal role in the disease rather than merely correlate with it? Genetic evidence carries the most weight in most organizations. That includes GWAS hits, population-scale studies that link genomic loci to disease, Mendelian randomization, which uses inherited variants as a quasi-natural experiment, and rare-variant burden tests, which ask whether damaging variants in a gene accumulate more often in cases than controls. Multi-omics integration, which combines several molecular data layers such as RNA, protein, and chromatin measurements, acts as supporting evidence.</p> <h3 id="target-related-safety">Target-related safety</h3> <p>Can humans tolerate modulation of this protein for years? Evolutionary conservation predicts essentiality and toxicity, wwhereas tissue specificity predicts off-target effects. Knockout phenotypes also provide a useful preview into safety implications if we were to neutralize the target.</p> <h3 id="commercial-strategy">Commercial strategy</h3> <p>Is there a path to reimbursement? Is this a first-in-class program, meaning a new mechanism, a best-in-class program, meaning a superior drug against an established mechanism, or a repurposing play, meaning an existing drug or target used in a new disease? These questions inform which target to prioritize even when several are scientifically equivalent.</p> <h3 id="technical-feasibility">Technical feasibility</h3> <p>Is the target druggable, meaning some therapeutic class can plausibly bind or modulate it? Can the organization develop the right modality, such as a small molecule, antibody, or oligonucleotide? A protein-protein interaction with no obvious pocket may be validated yet undruggable.</p> <h3 id="data-quality">Data quality</h3> <p>Is the evidence base solid, or built on a few underpowered studies that never replicated? Manual curation, field-standard evaluation, and in some cases meta-analyses feed into this.</p> <p>Taken together, a “good target” satisfies a multi-objective optimization problem across all five properties, involving a variety of data modalities. In practice, these criteria produce <em>weighted votes</em> rather than verdicts. Few targets get a clean yes across all five. More often, we see mixed signals and have to decide whether one target candidate’s profile beats the other eight candidate targets on our list. Target-discovery ML tries to improve this decision-making process.</p> <h2 id="where-ml-helps-in-drug-target-discovery">Where ML helps in drug target discovery</h2> <p>The chemistry-side ML playbook of generating candidates, predicting properties, and optimizing molecules does not transfer cleanly to target discovery. There is not an explicity accuracy metric to optimize against, training data is fragmented across a dozen modalities, and ground truth, i.e., “was this the right target?”, takes a decade and often a Phase III trial to establish. Despite these limitations, ML does three things well here.</p> <h3 id="scale">Scale</h3> <p>Datasets for target discovery are enormous and heterogeneous:</p> <ul> <li>GenBank contains over 300 billion base pairs of sequencing data.</li> <li>UniProt has more than 180 million protein sequences.</li> <li>The ENCODE multi-omics database exceeds one petabyte.</li> <li>PubMed has more than 35 million citations</li> <li>The USPTO database has hundreds of thousands of patents</li> <li>ClinicalTrials.gov has more than 438,000 clinical studies</li> </ul> <p>No human team can read or integrate all of this. ML can convert it into searchable, queryable representations. Knowledge graphs built from biomedical text, which turn papers into linked entities and relationships, compressed numerical representations of multi-omics datasets, and network-proximity algorithms that rank genes by their topological distance from disease-associated nodes are all methods that scale to the size of the actual evidence base.</p> <h3 id="cryptic-patterns">Cryptic patterns</h3> <p>A well-known example comes from SARS-CoV-2 repurposing, where we consider existing drugs that might work in a new or unaddressed disease. One study used network-diffusion and network-proximity algorithms on a combined dataset of the human interactome, the network of protein-protein interactions in cells, viral targets, and drug interaction data to rank 6,340 drugs for expected efficacy. The top-ranked drugs, tested experimentally, showed a 62% success rate in reducing viral infection, and 76 of the 77 drugs that worked did <em>not</em> bind to proteins directly targeted by the virus, suggesting network-based mechanisms rather than direct target engagement. That kind of signal, which requires multiple hops of complex reasoning, is hard to recover by manual review. In this case, graph algorithms operating across the interactome were able to surface it.</p> <h3 id="automation">Automation</h3> <p>Literature review for a new target can take days or months. Named-entity recognition, which identifies mentions of genes, diseases, drugs, authors, and institutions in text, can be automated with usable quality (state-of-the-art NER reaches roughly 90% F-score). Relation extraction, which asks whether a paper asserts a specific link between those entities, is harder, around 50%. A knowledge graph that updates as new papers are published treats the evidence base as a living object, instead of freezing it at the last manual review.</p> <p>These are front-of-funnel tasks that ML handles well. Final prioritization is a different matter*. In practice, target selection remains heavily human-curated. ML produces ranked candidate lists, aggregated evidence dashboards, and network-scored hypotheses, but the decision to allocate a program’s budget to target #3 rather than #7 still depends on organizational expertise, portfolio balance, competitive intelligence, and the intellectual-property landscape. Those factors are usually absent from training data. For now, target-discovery ML compresses the evidence into a form a human committee can reason about.</p> <p>*If you’re coming from a background in recommender systems, this is similar to a cascading systems approach where we care about maximizing recall at the start (i.e., we don’t want to preemptively filter out a candidate and cut a potential multibillion dollar revenue stream) and, by the end, are most trying to maximize precision (i.e., avoid a costly false positive that damages our corporate brand, wallet, and sanity).</p> <h2 id="worked-example-rentosertib">Worked Example: Rentosertib</h2> <p>Insilico Medicine’s rentosertib (also known as ISM001-055 / INS018_055) for idiopathic pulmonary fibrosis (IPF) serves as an instructional walkthrough (though I do not go so far as to categorize it as a gold-standard benchmark) of one end-to-end AI-driven target-and-molecule program.</p> <p>Target identification began with multi-omics analysis of lung tissue from IPF patients versus healthy controls, combined with text mining across IPF literature and knowledge-graph reasoning over protein-protein interactions and pathway data. TNIK, a TRAF2- and NCK-interacting kinase, emerged as a novel regulator of fibrotic pathways. I.e., TNIK was not an established IPF target when the program started. The timeline is also notable. Project initiation to preclinical candidate nomination, the point where a program chooses the molecule it will advance into formal preclinical testing, took roughly 12–18 months, and the asset reached human clinical trials in under 30 months total, against an industry average of 4.5–6 years for early-stage discovery. Target validation ran in parallel with early chemistry; preclinical models confirmed that TNIK inhibition reduced fibrotic phenotypes, and generative-chemistry tools designed a series of inhibitors. The knowledge-graph workflow behind this will be the subject of a future article in this series.</p> <p>Phase I established safety and pharmacokinetics, meaning how the body absorbs, distributes, and clears the drug. The Phase IIa trial (GENESIS-IPF, NCT05938920) enrolled 71 IPF patients across 21 sites in China over 12 weeks. The secondary endpoint, change in forced vital capacity (a standard measure of lung function) from baseline at week 12, showed +98.4 mL for the 60 mg daily dose versus −20.3 mL for placebo. <em>Nature Medicine</em> published the results in 2025, where they were positioned as the first clinical proof-of-concept for an end-to-end AI-discovered drug.</p> <p>Keeping limitations in view, seventy-one patients over 12 weeks is a hypothesis-generating dataset, not a study sized or designed to support approval. No pivotal trial is running as of April 2026; Insilico is in regulatory discussions about a Phase IIb pivotal study, and a separate US Phase IIa (NCT05975983) is enrolling, with eight of the planned 60 patients having completed the 12-week treatment as of mid-2025. Keeping within the subject matter of this article series, TNIK had to be identified as a target before any chemistry could start. That identification came from a workflow combining multi-omics data, literature mining, and knowledge-graph reasoning. The Phase IIa result matters because it is consistent with the target call having been right. Whether the methodology generalizes is a question the next decade of readouts will answer.</p> <p>A useful counterweight is Recursion’s REC-994, a lead pre-merger AI-discovered candidate for cerebral cavernous malformation, which was discontinued in May 2025 after long-term Phase II data failed to confirm earlier efficacy trends. High-profile AI-guided programs can fail in the clinic for the same reason other programs do. Through curating better evidence, we hope to lower the failure rate.</p> <h2 id="who-does-drug-target-discovery">Who does drug target discovery</h2> <p>For readers whose mental model of drug discovery centers on Big Pharma, it’s motivating to keep in mind that, in the United States, nearly 60% of newly approved drugs were discovered in universities or biotechnology companies, not by Big Pharma’s own R&amp;D. Small, often academic-adjacent biotechs take the early-stage innovation risk. Big Pharma licenses, acquires, or partners to bring the late-stage program through approval and marketing.</p> <p>The reasons are structural. Big Pharma faces the “better than the Beatles” problem (the bar for new drug approval keeps rising because better drugs already exist), the “low-hanging fruit” problem (the easy targets are mostly drugged), the “cautious regulator” problem (FDA standards ratchet up after each safety scare and rarely relax), and a tendency to industrialize the wrong activities; scaling basic research and brute-force screening has not improved clinical success rates in aggregate. For more information, we discuss these problems and the related Eroom’s Law in detail within chapter 1 of “Machine Learning for Drug Discovery.”</p> <p>For a practitioner in 2026, this means target-discovery tooling is disproportionately built inside smaller, often AI-native companies, licensed into Big Pharma programs, or run academically against public data. Big Pharma’s target-discovery groups increasingly act as evaluators and integrators rather than primary generators.</p> <h2 id="the-rest-of-the-drug-target-discovery-series">The rest of the drug target discovery series</h2> <p>The rest of this series goes deep on what this pillar surveys:</p> <ul> <li>Evidence frameworks → <em>How to tell a drug target matters</em> Driver vs passenger mutations, oncogene addiction (tumors becoming unusually dependent on one gene), multi-omics integration, GWAS, and practical target-assessment frameworks such as AstraZeneca’s 5R and GOT-IT.</li> <li>Druggability in the AlphaFold 3 era → <em>Druggability, ligandability, and modality choice in the AlphaFold 3 era</em> Classical druggability, the expanding druggable-genome concept, therapeutic modalities such as antisense oligonucleotides (ASOs), PROTAC degraders, and antibody-drug conjugates (ADCs), and what AlphaFold 3 changed about structure-based drug design.</li> <li>Tissue specificity → <em>Tissue specificity: the safety half of target selection</em> Single-cell RNA-seq and reference atlases such as GTEx and the Human Protein Atlas, and how tissue-level gene-expression data feeds into safety prediction.</li> <li>Likelihood of approval by therapeutic area → <em>The most crowded and abandoned therapeutic areas</em> Likelihood of approval by therapeutic area, the economics of repurposing, and why some disease areas are systematically more tractable.</li> <li>Novelty vs repurposing → <em>Novelty vs repurposing: when to invent a target and when to reuse one</em> When to invent a new target and when to reuse one. The Illuminating the Druggable Genome program, which focuses on understudied proteins, and research bias toward well-studied proteins.</li> <li>Synthetic lethality → <em>Synthetic lethality and combination targets: ML methods for finding drug pairs that work together</em> CRISPR-based screening, synthetic lethality, where dual perturbation of two genes kills a cell even though either single perturbation does not, and ML methods for drug-synergy prediction, including tools such as MAGeCK, CRISPRi, and Perturb-seq.</li> <li>Knowledge graphs and the rentosertib case study → <em>Knowledge graphs and case studies in AI-driven target discovery</em> Biomedical named-entity recognition, relation extraction, knowledge-graph embedding models, and walkthroughs of publicly documented AI-discovered-drug programs.</li> <li>Virtual cells → <em>Virtual cells for target discovery, perturbation models, and benchmarks</em> Single-cell atlases, perturbation-response models, Perturb-seq benchmarks, and how simulated interventions can prioritize target-validation experiments.</li> </ul> <h2 id="how-drug-target-discovery-connects-to-machine-learning-for-drug-discovery">How drug target discovery connects to <em>Machine Learning for Drug Discovery</em></h2> <p><a href="/book/"><em>Machine Learning for Drug Discovery</em></a> concentrates on methods that begin after a target has been chosen. Molecular property prediction, virtual screening, generative chemistry, protein structure prediction, drug repurposing, and multimodal pipelines are chapter-length topics because they are well-defined ML problems with benchmarks and data.</p> <p>If you came here from the book and have already built a property predictor, screened a virtual library, or trained a generative model for lead compounds, this series asks a prior question: <em>how did anyone decide that was the right target in the first place?</em> As you read through this series, you might notice that the methods change with the problem. For example, the work may involve graph learning over biomedical knowledge graphs, natural-language processing over biomedical literature, or exploiting multi-omics integration with multimodal models. However, at their core, these methods are more like variants of the methods we discuss in the book, rather than new or alien concepts.</p> <h2 id="whats-next-in-drug-target-discovery-open-questions">What’s next in drug target discovery: open questions</h2> <p><strong>Near-term readouts to watch.</strong> The 15 or so AI-discovered programs expected to enter pivotal Phase III trials in 2026 will provide the first broad test of whether front-of-funnel AI improves late-stage success rates. Rentosertib is the clearest test of AI-led target identification, because the target-selection step itself was AI-derived. Schrödinger/Takeda’s zasocitinib (TYK2 inhibitor) is a corresponding test of physics-based AI design.</p> <p><strong>Methodological open questions.</strong> First, how do you <em>benchmark</em> target-discovery ML without waiting a decade for each label? Surrogate endpoints, or proxy benchmarks, such as asking whether a method can rediscover known targets for known diseases using only pre-discovery literature, help but do not close the loop and may suffer from data leakage, data snooping, and related problems. Second, how far does the current generation of target-discovery ML extend beyond well-studied indications? The “ignorome,” the large set of human proteins that remain thinly studied, biases the training data toward what has already been studied. A target-discovery method that mostly finds targets adjacent to known ones misses the hardest part of the problem. The ignorome is the subject of a future article on novelty vs repurposing.</p> <p><strong>Regulatory shift.</strong> The FDA’s January 2025 draft guidance on AI in drug development established a risk-based credibility framework for when and how AI contributions must be qualified for regulatory submissions. Final guidance is expected within 2026, which will influence how target-discovery AI can be cited in an IND package (the dossier submitted to request permission for human testing).</p> <p>For the first time, the ML methods used in target discovery are starting to produce clinical readouts. However, the jury’s still out and the next decade will show how much they change outcomes – anyone telling you differently has a financial interest competing with transparency ;)</p> <hr/> <h2 id="further-reading">Further reading</h2> <ul> <li> <p>Citeline / Norstella (2024). <em>Why are clinical development success rates falling?</em> Industry report. <a href="https://www.norstella.com/why-clinical-development-success-rates-falling/">Norstella/Citeline article</a>. The primary source for the 6.7% Phase I LOA figure and the phase-by-phase transition rates cited throughout this post.</p> </li> <li> <p>Sun, D. et al. (2025). <em>A generative AI-discovered TNIK inhibitor for idiopathic pulmonary fibrosis: a randomized phase 2a trial.</em> <em>Nature Medicine.</em> <a href="https://doi.org/10.1038/s41591-025-03743-2">doi:10.1038/s41591-025-03743-2</a>. The rentosertib Phase IIa paper that the “worked example” section walks through; the primary source for readers who want to see the trial design, endpoints, and safety data in full.</p> </li> <li> <p>FDA Center for Drug Evaluation and Research (2025). <em>Considerations for the Use of Artificial Intelligence to Support Regulatory Decision Making for Drug and Biological Products.</em> Draft guidance, January 2025. <a href="https://www.fda.gov/regulatory-information/search-fda-guidance-documents/considerations-use-artificial-intelligence-support-regulatory-decision-making-drug-and-biological">FDA draft guidance</a>. The regulatory framework being written for how AI contributions - including target-discovery AI - must be qualified in IND and NDA submissions, that is, submissions to begin human testing or request approval.</p> </li> <li> <p>Wouters, O. J., McKee, M., Luyten, J. (2020). <em>Estimated Research and Development Investment Needed to Bring a New Medicine to Market, 2009–2018.</em> <em>JAMA</em>, 323(9): 844–853. The JAMA evaluation behind the $985M median / $1.3B mean figures cited in the funnel-economics section.</p> </li> <li> <p>RAND Corporation (2025). <em>Typical Cost of Developing a New Drug Is Skewed by Few High-Cost Outliers.</em> <a href="https://www.rand.org/news/press/2025/01/07.html">RAND press release</a>. The 2025 analysis that put the median direct R&amp;D cost at $150M (rising to $708M when capitalized for failures), and showed how a small number of outliers skew the industry-wide mean.</p> </li> </ul>]]></content><author><name>Noah Flynn</name><email>noahrflynn@gmail.com</email></author><category term="target discovery"/><category term="machine learning"/><category term="drug discovery"/><category term="target discovery"/><category term="AI drug discovery"/><summary type="html"><![CDATA[Drug target discovery is the least quantified stage of the drug pipeline and a major determinant of Phase II outcomes. Why it matters, and where ML fits.]]></summary></entry><entry><title type="html">What This Site Is For</title><link href="https://noahrflynn.com/blog/2026/what-this-site-is-for/" rel="alternate" type="text/html" title="What This Site Is For"/><published>2026-04-29T17:00:00+00:00</published><updated>2026-04-29T17:00:00+00:00</updated><id>https://noahrflynn.com/blog/2026/what-this-site-is-for</id><content type="html" xml:base="https://noahrflynn.com/blog/2026/what-this-site-is-for/"><![CDATA[<p>This site is my home base for work that sits between research, production AI systems, drug discovery, and teaching.</p> <p>The near-term plan is simple: publish notes that make complicated systems easier to reason about. Some posts will unpack ideas from <em>Machine Learning for Drug Discovery</em>. Others will come from teaching graduate cheminformatics and machine learning at UC Berkeley, or from the practical messiness of building agentic systems that have to work outside a benchmark.</p> <p>I want the writing here to be useful to people who build things: researchers, students, applied scientists, and drug discovery folks trying to make sense of modern ML without losing the thread of the underlying science.</p> <p>This site is also the canonical archive. I may syndicate posts to <a href="https://substack.com/@noaflynn">Substack</a>, LinkedIn, or X, but the version here is the one I will keep current. You can follow via <a href="/feed.xml">RSS</a> if you prefer the old, sturdy internet.</p> <p>For now, the best starting points are the <a href="/book/">book</a>, the <a href="/talks/">talks page</a>, and the <a href="/publications/">publications list</a>.</p>]]></content><author><name>Noah Flynn</name><email>noahrflynn@gmail.com</email></author><category term="launch"/><category term="machine learning"/><category term="drug discovery"/><category term="agents"/><category term="teaching"/><category term="writing"/><summary type="html"><![CDATA[A short note on InfoArk, the book, teaching, and the kinds of technical writing I plan to publish here.]]></summary></entry></feed>