Google DeepMind and Stanford just compressed a 5-7 year diagnostic odyssey into actionable insights. Their AI correctly identified causative genes for rare diseases - including a novel mutation for hearing loss that was later validated in the lab. Published in 𝘈𝘥𝘷𝘢𝘯𝘤𝘦𝘥 𝘚𝘤𝘪𝘦𝘯𝘤𝘦, this isn't just another AI research paper. It's a blueprint for fundamentally changing how rare disease patients get diagnosed. 𝗧𝗵𝗲 𝗕𝗿𝗲𝗮𝗸𝘁𝗵𝗿𝗼𝘂𝗴𝗵 Researchers at Google DeepMind and Stanford University demonstrated that large language models can dramatically accelerate rare genetic disease diagnosis. Using Google's Med-PaLM 2 and Gemini 2.5 Pro, the team analyzed complex genetic and clinical data to identify causative genetic factors in both mouse models and human patients. 𝗪𝗵𝗮𝘁 𝗠𝗮𝗸𝗲𝘀 𝗧𝗵𝗶𝘀 𝗦𝗶𝗴𝗻𝗶𝗳𝗶𝗰𝗮𝗻𝘁 The AI solved genetic problems of increasing complexity with remarkable precision: • Identified a novel causative gene for hearing loss in mice (later validated in the lab) • Analyzed genomic data from human patients with multifaceted symptom profiles • Successfully pinpointed underlying genetic variants, including variants of unknown significance (VUS) The system uses a retrieval and grounding pipeline to analyze vast amounts of genetic information and generate ranked hypotheses - essentially reasoning through genetic data the way a skilled clinical geneticist would, but at scale. 𝗜𝗺𝗽𝗹𝗶𝗰𝗮𝘁𝗶𝗼𝗻𝘀 𝗳𝗼𝗿 𝗣𝗵𝗮𝗿𝗺𝗮 & 𝗛𝗲𝗮𝗹𝘁𝗵𝗰𝗮𝗿𝗲 For rare disease drug development, faster diagnosis means: • Earlier patient identification for clinical trials • More accurate patient stratification • Accelerated pathway from genomic discovery to targeted therapy development • Reduced diagnostic odyssey costs (currently averaging $5M per patient over their journey) This represents more than incremental progress in AI-assisted diagnostics. It's a fundamental shift in how we might approach precision medicine - compressing years of diagnostic uncertainty into actionable insights that enable faster therapeutic intervention. The question isn't whether AI will transform rare disease diagnosis. It's how quickly we can validate and implement these tools in clinical practice. Follow Dr. Suzanne Morgan for more insights on AI and Rare Disease Source: https://lnkd.in/d-ki3nNu
Genetic Data Interpretation Technologies
Explore top LinkedIn content from expert professionals.
-
-
Honored to be quoted in TIME today on a new paper from Goodfire and Mayo Clinic that I believe represents a genuine inflection point for genomics. They took Evo 2, a genomic foundation model trained on 128,000 genomes, and applied interpretability techniques to predict which genetic variants cause disease. State-of-the-art performance across 4.2 million variants in ClinVar. Open-sourced. But the performance isn't the story. The story is what interpretability makes possible. A standard deep learning model gives you a score. This variant is probably pathogenic. Fine. But in medicine — and especially in genomics — a score without a mechanism is hard to act on. What this work does differently: it identifies the biological features of mutations driving the model's prediction. Not just what the model thinks, but what the model saw that led it there. I was asked to comment for the TIME piece, and what I said is what I genuinely believe: as genome sequencing costs continue to drop — whole genomes at near-zero cost are coming — the bottleneck shifts from data acquisition to interpretation. We'll be drowning in variants of uncertain significance. Tools that can identify the biology, not just the risk score, will be the ones that matter. This connects to something I think about constantly in my own work building foundation models for biology. We train large models on millions of cells, thousands of perturbations, billions of base pairs. These models encode enormous amounts of biological knowledge — patterns of gene regulation, cell state transitions, the logic of how perturbations propagate through networks. But that knowledge sits locked inside parameter matrices. We can probe it with benchmarks. We can't read it. Interpretability is the key. Not because we need to audit the model for safety — though that matters — but because the model has learned things about biology that we haven't formally characterized yet. The features it's using to make predictions are hypotheses about how biology works. Extracting those features is doing science. The Goodfire/Mayo Clinic work is a proof point that this isn't theoretical. Interpretability applied to a genomics model doesn't just validate the predictions — it generates biological insight. Features that correspond to real mechanisms. Features that could inform therapy design. We're at an early stage. The interpretability tooling for biology-specific foundation models is years behind what exists for language models. But the trajectory is clear: the labs and companies that invest in understanding what their models have learned — not just what they can predict — will extract far more scientific value from the same data. That's the bet I'm making. And this TIME coverage suggests the bet is starting to pay off. Link: https://lnkd.in/ehJuQJwG
-
What if your Practitioner could tell you, #BEFORE a prescription is written ,that a drug is likely to cause harm based on your DNA? That's not a future concept. I have been building it. Over the past few months I've been developing a Genetic Digital Twin protocol. The idea is straightforward but powerful: take each member's SNP (single nucleotide polymorphism) data, map it against a pharmacogenomic knowledge base, and construct a living genetic profile that layers: 🧬 General health risk SNPs — conditions like hypertension, methylation disorders, diabetes, and haemochromatosis 💊 PGx SNPs — variants in genes like CYP2D6, MTHFR, VKORC1, and SLCO1B1 that directly affect how a person metabolises drugs 📊 Risk scores — per condition, per member, across a 200+ person cohort from over 1 million raw genetic data points 🩺 Medication guidance — matched pharmacogenomic recommendations at the individual level But the part that's been most exciting to build is the What-If Scenario Engine. This is where the digital twin stops being descriptive and becomes decision-support: 1. Add a drug: prescribe warfarin to this patient, what does their CYP2C9/VKORC1 profile predict? 2. Remove a drug: if we de-prescribe codeine, how many high-risk flags are resolved? 3. Switch a drug: moving the population from clopidogrel to ticagrelor, who benefits genetically and who does not? 4. Condition-based: for all diabetic members, which co-prescriptions carry the highest genomic risk? The output is a fully interactive browser app with no server, no setup, where a clinician or medical advisor can select any member, view their 5-layer genetic twin, and run live scenarios in seconds. We are entering an era where population health management can be truly personalized, not by demographics or claims history alone, but by the genetic architecture of each individual. The technology exists. The data exists. The question now is how quickly healthcare systems are willing to operationalise it. Happy to discuss the methodology with anyone working at the intersection of genomics, health insurance, and precision medicine. #Pharmacogenomics #DigitalTwin #PrecisionMedicine #Genomics #DataScience
-
Synthetic biology is - quite literally - our future. A goundbreaking new biological foundation model Evo2 achieves state-of-the-art prediction of genetic variation impacts and generates coherent genome sequences, spanning all domains of life. A diverse team from leading research institutions including Arc Institute Stanford University NVIDIA University of California, Berkeley trained the model on 9.3 trillion DNA base pairs and has fully shared all code, parameters, and data. A few highlights from the paper (link in comments) 🔬 Zero-shot prediction achieves state-of-the-art accuracy in genetic variant interpretation. Evo 2 can predict the functional consequences of genetic mutations across all domains of life without specialized training. It surpasses existing models in assessing the pathogenicity of both coding and noncoding variants, including BRCA1 cancer-linked mutations. This generalist capability suggests Evo 2 could revolutionize genetic disease research, reducing reliance on expensive, manually curated datasets. 🛠 Genome-scale generation paves the way for synthetic life design. Evo 2 can generate full-length genome sequences with realistic structure and function, including mitochondrial genomes, bacterial chromosomes, and yeast DNA. Unlike prior models, Evo 2 ensures natural sequence coherence, improving synthetic biology applications like engineered microbes or artificial organelles. This sets the stage for programmable biology at an unprecedented scale. 🧬 Unprecedented long-context understanding revolutionizes genomic analysis. Evo 2 operates with a context window of up to 1 million nucleotides—far beyond the capabilities of previous models—allowing it to analyze genomic features across vast distances. This ability enables it to accurately identify regulatory elements, exon-intron boundaries, and structural components critical for understanding genome function. Its long-context recall is a major breakthrough for interpreting complex biological sequences. 🎛 Inference-time search enables controllable epigenomic design. Evo 2’s generative abilities extend beyond raw DNA sequence to epigenomic features, allowing researchers to design sequences with specific chromatin accessibility patterns. This approach successfully encoded Morse code messages into synthetic epigenomes, demonstrating a new method for controlling gene regulation via AI. This could lead to breakthroughs in gene therapy and epigenetic engineering. 🔮 Future potential: Toward AI-driven biological design and virtual cell modeling. Evo 2 represents a major leap toward AI-powered genomic engineering. Future iterations could integrate additional biological layers—such as transcriptomics and proteomics—to create virtual cell models that simulate complex cellular behaviors. This could revolutionize drug discovery, genetic therapy, and even synthetic life creation.
-
"All of life encodes information with DNA. While tools for sequencing, synthesis, and editing of genomic code have transformed biological research, intelligently composing new biological systems would also require a deep understanding of the immense complexity encoded by genomes. We introduce Evo 2, a biological foundation model trained on 9.3 trillion DNA base pairs from a highly curated genomic atlas spanning all domains of life. We train Evo 2 with 7B and 40B parameters to have an unprecedented 1 million token context window with single-nucleotide resolution. Evo 2 learns from DNA sequence alone to accurately predict the functional impacts of genetic variation—from noncoding pathogenic mutations to clinically significant BRCA1 variants—without task-specific finetuning. Applying mechanistic interpretability analyses, we reveal that Evo 2 autonomously learns a breadth of biological features, including exon–intron boundaries, transcription factor binding sites, protein structural elements, and prophage genomic regions. Beyond its predictive capabilities, Evo 2 generates mitochondrial, prokaryotic, and eukaryotic sequences at genome scale with greater naturalness and coherence than previous methods. Guiding Evo 2 via inference-time search enables controllable generation of epigenomic structure, for which we demonstrate the first inference-time scaling results in biology." https://lnkd.in/etVTRWiX
-
Human biology is generating data at a scale that is impossible to fully interpret without AI. HumanBase addresses this challenge by transforming massive, complex genomic datasets into clear, actionable biological insights. HumanBase is an interactive AI platform that integrates more than 62,000 public datasets and applies advanced machine learning and deep learning models to understand gene function, regulatory networks, and variant effects across specific tissues and cell types. Standardized data flows through AI driven modeling and prediction pipelines, enabling researchers to move from raw data to meaningful biological hypotheses. What makes HumanBase especially powerful is its ability to predict the impact of noncoding and previously unseen genetic variants at single nucleotide resolution. Its AI models estimate effects on epigenetics, gene expression, and post transcriptional regulation, helping prioritize pathogenic mutations and reveal underlying molecular mechanisms. Just as importantly, the platform is freely available, user friendly, and requires no coding experience, making advanced AI tools accessible to the broader research community. This is a strong example of how AI is shifting biology from descriptive to predictive, helping scientists ask better questions and design smarter experiments. Read more in Nature Methods https://lnkd.in/gcHCAzAr Follow Zain Khalpey, MD, PhD, FACS for more on Ai & Healthcare. #AIinBiology #ComputationalBiology #Genomics #HumanBiology #MachineLearning #DeepLearning #PrecisionMedicine #Bioinformatics #SystemsBiology #LifeSciences #DataDrivenScience #GeneticVariants #NoncodingDNA #ResearchInnovation #OpenScience #HealthcareAI #FutureOfMedicine #ScientificAI #TranslationalScience #DigitalBiology
-
The recent article published in Nature by Žiga Avsec and colleagues at Google DeepMind underscores the impact that the latest #genomics advances have on the #lifesciences ecosystem. Genetic variation plays a fundamental role in shaping biological processes and contributes significantly to disease susceptibility. However, accurately characterizing the functional consequences of #DNA sequence variation remains a major scientific challenge, particularly because approximately 98% of human genetic variants occur in non-coding regions of the genome. These regions, while not translated into proteins, exert critical regulatory control over gene expression, chromatin structure, and cellular function. Recent advances in deep-learning–based genomics have demonstrated substantial promise in addressing this challenge, yet most existing models are constrained by an inherent trade-off between the length of input DNA sequence they can process and the resolution of their predictions. In this manuscript, the authors introduce how Alphagenome Inc, a unified deep-learning model, can overcome these limitations. It can analyze DNA sequences spanning up to one million base pairs and deliver high-resolution predictions across a broad spectrum of genomic features. The authors suggest that AlphaGenome represents a significant step toward systematically interpreting non-coding variation, with potential applications in elucidating the genetic basis of disease, informing synthetic DNA design, and advancing fundamental understanding of genome regulation In my opinion, this article illustrates that we are at a critical inflection point for #precisionhealth and #personalizedmedicine, as the ability to accurately decode regulatory genomic variation at scale enables more individualized, predictive, and preventive approaches to disease management. However, the deployment of such powerful genomic #AI systems also introduces profound accountability imperatives. The scale, sensitivity, and longitudinal value of genomic data demand robust #data #governance frameworks, advanced #cyber-#ethics safeguards, and #privacy-preserving architectures by design. Given the sensitivity and societal impact of genomic data, equitable access must also be embedded from inception. In line with the DUBAI FUTURE FOUNDATION’s foresight-driven governance approach, #quantum-resilient security architectures are essential to safeguard genomic infrastructures against future cryptographic threats and to sustain public #trust in innovative #healthcare #ecosystems.
Explore categories
- Hospitality & Tourism
- Productivity
- Finance
- Soft Skills & Emotional Intelligence
- Project Management
- Education
- Technology
- Leadership
- Ecommerce
- User Experience
- Recruitment & HR
- Customer Experience
- Real Estate
- Marketing
- Sales
- Retail & Merchandising
- Supply Chain Management
- Future Of Work
- Consulting
- Writing
- Economics
- Artificial Intelligence
- Employee Experience
- Healthcare
- Workplace Trends
- Fundraising
- Networking
- Corporate Social Responsibility
- Negotiation
- Communication
- Engineering
- Career
- Business Strategy
- Change Management
- Organizational Culture
- Design
- Innovation
- Event Planning
- Training & Development