New Harvard-Developed AI System Unlocks Biology’s Source Code

DNA Technology Glowing

A groundbreaking study by Yunha Hwang and team has developed gLM, an AI system that decodes the complex language of genomics from extensive microbial data. This innovation enables a deeper understanding of gene functions and regulations, leading to new discoveries in genomics. gLM exemplifies the potential of AI in advancing life sciences and tackling global challenges. Credit: SciTechDaily.com

Artificial Intelligence (AI) systems, like ChatGPT, have taken the world by storm. Thre isn’t much they don’t have a hand in, from recommending the next binge-worthy TV show to helping navigate through traffic. But, can AI systems learn the language of life and help biologists reveal exciting breakthroughs in science?

In a new study published in Nature Communications, an interdisciplinary team of researchers led by Yunha Hwang, PhD candidate in the Department of Organismic and Evolutionary Biology (OEB) at Harvard, have pioneered an artificial intelligence (AI) system capable of deciphering the intricate language of genomics.

Genomic language is the source code of biology. It describes the biological functions and regulatory grammar encoded in genomes. The researchers asked can we develop an AI engine to “read” the genomic language and become fluent in the language, understanding the meaning, or functions and regulations, of genes? The team fed the microbial metagenomic data set, the largest and most diverse genomic dataset available, to the machine to create the Genomic Language Model (gLM).

The Challenge of Genomic Data

“In biology, we have a dictionary of known words and researchers work within those known words. The problem is that this fraction of known words constitutes less than one percent of biological sequences,” said Hwang, “the quantity and diversity of genomic data is exploding, but humans are incapable of processing such a large amount of complex data.”

Large language models (LLMs), like GPT4, learn meanings of words by processing massive amounts of diverse text data that enables understanding the relationships between words. Genomic language model (gLM) learns from highly diverse metagenomic data, sourced from microbes inhabiting various environments including the ocean, soil, and human gut. With this data, gLM learns to understand the functional “semantics” and regulatory “syntax” of each gene by learning the relationship between the gene and its genomic context. gLM, like LLMs, is a self-supervised model – this means that it learns meaningful representations of genes from data alone and does not require human-assigned labels.

Unveiling the Unknown in Genomics

Researchers have sequenced some of the most commonly studied organisms like people, E. coli, and fruit flies. However, even for the most studied genomes, the majority of the genes remain poorly characterized. “We’ve learned so much in this revolutionary age of ‘omics’, including how much we don’t know,” said senior author Professor Peter Girguis, also in OEB at Harvard. “We asked, how can we glean meaning from something without relying on a proverbial dictionary? How do we better understand the content and context of a genome?”

The study demonstrates that gLM learns enzymatic functions and co-regulated gene modules (called operons), and provides genomic context that can predict gene function. The model also learns taxonomic information and context-dependencies of gene functions. Strikingly, gLM does not know which enzyme it is seeing, nor what bacteria the sequence comes from. However, because it has seen many sequences and understands the evolutionary relationships between the sequences during training, it is able to derive the functional and evolutionary relationships between sequences.

The Potential of gLM in Biology

“Like words, genes can have different “meanings” depending on the context they are found in. Conversely, highly differentiated genes can be “synonymous” in function. gLM allows for a much more nuanced framework for understanding gene function. This is in contrast to the existing method of one-to-one mapping from sequence to annotation, which is not representative of the dynamic and context-dependent nature of the genomic language,” said Hwang.

Hwang teamed with co-authors Andre Cornman (an independent researcher in machine learning and biology), Sergey Ovchinnikov (former John Harvard Distinguished Fellow and current Assistant Professor at MIT), and Elizabeth Kellogg (Associate Faculty at St. Jude Children’s Research Hospital) to form an interdisciplinary team with strong backgrounds in microbiology, genomes, bioinformatics, protein science, and machine learning.

“In the lab, we are stuck in a step-by-step process of finding a gene, making a protein, purifying it, characterizing it, etc. and so we kind of discover only what we already know,” Girguis said. gLM, however, allows biologists to look at the context of an unknown gene and its role when it’s often found in similar groups of genes. The model can tell researchers that these groups of genes work together to achieve something, and it can provide the answers that do not appear in the “dictionary”.

“Genomic context contains critical information for understanding the evolutionary history and evolutionary trajectories of different proteins and genes,” Hwang said. “Ultimately, gLM learns this contextual information to help researchers understand the functions of genes that previously were unannotated.”

“Traditional functional annotation methods typically focus on one protein at a time, ignoring the interactions across proteins. gLM represents a major advancement by integrating the concept of gene neighborhoods with language models, thereby providing a more comprehensive view of protein interactions,” stated Martin Steinegger (Assistant Professor, Seoul National University), an expert in bioinformatics and machine learning, who was not involved in the study.

With genomic language modeling, biologists can discover new genomic patterns and uncover novel biology. gLM is a significant milestone in interdisciplinary collaboration driving advancements in the life sciences.

“With gLM we can gain new insights into poorly annotated genomes,” said Hwang. “gLM can also guide experimental validation of functions and enable discoveries of novel functions and biological mechanisms. We hope gLM can accelerate the discovery of novel biotechnological solutions for climate change and bioeconomy.”

Reference: “Genomic language model predicts protein co-regulation and function” by Yunha Hwang, Andre L. Cornman, Elizabeth H. Kellogg, Sergey Ovchinnikov and Peter R. Girguis, 3 April 2024, Nature Communications.
DOI: 10.1038/s41467-024-46947-9

1 Comment on "New Harvard-Developed AI System Unlocks Biology’s Source Code"

  1. Ashley Ferguson | April 29, 2024 at 5:55 am | Reply

    It’s chatgpt not gtp. First paragraph

Leave a comment

Email address is optional. If provided, your email will not be published or shared.