Machine learning is a powerful tool in computational biology, enabling the analysis of a wide range of biomedical data such as genomic sequences and biological imaging. But when researchers use machine learning in computational biology, understanding model behavior remains crucial for uncovering the underlying biological mechanisms in health and disease.
In a recent article in Nature Methods, researchers at Carnegie Mellon University’s School of Computer Science propose guidelines that outline pitfalls and opportunities for using interpretable machine learning methods to tackle computational biology problems. The Perspectives article, “Applying Interpretable Machine Learning in Computational Biology — Pitfalls, Recommendations and Opportunities for New Developments,” is featured in the journal’s August special issue on AI.
“Interpretable machine learning has generated significant excitement as machine learning and artificial intelligence tools are being applied to increasingly important problems,” said Ameet Talwalkar, an associate professor in CMU’s Machine Learning Department (MLD). “As these models grow in complexity, there is great promise not only in developing highly predictive models but also in creating tools that help end users understand how and why these models make certain predictions. However, it is crucial to acknowledge that interpretable machine learning has yet to deliver turnkey solutions to this interpretability problem.”
The paper is a collaboration between doctoral students Valerie Chen in MLD and Muyu (Wendy) Yang in the Ray and Stephanie Lane Computational Biology Department. Chen’s earlier work critiquing the interpretable machine learning community’s lack of grounding in downstream use cases inspired the article, and the idea was developed through discussions with Yang and Jian Ma, the Ray and Stephanie Lane Professor of Computational Biology.
“Our collaboration began with a deep dive into computational biology papers to survey the application of interpretable machine learning methods,” Yang said. “We noticed that many applications used these methods in a somewhat ad hoc manner. Our goal with this paper was to provide guidelines for more robust and consistent use of interpretable machine learning methods in computational biology.”
One major pitfall the paper addresses is the reliance on a single interpretable machine learning method. Instead, the researchers recommend using multiple interpretable machine learning methods with diverse sets of hyperparameters and comparing their results to obtain a more comprehensive understanding of the model behavior and its underlying interpretations.
“While some machine learning models seem to work surprisingly well, we often do not fully understand why,” Ma said. “In scientific domains like biomedicine, understanding why models work is crucial for discovering fundamental biological mechanisms.”
The paper also warns against cherry-picking results when evaluating interpretable machine learning methods, as this can lead to incomplete or biased interpretations of scientific findings.
Chen emphasized that the guidelines may have broader implications for a wider audience of researchers interested in applying interpretable machine-learning methods to their work.
“We hope that machine learning researchers developing new interpretable machine learning methods and tools — particularly those working on explaining large language models — will carefully consider the human-centric aspects of interpretable machine learning,” Chen said. “This includes understanding who their target user is and how the method will be used and evaluated.”
While understanding model behavior remains crucially important for scientific discovery and a fundamentally unsolved machine learning problem, the authors hope these challenges spur further interdisciplinary collaborations to facilitate the broader use of AI for scientific impact.