TY - JOUR
T1 - Deciphering RNA splicing logic with interpretable machine learning
AU - Liao, Susan E.
AU - Sudarshan, Mukund
AU - Regev, Oded
N1 - Publisher Copyright:
Copyright © 2023 the Author(s).
PY - 2023
Y1 - 2023
N2 - Machine learning methods, particularly neural networks trained on large datasets, are transforming how scientists approach scientific discovery and experimental design. However, current state-of-the-art neural networks are limited by their uninterpretability: Despite their excellent accuracy, they cannot describe how they arrived at their predictions. Here, using an “interpretable-by-design” approach, we present a neural network model that provides insights into RNA splicing, a fundamental process in the transfer of genomic information into functional biochemical products. Although we designed our model to emphasize interpretability, its predictive accuracy is on par with state-of-the-art models. To demonstrate the model’s interpretability, we introduce a visualization that, for any given exon, allows us to trace and quantify the entire decision process from input sequence to output splicing prediction. Importantly, the model revealed uncharacterized components of the splicing logic, which we experimentally validated. This study highlights how interpretable machine learning can advance scientific discovery.
AB - Machine learning methods, particularly neural networks trained on large datasets, are transforming how scientists approach scientific discovery and experimental design. However, current state-of-the-art neural networks are limited by their uninterpretability: Despite their excellent accuracy, they cannot describe how they arrived at their predictions. Here, using an “interpretable-by-design” approach, we present a neural network model that provides insights into RNA splicing, a fundamental process in the transfer of genomic information into functional biochemical products. Although we designed our model to emphasize interpretability, its predictive accuracy is on par with state-of-the-art models. To demonstrate the model’s interpretability, we introduce a visualization that, for any given exon, allows us to trace and quantify the entire decision process from input sequence to output splicing prediction. Importantly, the model revealed uncharacterized components of the splicing logic, which we experimentally validated. This study highlights how interpretable machine learning can advance scientific discovery.
KW - RNA splicing
KW - artificial intelligence
KW - interpretable machine learning
UR - http://www.scopus.com/inward/record.url?scp=85174324424&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85174324424&partnerID=8YFLogxK
U2 - 10.1073/pnas.2221165120
DO - 10.1073/pnas.2221165120
M3 - Article
C2 - 37796983
AN - SCOPUS:85174324424
SN - 0027-8424
VL - 120
JO - Proceedings of the National Academy of Sciences of the United States of America
JF - Proceedings of the National Academy of Sciences of the United States of America
IS - 41
M1 - e2221165120
ER -