Molecular Transformer unifies reaction prediction and retrosynthesis across pharma chemical space

Predicting how a complex molecule reacts with different reagents, and how to synthesise complex molecules from simpler starting materials, are fundamental to organic chemistry. We show that an attention-based machine translation model - Molecular Transformer - tackles both reaction prediction and re...

Ausführliche Beschreibung

Gespeichert in:
Bibliographische Detailangaben
Veröffentlicht in:Chemical communications (Cambridge, England) England), 2019-10, Vol.55 (81), p.12152-12155
Hauptverfasser: Lee, Alpha A, Yang, Qingyi, Sresht, Vishnu, Bolgar, Peter, Hou, Xinjun, Klug-McLeod, Jacquelyn L, Butler, Christopher R
Format: Artikel
Sprache:eng
Schlagworte:
Online-Zugang:Volltext
Tags: Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
Beschreibung
Zusammenfassung:Predicting how a complex molecule reacts with different reagents, and how to synthesise complex molecules from simpler starting materials, are fundamental to organic chemistry. We show that an attention-based machine translation model - Molecular Transformer - tackles both reaction prediction and retrosynthesis by learning from the same dataset. Reagents, reactants and products are represented as SMILES text strings. For reaction prediction, the model "translates" the SMILES of reactants and reagents to product SMILES, and the converse for retrosynthesis. Moreover, a model trained on publicly available data is able to make accurate predictions on proprietary molecules extracted from pharma electronic lab notebooks, demonstrating generalisability across chemical space. We expect our versatile framework to be broadly applicable to problems such as reaction condition prediction, reagent prediction and yield prediction. We develop a machine learning model that tackles both reaction prediction and retrosynthesis by learning from the same dataset. The model is generalizable across chemical space.
ISSN:1359-7345
1364-548X
DOI:10.1039/c9cc05122h