Command Palette
Search for a command to run...
{Junzhou Huang Xi Xiao Yu Rong Tingyang Xu Peilin Zhao Kelong Mao}

Abstract
With massive possible synthetic routes in chemistry, retrosynthesis prediction is still a challenge for researchers. Recently, retrosynthesis prediction is formulated as a Machine Translation (MT) task. Namely, since each molecule can be represented as a Simplified Molecular-Input Line-Entry System (SMILES) string, the process of synthesis is analogized to a process of language translation from reactants to products. However, the MT models that applied on SMILES data usually ignore the information of natural atomic connections and the topology of molecules. In this paper, we propose a Graph Enhanced Transformer (GET) framework, which adopts both the sequential and graphical information of molecules. Four different GET designs are proposed, which fuse the SMILES representations with atom embedding learned from our improved Graph Neural Network (GNN). Empirical results show that our model significantly outperforms the Transformer model in test accuracy.
Benchmarks
| Benchmark | Methodology | Metrics |
|---|---|---|
| single-step-retrosynthesis-on-uspto-50k | GET-LT1 (reaction class unknown) | Top-1 accuracy: 44.9 Top-10 accuracy: 65.9 Top-3 accuracy: 58.8 Top-5 accuracy: 62.4 |
| single-step-retrosynthesis-on-uspto-50k | GET-LT1 (reaction class as prior) | Top-1 accuracy: 57.4 Top-10 accuracy: 77.4 Top-3 accuracy: 71.3 Top-5 accuracy: 74.8 |
| single-step-retrosynthesis-on-uspto-50k | GET-LT2 (reaction class as prior) | Top-1 accuracy: 56.2 Top-10 accuracy: 74.7 Top-3 accuracy: 69.4 Top-5 accuracy: 72.5 |
Build AI with AI
From idea to launch — accelerate your AI development with free AI co-coding, out-of-the-box environment and best price of GPUs.