Transformers for Molecular Property Prediction: Lessons Learned from the Past Five Years

J Chem Inf Model. 2024 Aug 26;64(16):6259-6280. doi: 10.1021/acs.jcim.4c00747. Epub 2024 Aug 13.

Abstract

Molecular Property Prediction (MPP) is vital for drug discovery, crop protection, and environmental science. Over the last decades, diverse computational techniques have been developed, from using simple physical and chemical properties and molecular fingerprints in statistical models and classical machine learning to advanced deep learning approaches. In this review, we aim to distill insights from current research on employing transformer models for MPP. We analyze the currently available models and explore key questions that arise when training and fine-tuning a transformer model for MPP. These questions encompass the choice and scale of the pretraining data, optimal architecture selections, and promising pretraining objectives. Our analysis highlights areas not yet covered in current research, inviting further exploration to enhance the field's understanding. Additionally, we address the challenges in comparing different models, emphasizing the need for standardized data splitting and robust statistical analysis.

Keywords: Molecular property prediction; SMILES; benchmarking; chemical language models; deep learning; domain knowledge; fine-tuning; pretraining; sequence-based chemical models; systematic analysis; transformers.

Publication types

  • Review

MeSH terms

  • Deep Learning
  • Drug Discovery / methods
  • Machine Learning*