TY - JOUR
T1 - Molecular pretraining models towards molecular property prediction
AU - Qiao, Jianbo
AU - Gao, Wenjia
AU - Jin, Junru
AU - Wang, Ding
AU - Guo, Xu
AU - Manavalan, Balachandran
AU - Wei, Leyi
N1 - Publisher Copyright:
© Science China Press 2025.
PY - 2025/7
Y1 - 2025/7
N2 - Molecular property prediction plays a pivotal role in advancing our understanding of molecular representations, serving as a key driver for progress in drug discovery. Leveraging deep learning to gain comprehensive insights into molecular properties has become increasingly critical. Recent breakthroughs in molecular property prediction have been achieved through molecular pretraining models, which utilize large-scale databases of unlabeled molecules for pretraining, followed by fine-tuning for specific downstream tasks. These models enable a deeper understanding of molecular properties. In this study, we review recent advancements in molecular property prediction using molecular pretraining models. Our focus includes molecular descriptors, the impact of pretraining dataset size, molecular characterization model architectures, and the diversity of pretraining task types. Additionally, we compare the performance of existing methods and propose future directions to enhance the effectiveness of molecular pretraining models.
AB - Molecular property prediction plays a pivotal role in advancing our understanding of molecular representations, serving as a key driver for progress in drug discovery. Leveraging deep learning to gain comprehensive insights into molecular properties has become increasingly critical. Recent breakthroughs in molecular property prediction have been achieved through molecular pretraining models, which utilize large-scale databases of unlabeled molecules for pretraining, followed by fine-tuning for specific downstream tasks. These models enable a deeper understanding of molecular properties. In this study, we review recent advancements in molecular property prediction using molecular pretraining models. Our focus includes molecular descriptors, the impact of pretraining dataset size, molecular characterization model architectures, and the diversity of pretraining task types. Additionally, we compare the performance of existing methods and propose future directions to enhance the effectiveness of molecular pretraining models.
KW - graph neural network (GNN)
KW - graph Transformer
KW - molecular pretraining models
KW - molecular property prediction
KW - MoleculeNet
KW - PubChem
UR - https://www.scopus.com/pages/publications/105008890283
U2 - 10.1007/s11432-024-4457-2
DO - 10.1007/s11432-024-4457-2
M3 - Review article
AN - SCOPUS:105008890283
SN - 1674-733X
VL - 68
JO - Science China Information Sciences
JF - Science China Information Sciences
IS - 7
M1 - 170104
ER -