MOLGAECL: Molecular Graph Contrastive Learning via Graph Auto-Encoder Pretraining and Fine-Tuning Based on Drug-Drug Interaction Prediction.

Yu Li, Lin-Xuan Hou, Hai-Cheng Yi, Zhu-Hong You, Shi-Hong Chen, Jia Zheng, Yang Yuan, Cheng-Gang Mi
Author Information
  1. Yu Li: School of Computer Science and Artificial Intelligence, Changzhou University, Changzhou 213164, China. ORCID
  2. Lin-Xuan Hou: School of Computer Science and Artificial Intelligence, Changzhou University, Changzhou 213164, China.
  3. Hai-Cheng Yi: School of Computer Science and Artificial Intelligence, Changzhou University, Changzhou 213164, China.
  4. Zhu-Hong You: School of Computer Science and Artificial Intelligence, Changzhou University, Changzhou 213164, China.
  5. Shi-Hong Chen: School of Computer Science and Artificial Intelligence, Changzhou University, Changzhou 213164, China.
  6. Jia Zheng: School of Computer Science and Artificial Intelligence, Changzhou University, Changzhou 213164, China.
  7. Yang Yuan: School of Computer Science and Artificial Intelligence, Changzhou University, Changzhou 213164, China.
  8. Cheng-Gang Mi: Foreign Language and Literature Institute, Xi'an International Studies University, Xi'an 710129, China.

Abstract

Drug-drug interactions influence drug efficacy and patient prognosis, providing substantial research value. Some existing methods struggle with the challenges posed by sparse networks or lack the capability to integrate data from multiple sources. In this study, we propose MOLGAECL, a novel approach based on graph autoencoder pretraining and molecular graph contrastive learning. Initially, a large number of unlabeled molecular graphs are pretrained using a graph autoencoder, where graph contrastive learning is applied for more accurate representation of the drugs. Subsequently, a full-parameter fine-tuning is performed on different data sets to adapt the model for drug interaction-related prediction tasks. To assess the effectiveness of MOLGAECL, comparison experiments with state-of-the-art methods, fine-tuning comparison experiments, and parameter sensitivity analysis are conducted. Extensive experimental results demonstrate the superior performance of MOLGAECL. Specifically, MOLGAECL achieves an average increase of 6.13% in accuracy, 6.14% in AUROC, and 8.16% in AUPRC across all data sets.

MeSH Term

Drug Interactions
Machine Learning
Humans

Word Cloud

Created with Highcharts 10.0.0MOLGAECLgraphdatadrugmethodsautoencodermolecularcontrastivelearningfine-tuningsetscomparisonexperiments6GraphDrug-druginteractionsinfluenceefficacypatientprognosisprovidingsubstantialresearchvalueexistingstrugglechallengesposedsparsenetworkslackcapabilityintegratemultiplesourcesstudyproposenovelapproachbasedpretrainingInitiallylargenumberunlabeledgraphspretrainedusingappliedaccuraterepresentationdrugsSubsequentlyfull-parameterperformeddifferentadaptmodelinteraction-relatedpredictiontasksassesseffectivenessstate-of-the-artparametersensitivityanalysisconductedExtensiveexperimentalresultsdemonstratesuperiorperformanceSpecificallyachievesaverageincrease13%accuracy14%AUROC816%AUPRCacrossMOLGAECL:MolecularContrastiveLearningviaAuto-EncoderPretrainingFine-TuningBasedDrug-DrugInteractionPrediction

Similar Articles

Cited By