skip to main content
10.1145/3534678.3539368acmconferencesArticle/Chapter ViewAbstractPublication PageskddConference Proceedingsconference-collections
research-article

Unified 2D and 3D Pre-Training of Molecular Representations

Authors Info & Claims
Published:14 August 2022Publication History

ABSTRACT

Molecular representation learning has attracted much attention recently. A molecule can be viewed as a 2D graph with nodes/atoms connected by edges/bonds, and can also be represented by a 3D conformation with 3-dimensional coordinates of all atoms. We note that most previous work handles 2D and 3D information separately, while jointly leveraging these two sources may foster a more informative representation. In this work, we explore this appealing idea and propose a new representation learning method based on a unified 2D and 3D pre-training. Atom coordinates and interatomic distances are encoded and then fused with atomic representations through graph neural networks. The model is pre-trained on three tasks: reconstruction of masked atoms and coordinates, 3D conformation generation conditioned on 2D graph, and 2D graph generation conditioned on 3D conformation. We evaluate our method on 11 downstream molecular property prediction tasks: 7 with 2D information only and 4 with both 2D and 3D information. Our method achieves state-of-the-art results on 10 tasks, and the average improvement on 2D-only tasks is 8.3%. Our method also achieves significant improvement on two 3D conformation generation tasks.

Skip Supplemental Material Section

Supplemental Material

KDD22-fp1497.mp4

mp4

86.1 MB

References

  1. Ravichandra Addanki, Peter W Battaglia, David Budden, Andreea Deac, Jonathan Godwin, Thomas Keck, Wai Lok Sibon Li, Alvaro Sanchez-Gonzalez, Jacklynn Stott, Shantanu Thakoor, et al. 2021. Large-scale graph representation learning with very deep GNNs and self-supervision. arXiv:2107.09422 (2021).Google ScholarGoogle Scholar
  2. Brandon Anderson, Truong Son Hy, and Risi Kondor. 2019Cormorant: Covariant molecular neural networks. NeurIPS, Vol. 32 (2019).Google ScholarGoogle Scholar
  3. Simon Axelrod and Rafael Gomez-Bombarelli. 2020. Geom: Energy-annotated molecular conformations for property prediction and molecular generation. arXiv preprint arXiv:2006.05531 (2020).Google ScholarGoogle Scholar
  4. Dong Chen, Kaifu Gao, Duc Duy Nguyen, Xin Chen, Yi Jiang, Guo-Wei Wei, and Feng Pan. 2021. Algebraic graph-assisted bidirectional transformers for molecular property prediction. Nature Communications, Vol. 12, 1 (2021), 1--9.Google ScholarGoogle Scholar
  5. Seyone Chithrananda, Gabriel Grand, and Bharath Ramsundar. 2020. Chemberta: Large-scale self-supervised pretraining for molecular property prediction. arXiv:2010.09885 (2020).Google ScholarGoogle Scholar
  6. Jacob Devlin, Ming-Wei Chang, Kenton Lee, and Kristina Toutanova. 2019. BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding. In NAACL. 4171--4186.Google ScholarGoogle Scholar
  7. Alexey Dosovitskiy, Lucas Beyer, Alexander Kolesnikov, Dirk Weissenborn, Xiaohua Zhai, Thomas Unterthiner, Mostafa Dehghani, Matthias Minderer, Georg Heigold, Sylvain Gelly, Jakob Uszkoreit, and Neil Houlsby. 2021. An Image is Worth 16x16 Words: Transformers for Image Recognition at Scale. In ICLR.Google ScholarGoogle Scholar
  8. Joseph L. Durant, Burton A. Leland, Douglas R. Henry, and James G. Nourse. 2002. Reoptimization of MDL Keys for Use in Drug Discovery. Journal of Chemical Information and Computer Sciences, Vol. 42, 6 (2002), 1273--1280.Google ScholarGoogle ScholarCross RefCross Ref
  9. Fabian Fuchs, Daniel Worrall, Volker Fischer, and Max Welling. [n.,d.]. SE(3)-Transformers: 3D Roto-Translation Equivariant Attention Networks. In NeurIPS.Google ScholarGoogle Scholar
  10. Kaifu Gao, Duc Duy Nguyen, Vishnu Sresht, Alan M. Mathiowetz, Meihua Tu, and Guo-Wei Wei. 2020. Are 2D fingerprints still valuable for drug discovery? Phys. Chem. Chem. Phys., Vol. 22 (2020), 8373--8390. Issue 16.Google ScholarGoogle ScholarCross RefCross Ref
  11. Justin Gilmer, Samuel S. Schoenholz, Patrick F. Riley, Oriol Vinyals, and George E. Dahl. 2017. Neural Message Passing for Quantum Chemistry. In ICML. 1263--1272.Google ScholarGoogle ScholarDigital LibraryDigital Library
  12. Shion Honda, Shoi Shi, and Hiroki R Ueda. 2019. Smiles transformer: Pre-trained molecular fingerprint for low data drug discovery. arXiv:1911.04738 (2019).Google ScholarGoogle Scholar
  13. Weihua Hu, Matthias Fey, Hongyu Ren, Maho Nakata, Yuxiao Dong, and Jure Leskovec. 2021. OGB-LSC: A Large-Scale Challenge for Machine Learning on Graphs. In Thirty-fifth Conference on Neural Information Processing Systems Datasets and Benchmarks Track (Round 2).Google ScholarGoogle Scholar
  14. Weihua Hu, Matthias Fey, Marinka Zitnik, Yuxiao Dong, Hongyu Ren, Bowen Liu, Michele Catasta, and Jure Leskovec. 2020 a. Open Graph Benchmark: Datasets for Machine Learning on Graphs. arxiv, Vol. abs/2005.00687 (2020).Google ScholarGoogle Scholar
  15. Weihua Hu, Bowen Liu, Joseph Gomes, Marinka Zitnik, Percy Liang, Vijay Pande, and Jure Leskovec. 2020 b. Strategies for Pre-training Graph Neural Networks. In ICLR.Google ScholarGoogle Scholar
  16. Sergey Ioffe and Christian Szegedy. 2015. Batch Normalization: Accelerating Deep Network Training by Reducing Internal Covariate Shift. In ICML, Vol. 37. Lille, France, 448--456.Google ScholarGoogle Scholar
  17. Wengong Jin, Jonathan M. Stokes, Richard T. Eastman, Zina Itkin, Alexey V. Zakharov, James J. Collins, Tommi S. Jaakkola, and Regina Barzilay. 2021. Deep learning identifies synergistic drug combinations for treating COVID-19. Proceedings of the National Academy of Sciences, Vol. 118, 39 (2021).Google ScholarGoogle ScholarCross RefCross Ref
  18. Lukasz Kaiser, Aidan N. Gomez, Noam Shazeer, Ashish Vaswani, Niki Parmar, Llion Jones, and Jakob Uszkoreit. 2017. One Model To Learn Them All. CoRR, Vol. abs/1706.05137 (2017). showeprint[arXiv]1706.05137Google ScholarGoogle Scholar
  19. Charles FF Karney. 2007. Quaternions in molecular modeling. Journal of Molecular Graphics and Modelling, Vol. 25, 5 (2007), 595--604.Google ScholarGoogle ScholarCross RefCross Ref
  20. Johannes Klicpera, Janek Groß, and Stephan Günnemann. 2020. Directional Message Passing for Molecular Graphs. In ICLR .Google ScholarGoogle Scholar
  21. Kezhi Kong, Guohao Li, Mucong Ding, Zuxuan Wu, Chen Zhu, Bernard Ghanem, Gavin Taylor, and Tom Goldstein. 2020. FLAG: Adversarial Data Augmentation for Graph Neural Networks. CoRR, Vol. abs/2010.09891 (2020).Google ScholarGoogle Scholar
  22. Shengchao Liu, Hanchen Wang, Weiyang Liu, Joan Lasenby, Hongyu Guo, and Jian Tang. 2022. Pre-training Molecular Graph Representation with 3D Geometry. In ICLR.Google ScholarGoogle Scholar
  23. Yinhan Liu, Myle Ott, Naman Goyal, Jingfei Du, Mandar Joshi, Danqi Chen, Omer Levy, Mike Lewis, Luke Zettlemoyer, and Veselin Stoyanov. 2019. Roberta: A robustly optimized bert pretraining approach. arXiv:1907.11692 (2019).Google ScholarGoogle Scholar
  24. Chengqiang Lu, Qi Liu, Chao Wang, Zhenya Huang, Peize Lin, and Lixin He. 2019. Molecular property prediction: A multilevel quantum interactions modeling perspective. In AAAI, Vol. 33. 1052--1060.Google ScholarGoogle ScholarDigital LibraryDigital Library
  25. Shitong Luo, Chence Shi, Minkai Xu, and Jian Tang. 2021. Predicting Molecular Conformation via Dynamic Graph Score Matching. In NeurIPS, Vol. 34.Google ScholarGoogle Scholar
  26. Elman Mansimov, Omar Mahmood, Seokho Kang, and Kyunghyun Cho. 2019. Molecular Geometry Prediction using a Deep Generative Graph Neural Network. Scientific Reports, Vol. 9, 1 (31 Dec 2019), 20381.Google ScholarGoogle Scholar
  27. Rocco Meli and Philip C. Biggin. 2020. spyrmsd: symmetry-corrected RMSD calculations in Python. Journal of Cheminformatics, Vol. 12, 1 (31 Aug 2020), 49.Google ScholarGoogle ScholarCross RefCross Ref
  28. Aditya Ramesh, Mikhail Pavlov, Gabriel Goh, Scott Gray, Chelsea Voss, Alec Radford, Mark Chen, and Ilya Sutskever. 2021. Zero-Shot Text-to-Image Generation. In ICML, Vol. 139. 8821--8831.Google ScholarGoogle Scholar
  29. Yu Rong, Yatao Bian, Tingyang Xu, Weiyang Xie, Ying WEI, Wenbing Huang, and Junzhou Huang. 2020. Self-Supervised Graph Transformer on Large-Scale Molecular Data. In NeurIPS, Vol. 33. 12559--12571.Google ScholarGoogle Scholar
  30. Kristof Schütt, Oliver Unke, and Michael Gastegger. 2021. Equivariant message passing for the prediction of tensorial properties and molecular spectra. In ICML.Google ScholarGoogle Scholar
  31. K. T. Schütt, P.-J. Kindermans, H. E. Sauceda, S. Chmiela, A. Tkatchenko, and K.-R. Müller. 2017. SchNet: A Continuous-Filter Convolutional Neural Network for Modeling Quantum Interactions. In NeurIPS (Long Beach, California, USA). Curran Associates Inc., Red Hook, NY, USA, 992--1002.Google ScholarGoogle Scholar
  32. Hari Om Sharan. 2021Artificial Intelligence in Bioinformatics. Springer Singapore, Singapore, 395--403. https://doi.org/10.1007/978-981-33-6191-1_21Google ScholarGoogle Scholar
  33. Chence Shi, Shitong Luo, Minkai Xu, and Jian Tang. 2021. Learning gradient fields for molecular conformation generation. In ICML. PMLR, 9558--9568.Google ScholarGoogle Scholar
  34. Zeren Shui and George Karypis. 2020. Heterogeneous molecular graph neural networks for predicting molecule properties. In ICDM. IEEE, 492--500.Google ScholarGoogle Scholar
  35. Gregor Simm and Jose Miguel Hernandez-Lobato. 2020. A Generative Model for Molecular Distance Geometry. In ICML, Vol. 119. PMLR, 8949--8958.Google ScholarGoogle Scholar
  36. Hannes Stärk, Dominique Beaini, Gabriele Corso, Prudencio Tossou, Christian Dallago, Stephan Günnemann, and Pietro Liò. 2021. 3D Infomax improves GNNs for Molecular Property Prediction. arXiv preprint arXiv:2110.04126 (2021).Google ScholarGoogle Scholar
  37. Hannes Stärk, Dominique Beaini, Gabriele Corso, Prudencio Tossou, Christian Dallago, Stephan Günnemann, and Pietro Liò. 2021. 3D Infomax improves GNNs for Molecular Property Prediction. CoRR, Vol. abs/2110.04126 (2021).Google ScholarGoogle Scholar
  38. Jonathan M Stokes, Kevin Yang, Kyle Swanson, Wengong Jin, Andres Cubillos-Ruiz, Nina M Donghia, Craig R MacNair, Shawn French, Lindsey A Carfrae, Zohar Bloom-Ackermann, et al. 2020. A deep learning approach to antibiotic discovery. Cell, Vol. 180, 4 (2020), 688--702.Google ScholarGoogle ScholarCross RefCross Ref
  39. Igor V. Tetko and Ola Engkvist. 2020. From Big Data to Artificial Intelligence: chemoinformatics meets new challenges. Journal of Cheminformatics, Vol. 12, 1 (18 Dec 2020), 74. https://doi.org/10.1186/s13321-020-00475-yGoogle ScholarGoogle ScholarCross RefCross Ref
  40. Ashish Vaswani, Noam Shazeer, Niki Parmar, Jakob Uszkoreit, Llion Jones, Aidan N Gomez, Łukasz Kaiser, and Illia Polosukhin. 2017. Attention is All you Need. In NeurIPS.Google ScholarGoogle Scholar
  41. Sheng Wang, Yuzhi Guo, Yuhong Wang, Hongmao Sun, and Junzhou Huang. 2019. Smiles-bert: large scale unsupervised pre-training for molecular property prediction. In Proceedings of the 10th ACM international conference on bioinformatics, computational biology and health informatics. 429--436.Google ScholarGoogle ScholarDigital LibraryDigital Library
  42. Yuyang Wang, Jianren Wang, Zhonglin Cao, and Amir Barati Farimani. 2021. MolCLR: molecular contrastive learning of representations via graph neural networks. arXiv preprint arXiv:2102.10056 (2021).Google ScholarGoogle Scholar
  43. David Weininger. 1988. SMILES, a chemical language and information system. 1. Introduction to methodology and encoding rules. Journal of Chemical Information and Computer Sciences, Vol. 28, 1 (1988), 31--36. https://doi.org/10.1021/ci00057a005Google ScholarGoogle ScholarDigital LibraryDigital Library
  44. Kedi Wu and Guo-Wei Wei. 2018. Quantitative Toxicity Prediction Using Topology Based Multitask Deep Neural Networks. Journal of Chemical Information and Modeling, Vol. 58, 2 (2018), 520--531.Google ScholarGoogle ScholarCross RefCross Ref
  45. Zhenqin Wu, Bharath Ramsundar, Evan N Feinberg, Joseph Gomes, Caleb Geniesse, Aneesh S Pappu, Karl Leswing, and Vijay Pande. 2018. MoleculeNet: a benchmark for molecular machine learning. Chemical science, Vol. 9, 2 (2018), 513--530.Google ScholarGoogle Scholar
  46. Keyulu Xu, Weihua Hu, Jure Leskovec, and Stefanie Jegelka. 2019. How Powerful are Graph Neural Networks?. In ICLR 2019.Google ScholarGoogle Scholar
  47. Minkai Xu, Lantao Yu, Yang Song, Chence Shi, Stefano Ermon, and Jian Tang. 2022. GeoDiff: A Geometric Diffusion Model for Molecular Conformation Generation. In ICLR.Google ScholarGoogle Scholar
  48. Chengxuan Ying, Tianle Cai, Shengjie Luo, Shuxin Zheng, Guolin Ke, Di He, Yanming Shen, and Tie-Yan Liu. 2021. Do Transformers Really Perform Badly for Graph Representation?. In Advances in Neural Information Processing Systems.Google ScholarGoogle Scholar
  49. Yuning You, Tianlong Chen, Yongduo Sui, Ting Chen, Zhangyang Wang, and Yang Shen. 2020. Graph Contrastive Learning with Augmentations. In NeurIPS.Google ScholarGoogle Scholar
  50. Hengshuang Zhao, Li Jiang, Jiaya Jia, Philip HS Torr, and Vladlen Koltun. 2021. Point transformer. In ICCV. 16259--16268.Google ScholarGoogle Scholar
  51. Jinhua Zhu, Yingce Xia, Chang Liu, Lijun Wu, Shufang Xie, Tong Wang, Yusong Wang, Wengang Zhou, Tao Qin, Houqiang Li, and Tie-Yan Liu. 2022. Direct molecular conformation generation. arXiv preprint arXiv:2202.01356 (2022).Google ScholarGoogle Scholar
  52. Jinhua Zhu, Yingce Xia, Tao Qin, Wengang Zhou, Houqiang Li, and Tie-Yan Liu. 2021. Dual-view molecule pre-training. arXiv preprint arXiv:2106.10234 (2021).Google ScholarGoogle Scholar

Index Terms

  1. Unified 2D and 3D Pre-Training of Molecular Representations

        Recommendations

        Comments

        Login options

        Check if you have access through your login credentials or your institution to get full access on this article.

        Sign in
        • Published in

          cover image ACM Conferences
          KDD '22: Proceedings of the 28th ACM SIGKDD Conference on Knowledge Discovery and Data Mining
          August 2022
          5033 pages
          ISBN:9781450393850
          DOI:10.1145/3534678

          Copyright © 2022 ACM

          Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than the author(s) must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected].

          Publisher

          Association for Computing Machinery

          New York, NY, United States

          Publication History

          • Published: 14 August 2022

          Permissions

          Request permissions about this article.

          Request Permissions

          Check for updates

          Qualifiers

          • research-article

          Acceptance Rates

          Overall Acceptance Rate1,133of8,635submissions,13%

          Upcoming Conference

          KDD '24

        PDF Format

        View or Download as a PDF file.

        PDF

        eReader

        View online with eReader.

        eReader