2023
DOI: 10.1126/sciadv.adf0873
|View full text |Cite
|
Sign up to set email alerts
|

Accurate global machine learning force fields for molecules with hundreds of atoms

Abstract: Global machine learning force fields, with the capacity to capture collective interactions in molecular systems, now scale up to a few dozen atoms due to considerable growth of model complexity with system size. For larger molecules, locality assumptions are introduced, with the consequence that nonlocal interactions are not described. Here, we develop an exact iterative approach to train global symmetric gradient domain machine learning (sGDML) force fields (FFs) for several hundred atoms, without resorting t… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1

Citation Types

0
71
0

Year Published

2023
2023
2024
2024

Publication Types

Select...
6
1
1

Relationship

1
7

Authors

Journals

citations
Cited by 43 publications
(73 citation statements)
references
References 58 publications
0
71
0
Order By: Relevance
“…Unlike the previous datasets which only investigate small organic molecules, MD22 74 and SPICE 73 cover a wider variety of molecule types. In particular, MD22 includes the AIMD trajectories of proteins, carbohydrates, nucleic acids, and supramolecules (i.e., buckyball catcher and nanotube).…”
Section: Datasetsmentioning
confidence: 99%
See 1 more Smart Citation
“…Unlike the previous datasets which only investigate small organic molecules, MD22 74 and SPICE 73 cover a wider variety of molecule types. In particular, MD22 includes the AIMD trajectories of proteins, carbohydrates, nucleic acids, and supramolecules (i.e., buckyball catcher and nanotube).…”
Section: Datasetsmentioning
confidence: 99%
“…SPICE,73 and MD2274 ), which cover different molecular systems from the pre-training datasets.Figure 2a and 2b compare the performance of GNN models with and without denoise pre-training on ISO17 containing isomers of C 7 O 2 H 10 . It is demonstrated that pre-trained GNN models achieve better prediction accuracy when evaluated by both RMSE and MAE.…”
mentioning
confidence: 99%
“…To retain the original PSD structure, we require that a symmetric decomposition P = LL T ( L double-struckR n × n ) exists. ,, Another key requirement is that the construction of the preconditioner P should not dominate the overall computational costs. Generally, there are two principled ways to achieve this: either via sparsification of the kernel matrix (e.g., zero-out entries to obtain a block-diagonal form) or via a compression into a low-rank representation.…”
Section: Scalable Kernel Solvers For Quantum Chemistrymentioning
confidence: 99%
“…In this setting, both neural networks and kernel-based approaches have been successfully applied. Kernel machines are generally considered to be more data efficient in modeling high-quality MLFFs, but yield large linear optimization problems when many training samples and/or large molecule sizes are involved (e.g., in materials ,, or biomolecules). Linear systems are generally solved in closed-form, which has quadratic memory complexity.…”
Section: Introductionmentioning
confidence: 99%
“…These for example include the use of global or non-local descriptors. [18][19][20] In many cases, physical baseline models can also provide the correct long-range physics at affordable computational cost (∆-ML). 16,17,[21][22][23] Finally, message-passing neural networks can extend the range of local interatomic potentials by a multiple of the employed cutoff, though without including the full long-range interactions present in a periodic system.…”
Section: Introductionmentioning
confidence: 99%