Authors
Riccardo Rende, Luciano Loris Viteritti, Lorenzo Bardone, Federico Becca, Sebastian Goldt
Publication date
2024/8/2
Journal
Communications Physics
Volume
7
Issue
1
Pages
260
Publisher
Nature Publishing Group UK
Description
Neural-network architectures have been increasingly used to represent quantum many-body wave functions. These networks require a large number of variational parameters and are challenging to optimize using traditional methods, as gradient descent. Stochastic reconfiguration (SR) has been effective with a limited number of parameters, but becomes impractical beyond a few thousand parameters. Here, we leverage a simple linear algebra identity to show that SR can be employed even in the deep learning scenario. We demonstrate the effectiveness of our method by optimizing a Deep Transformer architecture with 3 × 105 parameters, achieving state-of-the-art ground-state energy in the J1J2 Heisenberg model at J2/J1 = 0.5 on the 10 × 10 square lattice, a challenging benchmark in highly-frustrated magnetism. This work marks a significant step forward in the scalability and efficiency of SR for neural …
Total citations
Scholar articles
R Rende, LL Viteritti, L Bardone, F Becca, S Goldt - Communications Physics, 2024
R Rende, LL Viteritti, L Bardone, F Becca, S Goldt - arXiv preprint arXiv:2310.05715