diff --git a/bib/bibliography.bib b/bib/bibliography.bib new file mode 100644 index 0000000000000000000000000000000000000000..182782eddc488b8be3834065aad136bdabbf1cfb --- /dev/null +++ b/bib/bibliography.bib @@ -0,0 +1,7747 @@ +@software{AimhubioAim2021, + title = {Aimhubio/Aim}, + date = {2021-05-13T14:14:05Z}, + origdate = {2019-05-31T18:25:07Z}, + url = {https://github.com/aimhubio/aim}, + urldate = {2021-05-13}, + abstract = {Aim — a super-easy way to record, search and compare 1000s of ML training runs}, + organization = {{Aim}}, + keywords = {experiment-tracking,keras,ML,MLOps,nlp,pytorch,reinforcement-learning,tensorflow} +} + +@article{aiOCELOTInfrastructureDatadriven2021, + title = {{{OCELOT}}: {{An}} Infrastructure for Data-Driven Research to Discover and Design Crystalline Organic Semiconductors}, + shorttitle = {{{OCELOT}}}, + author = {Ai, Qianxiang and Bhat, Vinayak and Ryno, Sean M. and Jarolimek, Karol and Sornberger, Parker and Smith, Andrew and Haley, Michael M. and Anthony, John E. and Risko, Chad}, + date = {2021-05-04}, + journaltitle = {The Journal of Chemical Physics}, + shortjournal = {J. Chem. Phys.}, + volume = {154}, + number = {17}, + pages = {174705}, + publisher = {{American Institute of Physics}}, + issn = {0021-9606}, + doi = {10.1063/5.0048714}, + url = {https://aip.scitation.org/doi/10.1063/5.0048714}, + urldate = {2021-05-15}, + abstract = {Materials design and discovery are often hampered by the slow pace and materials and human costs associated with Edisonian trial-and-error screening approaches. Recent advances in computational power, theoretical methods, and data science techniques, however, are being manifest in a convergence of these tools to enable in silico materials discovery. Here, we present the development and deployment of computational materials data and data analytic approaches for crystalline organic semiconductors. The OCELOT (Organic Crystals in Electronic and Light-Oriented Technologies) infrastructure, consisting of a Python-based OCELOT application programming interface and OCELOT database, is designed to enable rapid materials exploration. The database contains a descriptor-based schema for high-throughput calculations that have been implemented on more than 56 000 experimental crystal structures derived from 47 000 distinct molecular structures. OCELOT is open-access and accessible via a web-user interface at https://oscar.as.uky.edu.}, + keywords = {chemistry,descriptors,materials database,molecules,organic chemistry,visualization}, + file = {/home/johannes/Nextcloud/Zotero/Ai et al_2021_OCELOT.pdf;/home/johannes/Zotero/storage/DW64W25V/5.html} +} + +@article{alberi2019MaterialsDesign2018, + title = {The 2019 Materials by Design Roadmap}, + author = {Alberi, Kirstin and Nardelli, Marco Buongiorno and Zakutayev, Andriy and Mitas, Lubos and Curtarolo, Stefano and Jain, Anubhav and Fornari, Marco and Marzari, Nicola and Takeuchi, Ichiro and Green, Martin L. and Kanatzidis, Mercouri and Toney, Mike F. and Butenko, Sergiy and Meredig, Bryce and Lany, Stephan and Kattner, Ursula and Davydov, Albert and Toberer, Eric S. and Stevanovic, Vladan and Walsh, Aron and Park, Nam-Gyu and Aspuru-Guzik, Alán and Tabor, Daniel P. and Nelson, Jenny and Murphy, James and Setlur, Anant and Gregoire, John and Li, Hong and Xiao, Ruijuan and Ludwig, Alfred and Martin, Lane W. and Rappe, Andrew M. and Wei, Su-Huai and Perkins, John}, + date = {2018-10}, + journaltitle = {Journal of Physics D: Applied Physics}, + shortjournal = {J. Phys. D: Appl. Phys.}, + volume = {52}, + number = {1}, + pages = {013001}, + publisher = {{IOP Publishing}}, + issn = {0022-3727}, + doi = {10.1088/1361-6463/aad926}, + url = {https://doi.org/10.1088/1361-6463/aad926}, + urldate = {2022-05-13}, + abstract = {Advances in renewable and sustainable energy technologies critically depend on our ability to design and realize materials with optimal properties. Materials discovery and design efforts ideally involve close coupling between materials prediction, synthesis and characterization. The increased use of computational tools, the generation of materials databases, and advances in experimental methods have substantially accelerated these activities. It is therefore an opportune time to consider future prospects for materials by design approaches. The purpose of this Roadmap is to present an overview of the current state of computational materials prediction, synthesis and characterization approaches, materials design needs for various technologies, and future challenges and opportunities that must be addressed. The various perspectives cover topics on computational techniques, validation, materials databases, materials informatics, high-throughput combinatorial methods, advanced characterization approaches, and materials design issues in thermoelectrics, photovoltaics, solid state lighting, catalysts, batteries, metal alloys, complex oxides and transparent conducting materials. It is our hope that this Roadmap will guide researchers and funding agencies in identifying new prospects for materials design.}, + langid = {english}, + keywords = {DFT,materials database,MD,multiscale,roadmap}, + file = {/home/johannes/Nextcloud/Zotero/Alberi et al_2018_The 2019 materials by design roadmap.pdf} +} + +@article{amorosoInterplaySingleIonTwoIon2021, + title = {Interplay between {{Single-Ion}} and {{Two-Ion Anisotropies}} in {{Frustrated 2D Semiconductors}} and {{Tuning}} of {{Magnetic Structures Topology}}}, + author = {Amoroso, Danila and Barone, Paolo and Picozzi, Silvia}, + date = {2021-08}, + journaltitle = {Nanomaterials}, + volume = {11}, + number = {8}, + pages = {1873}, + publisher = {{Multidisciplinary Digital Publishing Institute}}, + doi = {10.3390/nano11081873}, + url = {https://www.mdpi.com/2079-4991/11/8/1873}, + urldate = {2021-07-28}, + abstract = {The effects of competing magnetic interactions in stabilizing different spin configurations are drawing renewed attention in order to unveil emerging topological spin textures and to highlight microscopic mechanisms leading to their stabilization. The possible key role of the two-site exchange anisotropy in selecting specific helicity and vorticity of skyrmionic lattices has only recently been proposed. In this work, we explore the phase diagram of a frustrated localized magnet characterized by a two-dimensional centrosymmetric triangular lattice, focusing on the interplay between the two-ion anisotropy and the single-ion anisotropy. The effects of an external magnetic field applied perpendicularly to the magnetic layer, are also investigated. By means of Monte Carlo simulations, we find an abundance of different spin configurations, going from trivial to high-order Q skyrmionic and meronic lattices. In closer detail, we find that a dominant role is played by the two-ion over the single-ion anisotropy in determining the planar spin texture; the strength and the sign of single ion anisotropy, together with the magnitude of the magnetic field, tune the perpendicular spin components, mostly affecting the polarity (and, in turn, the topology) of the spin texture. Our analysis confirms the crucial role of the anisotropic symmetric exchange in systems with dominant short-range interactions; at the same time, we predict a rich variety of complex magnetic textures, which may arise from a fine tuning of competing anisotropic mechanisms.}, + issue = {8}, + langid = {english}, + keywords = {atomic scale magnetic properties,magnetic interactions,magnetic nanostructures,pgi-1 seminar,skyrmions,topological spin textures,topology and magnetism}, + file = {/home/johannes/Nextcloud/Zotero/Amoroso et al_2021_Interplay between Single-Ion and Two-Ion Anisotropies in Frustrated 2D.pdf;/home/johannes/Zotero/storage/FN7Y4K4H/htm.html} +} + +@article{andersenOPTIMADEAPIExchanging2021, + title = {{{OPTIMADE}}, an {{API}} for Exchanging Materials Data}, + author = {Andersen, Casper W. and Armiento, Rickard and Blokhin, Evgeny and Conduit, Gareth J. and Dwaraknath, Shyam and Evans, Matthew L. and Fekete, Ãdám and Gopakumar, Abhijith and Gražulis, Saulius and Merkys, Andrius and Mohamed, Fawzi and Oses, Corey and Pizzi, Giovanni and Rignanese, Gian-Marco and Scheidgen, Markus and Talirz, Leopold and Toher, Cormac and Winston, Donald and Aversa, Rossella and Choudhary, Kamal and Colinet, Pauline and Curtarolo, Stefano and Di Stefano, Davide and Draxl, Claudia and Er, Suleyman and Esters, Marco and Fornari, Marco and Giantomassi, Matteo and Govoni, Marco and Hautier, Geoffroy and Hegde, Vinay and Horton, Matthew K. and Huck, Patrick and Huhs, Georg and Hummelshøj, Jens and Kariryaa, Ankit and Kozinsky, Boris and Kumbhar, Snehal and Liu, Mohan and Marzari, Nicola and Morris, Andrew J. and Mostofi, Arash and Persson, Kristin A. and Petretto, Guido and Purcell, Thomas and Ricci, Francesco and Rose, Frisco and Scheffler, Matthias and Speckhard, Daniel and Uhrin, Martin and Vaitkus, Antanas and Villars, Pierre and Waroquiers, David and Wolverton, Chris and Wu, Michael and Yang, Xiaoyu}, + date = {2021-12}, + journaltitle = {Scientific Data}, + shortjournal = {Sci Data}, + volume = {8}, + number = {1}, + eprint = {2103.02068}, + eprinttype = {arxiv}, + pages = {217}, + issn = {2052-4463}, + doi = {10.1038/s41597-021-00974-z}, + url = {http://arxiv.org/abs/2103.02068}, + urldate = {2021-10-15}, + abstract = {The Open Databases Integration for Materials Design (OPTIMADE) consortium has designed a universal application programming interface (API) to make materials databases accessible and interoperable. We outline the first stable release of the specification, v1.0, which is already supported by many leading databases and several software packages. We illustrate the advantages of the OPTIMADE API through worked examples on each of the public materials databases that support the full API specification.}, + archiveprefix = {arXiv}, + keywords = {Condensed Matter - Materials Science}, + file = {/home/johannes/Nextcloud/Zotero/Andersen et al_2021_OPTIMADE, an API for exchanging materials data.pdf;/home/johannes/Zotero/storage/PI7C4VKS/2103.html} +} + +@misc{andersonCormorantCovariantMolecular2019, + title = {Cormorant: {{Covariant Molecular Neural Networks}}}, + shorttitle = {Cormorant}, + author = {Anderson, Brandon and Hy, Truong-Son and Kondor, Risi}, + date = {2019-11-25}, + number = {arXiv:1906.04015}, + eprint = {1906.04015}, + eprinttype = {arxiv}, + primaryclass = {physics, stat}, + publisher = {{arXiv}}, + doi = {10.48550/arXiv.1906.04015}, + url = {http://arxiv.org/abs/1906.04015}, + urldate = {2022-10-04}, + abstract = {We propose Cormorant, a rotationally covariant neural network architecture for learning the behavior and properties of complex many-body physical systems. We apply these networks to molecular systems with two goals: learning atomic potential energy surfaces for use in Molecular Dynamics simulations, and learning ground state properties of molecules calculated by Density Functional Theory. Some of the key features of our network are that (a) each neuron explicitly corresponds to a subset of atoms; (b) the activation of each neuron is covariant to rotations, ensuring that overall the network is fully rotationally invariant. Furthermore, the non-linearity in our network is based upon tensor products and the Clebsch-Gordan decomposition, allowing the network to operate entirely in Fourier space. Cormorant significantly outperforms competing algorithms in learning molecular Potential Energy Surfaces from conformational geometries in the MD-17 dataset, and is competitive with other methods at learning geometric, energetic, electronic, and thermodynamic properties of molecules on the GDB-9 dataset.}, + archiveprefix = {arXiv}, + keywords = {equivariant,GNN,MD17,ML,MLP,MPNN,O(3),QM9,representation learning,SchNet,SO(3)}, + file = {/home/johannes/Nextcloud/Zotero/Anderson et al_2019_Cormorant.pdf;/home/johannes/Zotero/storage/RY359LWP/1906.html} +} + +@article{artrithEfficientAccurateMachinelearning2017, + title = {Efficient and Accurate Machine-Learning Interpolation of Atomic Energies in Compositions with Many Species}, + author = {Artrith, Nongnuch and Urban, Alexander and Ceder, Gerbrand}, + date = {2017-07-21}, + journaltitle = {Physical Review B}, + shortjournal = {Phys. Rev. B}, + volume = {96}, + number = {1}, + pages = {014112}, + publisher = {{American Physical Society}}, + doi = {10.1103/PhysRevB.96.014112}, + url = {https://link.aps.org/doi/10.1103/PhysRevB.96.014112}, + urldate = {2021-10-18}, + abstract = {Machine-learning potentials (MLPs) for atomistic simulations are a promising alternative to conventional classical potentials. Current approaches rely on descriptors of the local atomic environment with dimensions that increase quadratically with the number of chemical species. In this paper, we demonstrate that such a scaling can be avoided in practice. We show that a mathematically simple and computationally efficient descriptor with constant complexity is sufficient to represent transition-metal oxide compositions and biomolecules containing 11 chemical species with a precision of around 3 meV/atom. This insight removes a perceived bound on the utility of MLPs and paves the way to investigate the physics of previously inaccessible materials with more than ten chemical species.}, + file = {/home/johannes/Nextcloud/Zotero/Artrith et al_2017_Efficient and accurate machine-learning interpolation of atomic energies in.pdf;/home/johannes/Zotero/storage/77VRNTN7/Artrith et al. - 2017 - Efficient and accurate machine-learning interpolat.pdf;/home/johannes/Zotero/storage/RL7TSVEA/PhysRevB.96.html} +} + +@article{atzGeometricDeepLearning2021, + title = {Geometric Deep Learning on Molecular Representations}, + author = {Atz, Kenneth and Grisoni, Francesca and Schneider, Gisbert}, + date = {2021-12}, + journaltitle = {Nature Machine Intelligence}, + shortjournal = {Nat Mach Intell}, + volume = {3}, + number = {12}, + pages = {1023--1032}, + publisher = {{Nature Publishing Group}}, + issn = {2522-5839}, + doi = {10.1038/s42256-021-00418-8}, + url = {https://www.nature.com/articles/s42256-021-00418-8}, + urldate = {2022-01-02}, + abstract = {Geometric deep learning (GDL) is based on neural network architectures that incorporate and process symmetry information. GDL bears promise for molecular modelling applications that rely on molecular representations with different symmetry properties and levels of abstraction. This Review provides a structured and harmonized overview of molecular GDL, highlighting its applications in drug discovery, chemical synthesis prediction and quantum chemistry. It contains an introduction to the principles of GDL, as well as relevant molecular representations, such as molecular graphs, grids, surfaces and strings, and their respective properties. The current challenges for GDL in the molecular sciences are discussed, and a forecast of future opportunities is attempted. Geometric representations are becoming more important in molecular deep learning as the spatial structure of molecules contains important information about their properties. Kenneth Atz and colleagues review current progress and challenges in this emerging field of geometric deep learning.}, + issue = {12}, + langid = {english}, + keywords = {CNN,equivariant,GCN,GDL,GNN,invariance,molecules,MPNN,review,review-of-GDL}, + annotation = {Primary\_atype: Reviews Subject\_term: Cheminformatics;Computational models;Computational science Subject\_term\_id: cheminformatics;computational-models;computational-science}, + file = {/home/johannes/Nextcloud/Zotero/Atz et al_2021_Geometric deep learning on molecular representations.pdf;/home/johannes/Zotero/storage/WJWQFR9K/s42256-021-00418-8.html} +} + +@article{bacTopologicalResponseAnomalous2022, + title = {Topological Response of the Anomalous {{Hall}} Effect in {{MnBi2Te4}} Due to Magnetic Canting}, + author = {Bac, S.-K. and Koller, K. and Lux, F. and Wang, J. and Riney, L. and Borisiak, K. and Powers, W. and Zhukovskyi, M. and Orlova, T. and Dobrowolska, M. and Furdyna, J. K. and Dilley, N. R. and Rokhinson, L. P. and Mokrousov, Y. and McQueeney, R. J. and Heinonen, O. and Liu, X. and Assaf, B. A.}, + date = {2022-04-20}, + journaltitle = {npj Quantum Materials}, + shortjournal = {npj Quantum Mater.}, + volume = {7}, + number = {1}, + pages = {1--7}, + publisher = {{Nature Publishing Group}}, + issn = {2397-4648}, + doi = {10.1038/s41535-022-00455-5}, + url = {https://www.nature.com/articles/s41535-022-00455-5}, + urldate = {2022-06-01}, + abstract = {Three-dimensional (3D) compensated MnBi2Te4 is antiferromagnetic, but undergoes a spin-flop transition at intermediate fields, resulting in a canted phase before saturation. In this work, we experimentally show that the anomalous Hall effect (AHE) in MnBi2Te4 originates from a topological response that is sensitive to the perpendicular magnetic moment and to its canting angle. Synthesis by molecular beam epitaxy allows us to obtain a large-area quasi-3D 24-layer MnBi2Te4 with near-perfect compensation that hosts the phase diagram observed in bulk which we utilize to probe the AHE. This AHE is seen to exhibit an antiferromagnetic response at low magnetic fields, and a clear evolution at intermediate fields through surface and bulk spin-flop transitions into saturation. Throughout this evolution, the AHE is super-linear versus magnetization rather than the expected linear relationship. We reveal that this discrepancy is related to the canting angle, consistent with the symmetry of the crystal. Our findings bring to light a topological anomalous Hall response that can be found in non-collinear ferromagnetic, and antiferromagnetic phases.}, + issue = {1}, + langid = {english}, + keywords = {Ferromagnetism,Magnetic properties and materials,Topological insulators}, + file = {/home/johannes/Nextcloud/Zotero/Bac et al_2022_Topological response of the anomalous Hall effect in MnBi2Te4 due to magnetic.pdf;/home/johannes/Nextcloud/Zotero/Bac et al_2022_Topological response of the anomalous Hall effect in MnBi2Te4 due to magnetic2_supplementary.pdf;/home/johannes/Zotero/storage/E6I5UGGJ/s41535-022-00455-5.html} +} + +@article{bankoFastTrackResearchData2020, + title = {Fast-{{Track}} to {{Research Data Management}} in {{Experimental Material Science}}–{{Setting}} the {{Ground}} for {{Research Group Level Materials Digitalization}}}, + author = {Banko, Lars and Ludwig, Alfred}, + date = {2020-08-10}, + journaltitle = {ACS Combinatorial Science}, + shortjournal = {ACS Comb. Sci.}, + volume = {22}, + number = {8}, + pages = {401--409}, + publisher = {{American Chemical Society}}, + issn = {2156-8952}, + doi = {10.1021/acscombsci.0c00057}, + url = {https://doi.org/10.1021/acscombsci.0c00057}, + urldate = {2022-05-12}, + abstract = {Research data management is a major necessity for the digital transformation in material science. Material science is multifaceted and experimental data, especially, is highly diverse. We demonstrate an adjustable approach to a group level data management based on a customizable document management software. Our solution is to continuously transform data management workflows from generalized to specialized data management. We start up fast with a relatively unregulated base setting and adapt continuously over the period of use to transform more and more data procedures into specialized data management workflows. By continuous adaptation and integration of analysis workflows and metadata schemes, the amount and the quality of the data improves. As an example of this process, in a period of 36 months, data on over 1800 samples, mainly materials libraries with hundreds of individual samples, were collected. The research data management system now contains over 1700 deposition processes and more than 4000 characterization documents. From initially mainly user-defined data input, an increased number of specialized data processing workflows was developed allowing the collection of more specialized, quality-assured data sets.}, + keywords = {experimental science,RDM}, + file = {/home/johannes/Nextcloud/Zotero/Banko_Ludwig_2020_Fast-Track to Research Data Management in Experimental Material Science–Setting.pdf;/home/johannes/Zotero/storage/7HEKG4XK/acscombsci.html} +} + +@article{barthLocalExchangecorrelationPotential1972, + title = {A Local Exchange-Correlation Potential for the Spin Polarized Case. i}, + author = {von Barth, U. and Hedin, L.}, + date = {1972-07}, + journaltitle = {Journal of Physics C: Solid State Physics}, + shortjournal = {J. Phys. C: Solid State Phys.}, + volume = {5}, + number = {13}, + pages = {1629--1642}, + publisher = {{IOP Publishing}}, + issn = {0022-3719}, + doi = {10.1088/0022-3719/5/13/012}, + url = {https://doi.org/10.1088/0022-3719/5/13/012}, + urldate = {2021-10-13}, + abstract = {The local density theory is developed by Hohenberg, Kohn and Sham is extended to the spin polarized case. A spin dependent one- electron potential pertinent to ground state properties is obtained from calculations of the total energy per electron made with a 'bubble' (or random phase) type of dielectric function. The potential is found to be well represented by an analytic expression corresponding to a shifted and rescaled spin dependent Slater potential. To test this potential the momentum dependent spin susceptibility of an electron gas is calculated. The results compare favourably with available information from other calculations and from experiment. The potential obtained in this paper should be useful for split band calculations of magnetic materials.}, + langid = {english}, + keywords = {DFT,LDA,LSDA,original publication,xc functional}, + file = {/home/johannes/Nextcloud/Zotero/Barth_Hedin_1972_A local exchange-correlation potential for the spin polarized case.pdf} +} + +@book{bartok-partayGaussianApproximationPotential2010, + title = {The {{Gaussian Approximation Potential}}}, + author = {BartÏŒk-Pártay, Albert}, + date = {2010}, + series = {Springer {{Theses}}}, + publisher = {{Springer Berlin Heidelberg}}, + location = {{Berlin, Heidelberg}}, + doi = {10.1007/978-3-642-14067-9}, + url = {http://link.springer.com/10.1007/978-3-642-14067-9}, + urldate = {2021-07-06}, + isbn = {978-3-642-14066-2 978-3-642-14067-9}, + langid = {english}, + keywords = {GAP,GPR,ML,models,original publication}, + file = {/home/johannes/Books/scientific_machine_learning/Bartók-Pártay_The Gaussian Approximation Potential_thesis-2010.pdf} +} + +@article{bartokGaussianApproximationPotentials2010, + title = {Gaussian {{Approximation Potentials}}: {{The Accuracy}} of {{Quantum Mechanics}}, without the {{Electrons}}}, + shorttitle = {Gaussian {{Approximation Potentials}}}, + author = {Bartók, Albert P. and Payne, Mike C. and Kondor, Risi and Csányi, Gábor}, + date = {2010-04-01}, + journaltitle = {Physical Review Letters}, + shortjournal = {Phys. Rev. Lett.}, + volume = {104}, + number = {13}, + pages = {136403}, + publisher = {{American Physical Society}}, + doi = {10.1103/PhysRevLett.104.136403}, + url = {https://link.aps.org/doi/10.1103/PhysRevLett.104.136403}, + urldate = {2021-07-06}, + abstract = {We introduce a class of interatomic potential models that can be automatically generated from data consisting of the energies and forces experienced by atoms, as derived from quantum mechanical calculations. The models do not have a fixed functional form and hence are capable of modeling complex potential energy landscapes. They are systematically improvable with more data. We apply the method to bulk crystals, and test it by calculating properties at high temperatures. Using the interatomic potential to generate the long molecular dynamics trajectories required for such calculations saves orders of magnitude in computational cost.}, + keywords = {GAP,ML,models,original publication}, + file = {/home/johannes/Nextcloud/Zotero/Bartók et al_2010_Gaussian Approximation Potentials.pdf;/home/johannes/Zotero/storage/DQIZDC4R/Bartók et al. - 2010 - Gaussian Approximation Potentials The Accuracy of.pdf;/home/johannes/Zotero/storage/QQUERR3G/PhysRevLett.104.html} +} + +@unpublished{bartokGaussianApproximationPotentials2020, + title = {Gaussian {{Approximation Potentials}}: A Brief Tutorial Introduction}, + shorttitle = {Gaussian {{Approximation Potentials}}}, + author = {Bartók, Albert P. and Csányi, Gábor}, + date = {2020-02-05}, + eprint = {1502.01366}, + eprinttype = {arxiv}, + primaryclass = {cond-mat, physics:physics}, + url = {http://arxiv.org/abs/1502.01366}, + urldate = {2021-07-06}, + abstract = {We present a swift walk-through of our recent work that uses machine learning to fit interatomic potentials based on quantum mechanical data. We describe our Gaussian Approximation Potentials (GAP) framework, discussing a variety of descriptors, how to train the model on total energies and derivatives and the simultaneous use of multiple models. We also show a small example using QUIP, the software sandbox implementation of GAP that is available for non-commercial use.}, + archiveprefix = {arXiv}, + keywords = {_tablet,GAP,ML,models,QUIP,SOAP,tutorial}, + file = {/home/johannes/Nextcloud/Zotero/Bartók_Csányi_2020_Gaussian Approximation Potentials.pdf;/home/johannes/Zotero/storage/SBML3RKM/1502.html} +} + +@article{bartokMachineLearningUnifies2017, + title = {Machine Learning Unifies the Modeling of Materials and Molecules}, + author = {Bartók, Albert P. and De, Sandip and Poelking, Carl and Bernstein, Noam and Kermode, James R. and Csányi, Gábor and Ceriotti, Michele}, + date = {2017-12}, + journaltitle = {Science Advances}, + publisher = {{American Association for the Advancement of Science}}, + doi = {10.1126/sciadv.1701816}, + url = {https://www.science.org/doi/10.1126/sciadv.1701816}, + urldate = {2022-10-03}, + abstract = {Statistical learning based on a local representation of atomic structures provides a universal model of chemical stability.}, + langid = {english}, + keywords = {CCSD(T),coupled cluster,DFT,GAP,GGA,kernel methods,ML,molecules,silicon,SOAP,solids,surface physics}, + file = {/home/johannes/Nextcloud/Zotero/Bartók et al_2017_Machine learning unifies the modeling of materials and molecules.pdf;/home/johannes/Zotero/storage/DZL84DP7/sciadv.html} +} + +@article{bartokRepresentingChemicalEnvironments2013, + title = {On Representing Chemical Environments}, + author = {Bartók, Albert P.}, + date = {2013}, + journaltitle = {Physical Review B}, + shortjournal = {Phys. Rev. B}, + volume = {87}, + number = {18}, + doi = {10.1103/PhysRevB.87.184115}, + keywords = {descriptors,ML,original publication,SOAP}, + file = {/home/johannes/Nextcloud/Zotero/Bartók_2013_On representing chemical environments.pdf;/home/johannes/Zotero/storage/VRNA6FAC/PhysRevB.87.html} +} + +@article{basovPropertiesDemandQuantum2017, + title = {Towards Properties on Demand in Quantum Materials}, + author = {Basov, D. N. and Averitt, R. D. and Hsieh, D.}, + date = {2017-11}, + journaltitle = {Nature Materials}, + shortjournal = {Nature Mater}, + volume = {16}, + number = {11}, + pages = {1077--1088}, + publisher = {{Nature Publishing Group}}, + issn = {1476-4660}, + doi = {10.1038/nmat5017}, + url = {https://www.nature.com/articles/nmat5017}, + urldate = {2021-08-24}, + abstract = {The past decade has witnessed an explosion in the field of quantum materials, headlined by the predictions and discoveries of novel Landau-symmetry-broken phases in correlated electron systems, topological phases in systems with strong spin–orbit coupling, and ultra-manipulable materials platforms based on two-dimensional van der Waals crystals. Discovering pathways to experimentally realize quantum phases of matter and exert control over their properties is a central goal of modern condensed-matter physics, which holds promise for a new generation of electronic/photonic devices with currently inaccessible and likely unimaginable functionalities. In this Review, we describe emerging strategies for selectively perturbing microscopic interaction parameters, which can be used to transform materials into a desired quantum state. Particular emphasis will be placed on recent successes to tailor electronic interaction parameters through the application of intense fields, impulsive electromagnetic stimulation, and nanostructuring or interface engineering. Together these approaches outline a potential roadmap to an era of quantum phenomena on demand.}, + issue = {11}, + langid = {english}, + annotation = {Bandiera\_abtest: a Cg\_type: Nature Research Journals Primary\_atype: Reviews Subject\_term: Electronic properties and materials;Phase transitions and critical phenomena Subject\_term\_id: electronic-properties-and-materials;phase-transitions-and-critical-phenomena}, + file = {/home/johannes/Nextcloud/Zotero/Basov et al_2017_Towards properties on demand in quantum materials.pdf} +} + +@unpublished{batatiaDesignSpaceEquivariant2022, + title = {The {{Design Space}} of {{E}}(3)-{{Equivariant Atom-Centered Interatomic Potentials}}}, + author = {Batatia, Ilyes and Batzner, Simon and Kovács, Dávid Péter and Musaelian, Albert and Simm, Gregor N. C. and Drautz, Ralf and Ortner, Christoph and Kozinsky, Boris and Csányi, Gábor}, + date = {2022-05-13}, + number = {arXiv:2205.06643}, + eprint = {2205.06643}, + eprinttype = {arxiv}, + primaryclass = {cond-mat, physics:physics, stat}, + publisher = {{arXiv}}, + doi = {10.48550/arXiv.2205.06643}, + url = {http://arxiv.org/abs/2205.06643}, + urldate = {2022-05-21}, + abstract = {The rapid progress of machine learning interatomic potentials over the past couple of years produced a number of new architectures. Particularly notable among these are the Atomic Cluster Expansion (ACE), which unified many of the earlier ideas around atom density-based descriptors, and Neural Equivariant Interatomic Potentials (NequIP), a message passing neural network with equivariant features that showed state of the art accuracy. In this work, we construct a mathematical framework that unifies these models: ACE is generalised so that it can be recast as one layer of a multi-layer architecture. From another point of view, the linearised version of NequIP is understood as a particular sparsification of a much larger polynomial model. Our framework also provides a practical tool for systematically probing different choices in the unified design space. We demonstrate this by an ablation study of NequIP via a set of experiments looking at in- and out-of-domain accuracy and smooth extrapolation very far from the training data, and shed some light on which design choices are critical for achieving high accuracy. Finally, we present BOTNet (Body-Ordered-Tensor-Network), a much-simplified version of NequIP, which has an interpretable architecture and maintains accuracy on benchmark datasets.}, + archiveprefix = {arXiv}, + keywords = {_tablet,ACE,BOTNet,descriptors,equivariant,GNN,ML,MLP,MPNN,NequIP,NN,unified theory}, + file = {/home/johannes/Nextcloud/Zotero/Batatia et al_2022_The Design Space of E(3)-Equivariant Atom-Centered Interatomic Potentials.pdf;/home/johannes/Zotero/storage/2FLTPTA2/2205.html} +} + +@misc{batatiaMACEHigherOrder2022, + title = {{{MACE}}: {{Higher Order Equivariant Message Passing Neural Networks}} for {{Fast}} and {{Accurate Force Fields}}}, + shorttitle = {{{MACE}}}, + author = {Batatia, Ilyes and Kovács, Dávid Péter and Simm, Gregor N. C. and Ortner, Christoph and Csányi, Gábor}, + date = {2022-06-15}, + number = {arXiv:2206.07697}, + eprint = {2206.07697}, + eprinttype = {arxiv}, + primaryclass = {cond-mat, physics:physics, stat}, + publisher = {{arXiv}}, + doi = {10.48550/arXiv.2206.07697}, + url = {http://arxiv.org/abs/2206.07697}, + urldate = {2022-09-25}, + abstract = {Creating fast and accurate force fields is a long-standing challenge in computational chemistry and materials science. Recently, several equivariant message passing neural networks (MPNNs) have been shown to outperform models built using other approaches in terms of accuracy. However, most MPNNs suffer from high computational cost and poor scalability. We propose that these limitations arise because MPNNs only pass two-body messages leading to a direct relationship between the number of layers and the expressivity of the network. In this work, we introduce MACE, a new equivariant MPNN model that uses higher body order messages. In particular, we show that using four-body messages reduces the required number of message passing iterations to just \textbackslash emph\{two\}, resulting in a fast and highly parallelizable model, reaching or exceeding state-of-the-art accuracy on the rMD17, 3BPA, and AcAc benchmark tasks. We also demonstrate that using higher order messages leads to an improved steepness of the learning curves.}, + archiveprefix = {arXiv}, + keywords = {_tablet,ACE,chemical species scaling problem,descriptors,equivariant,library,MACE,ML,MLP,models,MPNN,Multi-ACE,NequIP,original publication,unified theory,with-code}, + file = {/home/johannes/Nextcloud/Zotero/Batatia et al_2022_MACE.pdf;/home/johannes/Zotero/storage/LDAKZMRF/2206.html} +} + +@article{batraEmergingMaterialsIntelligence2020, + title = {Emerging Materials Intelligence Ecosystems Propelled by Machine Learning}, + author = {Batra, Rohit and Song, Le and Ramprasad, Rampi}, + date = {2020-11-09}, + journaltitle = {Nature Reviews Materials}, + pages = {1--24}, + publisher = {{Nature Publishing Group}}, + issn = {2058-8437}, + doi = {10.1038/s41578-020-00255-y}, + url = {https://www.nature.com/articles/s41578-020-00255-y}, + urldate = {2021-05-19}, + abstract = {The age of cognitive computing and artificial intelligence (AI) is just dawning. Inspired by its successes and promises, several AI ecosystems are blossoming, many of them within the domain of materials science and engineering. These materials intelligence ecosystems are being shaped by several independent developments. Machine learning (ML) algorithms and extant materials data are utilized to create surrogate models of materials properties and performance predictions. Materials data repositories, which fuel such surrogate model development, are mushrooming. Automated data and knowledge capture from the literature (to populate data repositories) using natural language processing approaches is being explored. The design of materials that meet target property requirements and of synthesis steps to create target materials appear to be within reach, either by closed-loop active-learning strategies or by inverting the prediction pipeline using advanced generative algorithms. AI and ML concepts are also transforming the computational and physical laboratory infrastructural landscapes used to create materials data in the first place. Surrogate models that can outstrip physics-based simulations (on which they are trained) by several orders of magnitude in speed while preserving accuracy are being actively developed. Automation, autonomy and guided high-throughput techniques are imparting enormous efficiencies and eliminating redundancies in materials synthesis and characterization. The integration of the various parts of the burgeoning ML landscape may lead to materials-savvy digital assistants and to a human–machine partnership that could enable dramatic efficiencies, accelerated discoveries and increased productivity. Here, we review these emergent materials intelligence ecosystems and discuss the imminent challenges and opportunities.}, + langid = {english}, + keywords = {materials informatics}, + file = {/home/johannes/Nextcloud/Zotero/Batra et al_2020_Emerging materials intelligence ecosystems propelled by machine learning.pdf;/home/johannes/Zotero/storage/A3A6TGKC/s41578-020-00255-y.html} +} + +@unpublished{batznerEquivariantGraphNeural2021, + title = {E(3)-{{Equivariant Graph Neural Networks}} for {{Data-Efficient}} and {{Accurate Interatomic Potentials}}}, + author = {Batzner, Simon and Musaelian, Albert and Sun, Lixin and Geiger, Mario and Mailoa, Jonathan P. and Kornbluth, Mordechai and Molinari, Nicola and Smidt, Tess E. and Kozinsky, Boris}, + date = {2021-12-16}, + eprint = {2101.03164}, + eprinttype = {arxiv}, + primaryclass = {cond-mat, physics:physics}, + url = {http://arxiv.org/abs/2101.03164}, + urldate = {2022-01-02}, + abstract = {This work presents Neural Equivariant Interatomic Potentials (NequIP), an E(3)-equivariant neural network approach for learning interatomic potentials from ab-initio calculations for molecular dynamics simulations. While most contemporary symmetry-aware models use invariant convolutions and only act on scalars, NequIP employs E(3)-equivariant convolutions for interactions of geometric tensors, resulting in a more information-rich and faithful representation of atomic environments. The method achieves state-of-the-art accuracy on a challenging and diverse set of molecules and materials while exhibiting remarkable data efficiency. NequIP outperforms existing models with up to three orders of magnitude fewer training data, challenging the widely held belief that deep neural networks require massive training sets. The high data efficiency of the method allows for the construction of accurate potentials using high-order quantum chemical level of theory as reference and enables high-fidelity molecular dynamics simulations over long time scales.}, + archiveprefix = {arXiv}, + version = {3}, + keywords = {_tablet,GNN,MD,ML,MLP,molecules,MPNN,NequIP,Neural networks,Physics - Computational Physics,solids}, + file = {/home/johannes/Nextcloud/Zotero/Batzner et al_2021_E(3)-Equivariant Graph Neural Networks for Data-Efficient and Accurate.pdf;/home/johannes/Nextcloud/Zotero/false;/home/johannes/Zotero/storage/85ATGPNR/s41467-022-29939-5.html;/home/johannes/Zotero/storage/V4Y8BWNW/2101.html} +} + +@thesis{bauerDevelopmentRelativisticFullpotential2014, + title = {Development of a Relativistic Full-Potential First-Principles Multiple Scattering {{Green}} Function Method Applied to Complex Magnetic Textures of Nano Structures at Surfaces}, + author = {Bauer, David Siegfried Georg}, + date = {2014}, + number = {FZJ-2014-01052}, + institution = {{Forschungszentrum Jülich GmbH Zentralbibliothek, Verlag}}, + url = {http://hdl.handle.net/2128/5899}, + urldate = {2022-08-12}, + abstract = {This thesis is concerned with the quantum mechanical investigation of a novel class of magnetic phenomena in atomic- and nanoscale-sized systems deposited on surfaces or embedded in bulk materials that result from a competition between the exchange and the relativistic spin-orbit interactions. The thesis is motivated by the observation of novel spin-textures of one- and two-dimensional periodicity of nanoscale pitchlength exhibiting a unique winding sense observed in ultra-thin magnetic lms on nonmagnetic metallic substrates with a large spin-orbit interaction. The goal is to extend this eld to magnetic clusters and nano-structures of nite size in order to investigate in how far the size of the cluster and the atoms at the edge of the cluster or ribbon that are particular susceptible to relativistic eects change the balance betweendierent interactions and thus lead to new magnetic phenomena. As an example, the challenging problem of Fe nano-islands on Ir(111) is addressed in detail as for an Fe monolayer on Ir(111) a magnetic nanoskyrmion lattice was observed as magnetic structure.To achieve this goal a new rst-principles all-electron electronic structure code based on density functional theory was developed. The method of choice is the Korringa-Kohn-Rostoker (KKR) impurity Green function method, resorting on a multiple scattering approach. This method has been conceptually further advanced to combine the neglect of any shape approximation to the full potential, with the treatment ofnon-collinear magnetism, of the spin-orbit interaction, as well as of the structural relaxation together with the perfect embedding of a nite size magnetic cluster of atoms into a surface or a bulk environment. For this purpose the formalism makes use of an expansion of the Green function involving explicitly left- and right-hand side scattering solutions. Relativistic eects are treated via the scalar-relativistic approximation and a spin-orbit coupling term treated self-consistently. This required the development of a new algorithm to solve the relativistic quantum mechanical scattering problem for a single atom with a non-spherical potential formulated in terms of the Lippmann-Schwinger integral equation. Prior to the investigation of the Fe nano-islands, the magnetic structure of an Fe monolayer is studied using atomistic spin-dynamics on the basis of a classical model Hamiltonian, which uses realistic coupling parameters obtained from rst principles. It is shown that this method is capable to nd the experimentally determined magnetic structure. [...] Bauer, David Siegfried Georg}, + isbn = {9783893369348}, + langid = {english}, + keywords = {_tablet,juKKR,KKR,KKRimp,PGI-1/IAS-1,thesis}, + file = {/home/johannes/Nextcloud/Zotero/Bauer_2014_Development of a relativistic full-potential first-principles multiple.pdf;/home/johannes/Zotero/storage/SYS2ZV93/151022.html} +} + +@article{beckeDensityfunctionalTheoryVs2022, + title = {Density-Functional Theory vs Density-Functional Fits: {{The}} Best of Both}, + shorttitle = {Density-Functional Theory vs Density-Functional Fits}, + author = {Becke, Axel D.}, + date = {2022-12-21}, + journaltitle = {The Journal of Chemical Physics}, + shortjournal = {J. Chem. Phys.}, + volume = {157}, + number = {23}, + pages = {234102}, + publisher = {{American Institute of Physics}}, + issn = {0021-9606}, + doi = {10.1063/5.0128996}, + url = {https://aip.scitation.org/doi/10.1063/5.0128996}, + urldate = {2022-12-29}, + abstract = {In a recent paper [A. D. Becke, J. Chem. Phys. 156, 214101 (2022)], we compared two Kohn–Sham density functionals based on physical modeling and theory with the best density-functional power series fits in the literature. With only a handful of physically motivated pre-factors, our functionals matched, and even slightly exceeded, the performance of the best power-series functionals on the general main group thermochemistry, kinetics, and noncovalent interactions (GMTKN55) chemical database of Goerigk et al. [Phys. Chem. Chem. Phys. 19, 32184 (2017)]. This begs the question: how much can their performance be improved by adding power-series terms of our own? We address this question in the present work. First, we describe a series expansion variable that we believe contains more local physics than any other variable considered to date. Then we undertake modest, one-dimensional fits to the GMTKN55 data with our theory-based functional corrected by power-series exchange and dynamical correlation terms. We settle on 12 power-series terms (plus six parent terms) and achieve the lowest GMTKN55 “WTMAD2†error yet reported, by a substantial margin, for a hybrid Kohn–Sham density functional. The new functional is called “B22plus.â€}, + keywords = {/unread,B22,B22plus,DFA,DFT,DM21}, + file = {/home/johannes/Nextcloud/Zotero/Becke_2022_Density-functional theory vs density-functional fits.pdf} +} + +@article{behlerAtomcenteredSymmetryFunctions2011, + title = {Atom-Centered Symmetry Functions for Constructing High-Dimensional Neural Network Potentials}, + author = {Behler, Jörg}, + date = {2011-02-16}, + journaltitle = {The Journal of Chemical Physics}, + shortjournal = {J. Chem. Phys.}, + volume = {134}, + number = {7}, + pages = {074106}, + publisher = {{American Institute of Physics}}, + issn = {0021-9606}, + doi = {10.1063/1.3553717}, + url = {https://aip.scitation.org/doi/full/10.1063/1.3553717}, + urldate = {2021-05-18}, + abstract = {Neural networks offer an unbiased and numerically very accurate approach to represent high-dimensional ab initio potential-energy surfaces. Once constructed, neural network potentials can provide the energies and forces many orders of magnitude faster than electronic structure calculations, and thus enable molecular dynamics simulations of large systems. However, Cartesian coordinates are not a good choice to represent the atomic positions, and a transformation to symmetry functions is required. Using simple benchmark systems, the properties of several types of symmetry functions suitable for the construction of high-dimensional neural network potential-energy surfaces are discussed in detail. The symmetry functions are general and can be applied to all types of systems such as molecules, crystalline and amorphous solids, and liquids.}, + keywords = {ACSF,descriptors,HDNNP,ML,models,original publication}, + file = {/home/johannes/Zotero/storage/HVL86IPH/1.html} +} + +@article{behlerConstructingHighdimensionalNeural2015, + title = {Constructing High-Dimensional Neural Network Potentials: {{A}} Tutorial Review}, + shorttitle = {Constructing High-Dimensional Neural Network Potentials}, + author = {Behler, Jörg}, + date = {2015}, + journaltitle = {International Journal of Quantum Chemistry}, + volume = {115}, + number = {16}, + pages = {1032--1050}, + issn = {1097-461X}, + doi = {10.1002/qua.24890}, + url = {https://onlinelibrary.wiley.com/doi/abs/10.1002/qua.24890}, + urldate = {2021-05-18}, + abstract = {A lot of progress has been made in recent years in the development of atomistic potentials using machine learning (ML) techniques. In contrast to most conventional potentials, which are based on physical approximations and simplifications to derive an analytic functional relation between the atomic configuration and the potential-energy, ML potentials rely on simple but very flexible mathematical terms without a direct physical meaning. Instead, in case of ML potentials the topology of the potential-energy surface is “learned†by adjusting a number of parameters with the aim to reproduce a set of reference electronic structure data as accurately as possible. Due to this bias-free construction, they are applicable to a wide range of systems without changes in their functional form, and a very high accuracy close to the underlying first-principles data can be obtained. Neural network potentials (NNPs), which have first been proposed about two decades ago, are an important class of ML potentials. Although the first NNPs have been restricted to small molecules with only a few degrees of freedom, they are now applicable to high-dimensional systems containing thousands of atoms, which enables addressing a variety of problems in chemistry, physics, and materials science. In this tutorial review, the basic ideas of NNPs are presented with a special focus on developing NNPs for high-dimensional condensed systems. A recipe for the construction of these potentials is given and remaining limitations of the method are discussed. © 2015 Wiley Periodicals, Inc.}, + langid = {english}, + keywords = {_tablet,HDNNP,ML,models,molecular dynamics,neural network potentials,review,tutorial}, + annotation = {\_eprint: https://onlinelibrary.wiley.com/doi/pdf/10.1002/qua.24890}, + file = {/home/johannes/Nextcloud/Zotero/Behler_2015_Constructing high-dimensional neural network potentials.pdf;/home/johannes/Zotero/storage/DQEEE6BV/qua.html} +} + +@article{behlerFourGenerationsHighDimensional2021, + title = {Four {{Generations}} of {{High-Dimensional Neural Network Potentials}}}, + author = {Behler, Jörg}, + date = {2021-03-29}, + journaltitle = {Chemical Reviews}, + shortjournal = {Chem. Rev.}, + publisher = {{American Chemical Society}}, + issn = {0009-2665}, + doi = {10.1021/acs.chemrev.0c00868}, + url = {https://doi.org/10.1021/acs.chemrev.0c00868}, + urldate = {2021-05-18}, + abstract = {Since their introduction about 25 years ago, machine learning (ML) potentials have become an important tool in the field of atomistic simulations. After the initial decade, in which neural networks were successfully used to construct potentials for rather small molecular systems, the development of high-dimensional neural network potentials (HDNNPs) in 2007 opened the way for the application of ML potentials in simulations of large systems containing thousands of atoms. To date, many other types of ML potentials have been proposed continuously increasing the range of problems that can be studied. In this review, the methodology of the family of HDNNPs including new recent developments will be discussed using a classification scheme into four generations of potentials, which is also applicable to many other types of ML potentials. The first generation is formed by early neural network potentials designed for low-dimensional systems. High-dimensional neural network potentials established the second generation and are based on three key steps: first, the expression of the total energy as a sum of environment-dependent atomic energy contributions; second, the description of the atomic environments by atom-centered symmetry functions as descriptors fulfilling the requirements of rotational, translational, and permutation invariance; and third, the iterative construction of the reference electronic structure data sets by active learning. In third-generation HDNNPs, in addition, long-range interactions are included employing environment-dependent partial charges expressed by atomic neural networks. In fourth-generation HDNNPs, which are just emerging, in addition, nonlocal phenomena such as long-range charge transfer can be included. The applicability and remaining limitations of HDNNPs are discussed along with an outlook at possible future developments.}, + keywords = {HDNNP,ML,MLP,models,review,review-of-MLP}, + file = {/home/johannes/Nextcloud/Zotero/Behler_2021_Four Generations of High-Dimensional Neural Network Potentials.pdf} +} + +@article{behlerGeneralizedNeuralNetworkRepresentation2007, + title = {Generalized {{Neural-Network Representation}} of {{High-Dimensional Potential-Energy Surfaces}}}, + author = {Behler, Jörg}, + date = {2007}, + journaltitle = {Physical Review Letters}, + shortjournal = {Phys. Rev. Lett.}, + volume = {98}, + number = {14}, + doi = {10.1103/PhysRevLett.98.146401}, + keywords = {BPNN,MD,ML,MLP,models,NN,original publication}, + file = {/home/johannes/Zotero/storage/RNTYUSXX/PhysRevLett.98.html} +} + +@article{behlerPerspectiveMachineLearning2016, + title = {Perspective: {{Machine}} Learning Potentials for Atomistic Simulations}, + shorttitle = {Perspective}, + author = {Behler, Jörg}, + date = {2016-11-01}, + journaltitle = {The Journal of Chemical Physics}, + shortjournal = {J. Chem. Phys.}, + volume = {145}, + number = {17}, + pages = {170901}, + publisher = {{American Institute of Physics}}, + issn = {0021-9606}, + doi = {10.1063/1.4966192}, + url = {https://aip.scitation.org/doi/full/10.1063/1.4966192}, + urldate = {2021-05-18}, + abstract = {Nowadays, computer simulations have become a standard tool in essentially all fields of chemistry, condensed matter physics, and materials science. In order to keep up with state-of-the-art experiments and the ever growing complexity of the investigated problems, there is a constantly increasing need for simulations of more realistic, i.e., larger, model systems with improved accuracy. In many cases, the availability of sufficiently efficient interatomic potentials providing reliable energies and forces has become a serious bottleneck for performing these simulations. To address this problem, currently a paradigm change is taking place in the development of interatomic potentials. Since the early days of computer simulations simplified potentials have been derived using physical approximations whenever the direct application of electronic structure methods has been too demanding. Recent advances in machine learning (ML) now offer an alternative approach for the representation of potential-energy surfaces by fitting large data sets from electronic structure calculations. In this perspective, the central ideas underlying these ML potentials, solved problems and remaining challenges are reviewed along with a discussion of their current applicability and limitations.}, + keywords = {ML,MLP,models,review}, + file = {/home/johannes/Nextcloud/Zotero/Behler_2016_Perspective.pdf} +} + +@article{bengioRepresentationLearningReview2013, + title = {Representation {{Learning}}: {{A Review}} and {{New Perspectives}}}, + shorttitle = {Representation {{Learning}}}, + author = {Bengio, Yoshua and Courville, Aaron and Vincent, Pascal}, + date = {2013-08}, + journaltitle = {IEEE Transactions on Pattern Analysis and Machine Intelligence}, + volume = {35}, + number = {8}, + pages = {1798--1828}, + issn = {1939-3539}, + doi = {10.1109/TPAMI.2013.50}, + abstract = {The success of machine learning algorithms generally depends on data representation, and we hypothesize that this is because different representations can entangle and hide more or less the different explanatory factors of variation behind the data. Although specific domain knowledge can be used to help design representations, learning with generic priors can also be used, and the quest for AI is motivating the design of more powerful representation-learning algorithms implementing such priors. This paper reviews recent work in the area of unsupervised feature learning and deep learning, covering advances in probabilistic models, autoencoders, manifold learning, and deep networks. This motivates longer term unanswered questions about the appropriate objectives for learning good representations, for computing representations (i.e., inference), and the geometrical connections between representation learning, density estimation, and manifold learning.}, + eventtitle = {{{IEEE Transactions}} on {{Pattern Analysis}} and {{Machine Intelligence}}}, + keywords = {Abstracts,autoencoder,Boltzmann machine,Deep learning,Feature extraction,feature learning,Learning systems,Machine learning,Manifolds,neural nets,Neural networks,representation learning,Speech recognition,unsupervised learning}, + file = {/home/johannes/Nextcloud/Zotero/Bengio et al_2013_Representation Learning.pdf;/home/johannes/Zotero/storage/PEAGSIHD/6472238.html} +} + +@article{benmahmoudLearningElectronicDensity2020, + title = {Learning the Electronic Density of States in Condensed Matter}, + author = {Ben Mahmoud, Chiheb and Anelli, Andrea and Csányi, Gábor and Ceriotti, Michele}, + date = {2020-12-14}, + journaltitle = {Physical Review B}, + shortjournal = {Phys. Rev. B}, + volume = {102}, + number = {23}, + pages = {235130}, + publisher = {{American Physical Society}}, + doi = {10.1103/PhysRevB.102.235130}, + url = {https://link.aps.org/doi/10.1103/PhysRevB.102.235130}, + urldate = {2021-10-20}, + abstract = {The electronic density of states (DOS) quantifies the distribution of the energy levels that can be occupied by electrons in a quasiparticle picture and is central to modern electronic structure theory. It also underpins the computation and interpretation of experimentally observable material properties such as optical absorption and electrical conductivity. We discuss the challenges inherent in the construction of a machine-learning (ML) framework aimed at predicting the DOS as a combination of local contributions that depend in turn on the geometric configuration of neighbors around each atom, using quasiparticle energy levels from density functional theory as training data. We present a challenging case study that includes configurations of silicon spanning a broad set of thermodynamic conditions, ranging from bulk structures to clusters and from semiconducting to metallic behavior. We compare different approaches to represent the DOS, and the accuracy of predicting quantities such as the Fermi level, the electron density at the Fermi level, or the band energy, either directly or as a side product of the evaluation of the DOS. We find that the performance of the model depends crucially on the resolution chosen to smooth the DOS and that there is a tradeoff to be made between the systematic error associated with the smoothing and the error in the ML model for a specific structure. We find however that the errors are not strongly correlated among similar structures, and so the average DOS over an ensemble of configurations is in very good agreement with the reference electronic structure calculations, despite the large nominal error on individual configurations. We demonstrate the usefulness of this approach by computing the density of states of a large amorphous silicon sample, for which it would be prohibitively expensive to compute the DOS by direct electronic structure calculations and show how the atom-centered decomposition of the DOS that is obtained through our model can be used to extract physical insights into the connections between structural and electronic features.}, + keywords = {DFT,GPR,KPCovR,ML,ML-DFT,ML-ESM,nonscalar learning target,PCovR,prediction of LDOS,SOAP}, + file = {/home/johannes/Nextcloud/Zotero/Ben Mahmoud et al_2020_Learning the electronic density of states in condensed matter.pdf;/home/johannes/Zotero/storage/AG3V7VGZ/Ben Mahmoud et al_2020_Learning the electronic density of states in condensed matter2.pdf;/home/johannes/Zotero/storage/BIS7Q3X7/PhysRevB.102.html} +} + +@article{benmahmoudPredictingHotelectronFree2022, + title = {Predicting Hot-Electron Free Energies from Ground-State Data}, + author = {Ben Mahmoud, Chiheb and Grasselli, Federico and Ceriotti, Michele}, + date = {2022-09-27}, + journaltitle = {Physical Review B}, + shortjournal = {Phys. Rev. B}, + volume = {106}, + number = {12}, + pages = {L121116}, + publisher = {{American Physical Society}}, + doi = {10.1103/PhysRevB.106.L121116}, + url = {https://link.aps.org/doi/10.1103/PhysRevB.106.L121116}, + urldate = {2022-09-28}, + abstract = {Machine-learning potentials are usually trained on the ground-state, Born-Oppenheimer energy surface, which depends exclusively on the atomic positions and not on the simulation temperature. This disregards the effect of thermally excited electrons, that is important in metals, and essential to the description of warm dense matter. An accurate physical description of these effects requires that the nuclei move on a temperature-dependent electronic free energy. We propose a method to obtain machine-learning predictions of this free energy at an arbitrary electron temperature using exclusively training data from ground-state calculations, avoiding the need to train temperature-dependent potentials, and benchmark it on metallic liquid hydrogen at the conditions of the core of gas giants and brown dwarfs. This Letter demonstrates the advantages of hybrid schemes that use physical consideration to combine machine-learning predictions, providing a blueprint for the development of similar approaches that extend the reach of atomistic modeling by removing the barrier between physics and data-driven methodologies.}, + keywords = {approximative GPR,EOS,extrapolate from gound state,finite-temperature DFT,forces,GAP,GPR,Hellmann-Feynman,ML,ML-DFT,ML-ESM,prediction from DOS,prediction of free energy,SOAP,warm dense matter}, + file = {/home/johannes/Nextcloud/Zotero/Ben Mahmoud et al_2022_Predicting hot-electron free energies from ground-state data.pdf;/home/johannes/Zotero/storage/6U9PWZG6/Ben Mahmoud et al. - 2022 - Predicting hot-electron free energies from ground-.pdf;/home/johannes/Zotero/storage/5YSTIB2N/PhysRevB.106.html} +} + +@unpublished{bernerModernMathematicsDeep2021, + title = {The {{Modern Mathematics}} of {{Deep Learning}}}, + author = {Berner, Julius and Grohs, Philipp and Kutyniok, Gitta and Petersen, Philipp}, + date = {2021-05-09}, + eprint = {2105.04026}, + eprinttype = {arxiv}, + primaryclass = {cs, stat}, + url = {http://arxiv.org/abs/2105.04026}, + urldate = {2022-01-02}, + abstract = {We describe the new field of mathematical analysis of deep learning. This field emerged around a list of research questions that were not answered within the classical framework of learning theory. These questions concern: the outstanding generalization power of overparametrized neural networks, the role of depth in deep architectures, the apparent absence of the curse of dimensionality, the surprisingly successful optimization performance despite the non-convexity of the problem, understanding what features are learned, why deep architectures perform exceptionally well in physical problems, and which fine aspects of an architecture affect the behavior of a learning task in which way. We present an overview of modern approaches that yield partial answers to these questions. For selected approaches, we describe the main ideas in more detail.}, + archiveprefix = {arXiv}, + keywords = {Computer Science - Machine Learning,Statistics - Machine Learning}, + file = {/home/johannes/Nextcloud/Zotero/Berner et al_2021_The Modern Mathematics of Deep Learning.pdf;/home/johannes/Zotero/storage/XDBSS3FE/2105.html} +} + +@article{bigiSmoothBasisAtomistic2022, + title = {A Smooth Basis for Atomistic Machine Learning}, + author = {Bigi, Filippo and Huguenin-Dumittan, Kevin K. and Ceriotti, Michele and Manolopoulos, David E.}, + date = {2022-12-21}, + journaltitle = {The Journal of Chemical Physics}, + shortjournal = {J. Chem. Phys.}, + volume = {157}, + number = {23}, + pages = {234101}, + publisher = {{American Institute of Physics}}, + issn = {0021-9606}, + doi = {10.1063/5.0124363}, + url = {https://aip.scitation.org/doi/10.1063/5.0124363}, + urldate = {2022-12-29}, + abstract = {Machine learning frameworks based on correlations of interatomic positions begin with a discretized description of the density of other atoms in the neighborhood of each atom in the system. Symmetry considerations support the use of spherical harmonics to expand the angular dependence of this density, but there is, as of yet, no clear rationale to choose one radial basis over another. Here, we investigate the basis that results from the solution of the Laplacian eigenvalue problem within a sphere around the atom of interest. We show that this generates a basis of controllable smoothness within the sphere (in the same sense as plane waves provide a basis with controllable smoothness for a problem with periodic boundaries) and that a tensor product of Laplacian eigenstates also provides a smooth basis for expanding any higher-order correlation of the atomic density within the appropriate hypersphere. We consider several unsupervised metrics of the quality of a basis for a given dataset and show that the Laplacian eigenstate basis has a performance that is much better than some widely used basis sets and competitive with data-driven bases that numerically optimize each metric. Finally, we investigate the role of the basis in building models of the potential energy. In these tests, we find that a combination of the Laplacian eigenstate basis and target-oriented heuristics leads to equal or improved regression performance when compared to both heuristic and data-driven bases in the literature. We conclude that the smoothness of the basis functions is a key aspect of successful atomic density representations.}, + keywords = {/unread,ACDC,ACE,density correlation,descriptor comparison,descriptors,equivariant,general body-order,GPR,Jacobian condition number,Laplacian eigenstate basis,MACE,NN,prediction of potential energy,radial basis,regression,residual variance,smooth basis,Supervised learning,unsupervised learning}, + file = {/home/johannes/Nextcloud/Zotero/Bigi et al_2022_A smooth basis for atomistic machine learning.pdf;/home/johannes/Nextcloud/Zotero/Bigi et al_2022_A smooth basis for atomistic machine learning2.pdf} +} + +@software{blaiszikChartingMLPublications2022, + title = {Charting {{ML Publications}} in {{Science}}}, + author = {Blaiszik, Ben}, + date = {2022-11-25T09:52:56Z}, + origdate = {2019-06-09T00:07:07Z}, + url = {https://github.com/blaiszik/ml_publication_charts}, + urldate = {2022-12-29}, + keywords = {/unread,AML,literature analysis,ML,popular science} +} + +@article{blankNeuralNetworkModels1995, + title = {Neural Network Models of Potential Energy Surfaces}, + author = {Blank, Thomas B. and Brown, Steven D. and Calhoun, August W. and Doren, Douglas J.}, + date = {1995-09-08}, + journaltitle = {The Journal of Chemical Physics}, + shortjournal = {J. Chem. Phys.}, + volume = {103}, + number = {10}, + pages = {4129--4137}, + publisher = {{American Institute of Physics}}, + issn = {0021-9606}, + doi = {10.1063/1.469597}, + url = {https://aip.scitation.org/doi/10.1063/1.469597}, + urldate = {2021-10-22}, + file = {/home/johannes/Nextcloud/Zotero/Blank et al_1995_Neural network models of potential energy surfaces.pdf} +} + +@report{bluegelComputationalCondensedMatter2006, + title = {Computational Condensed Matter Physics : lecture manuscripts of the 37th Spring School of the Institute of Solid State Research ; this Spring School was organized by the Institute of Solid State Research in the Forschungszentrum Jülich on March 6 - 17 ...}, + shorttitle = {Computational Condensed Matter Physics}, + author = {Blügel, S. and Müller-Krumbhaar, H. and Spatschek, R. and Koch, E. and Gompper, G. and Winkler, R. G.}, + date = {2006}, + number = {PreJuSER-56047}, + institution = {{Forschungszentrum, Zentralbibliothek}}, + url = {http://hdl.handle.net/2128/2396}, + urldate = {2021-12-12}, + abstract = {Blügel, S.; Gompper, G.; Koch, E.; Müller-Krumbhaar, H.; Spatschek, R.; Winkler, R. G.}, + isbn = {9783893364305}, + langid = {ngerman}, + keywords = {condensed matter,DFT,FZJ,IFF,IFF spring school,KKR,magnetism,PGI-1/IAS-1}, + file = {/home/johannes/Nextcloud/Zotero/Blügel et al_2006_Computational Condensed Matter Physics.pdf;/home/johannes/Zotero/storage/IUT3QPKV/56047.html} +} + +@report{blugelDensityFunctionalTheory2006, + title = {Density {{Functional Theory}} in {{Practice}}}, + author = {Blügel, S.}, + date = {2006}, + number = {PreJuSER-51316}, + institution = {{Theorie I}}, + url = {https://juser.fz-juelich.de/record/51316}, + urldate = {2021-12-12}, + abstract = {Blügel, S.}, + isbn = {9783893364305}, + langid = {english}, + keywords = {_tablet,bluegel,DFT,FLEUR,IFF,IFF spring school,PGI-1/IAS-1,rec-by-bluegel}, + file = {/home/johannes/Nextcloud/Zotero/Blügel_2006_Density Functional Theory in Practice.pdf;/home/johannes/Zotero/storage/ZL4WZAY7/51316.html} +} + +@book{blumFoundationsDataScience2020, + title = {Foundations of {{Data Science}}}, + author = {Blum, Avrim and Hopcroft, John and Kannan, Ravi}, + date = {2020-01-31}, + edition = {1}, + publisher = {{Cambridge University Press}}, + doi = {10.1017/9781108755528}, + url = {https://www.cambridge.org/core/product/identifier/9781108755528/type/book}, + urldate = {2021-05-04}, + isbn = {978-1-108-75552-8 978-1-108-48506-7}, + keywords = {data science,general,theory}, + file = {/home/johannes/Books/data_science/general_theory/Blum_FoundationsOfDataScience_1e-2020.pdf} +} + +@book{blundellMagnetismCondensedMatter2001, + title = {Magnetism in Condensed Matter}, + author = {Blundell, Stephen}, + date = {2001}, + publisher = {{Oxford University Press}}, + location = {{Oxford; New York}}, + url = {http://public.eblib.com/choice/publicfullrecord.aspx?p=4963266}, + urldate = {2022-06-18}, + abstract = {An understanding of the quantum mechanical nature of magnetism has led to the development of new magnetic materials which are used as permanent magnets, sensors, and information storage. Behind these practical applications lie a range of fundamental ideas, including symmetry breaking, order parameters, excitations, frustration, and reduced dimensionality. This superb new textbook presents a logical account of these ideas, staring from basic concepts in electromagnetsim and quantum mechanics. It outlines the origin of magnetic moments in atoms and how these moments can be affected by their local environment inside a crystal. The different types of interactions which can be present between magnetic moments are described. The final chapters of the book are devoted to the magnetic properties of metals, and to the complex behaviour which can occur when competing magnetic interactions are present and/or the system has a reduced dimensionality. Throughout the text, the theoretical principles are applied to real systems. There is substantial discussion of experimental techniques and current research topics.; The book is copiously illustrated and contains detailed appendices which cover the fundamental principles.}, + isbn = {9780585483603 9781280375132 9780191586644 9786610375134 9780198505921}, + langid = {english}, + keywords = {condensed matter,magnetism,textbook,undergraduate}, + annotation = {OCLC: 53956469}, + file = {/home/johannes/Nextcloud/Zotero/Blundell_2001_Magnetism in condensed matter.pdf} +} + +@article{bochkarevEfficientParametrizationAtomic2022, + title = {Efficient Parametrization of the Atomic Cluster Expansion}, + author = {Bochkarev, Anton}, + date = {2022}, + journaltitle = {Physical Review Materials}, + shortjournal = {Phys. Rev. Materials}, + volume = {6}, + number = {1}, + doi = {10.1103/PhysRevMaterials.6.013804}, + keywords = {_tablet,ACE,descriptors,library,ML,pacemaker,with-code}, + file = {/home/johannes/Nextcloud/Zotero/Bochkarev_2022_Efficient parametrization of the atomic cluster expansion.pdf;/home/johannes/Zotero/storage/LLPTMRGA/PhysRevMaterials.6.html} +} + +@unpublished{bochkarevMultilayerAtomicCluster2022, + title = {Multilayer Atomic Cluster Expansion for Semi-Local Interactions}, + author = {Bochkarev, Anton and Lysogorskiy, Yury and Ortner, Christoph and Csányi, Gábor and Drautz, Ralf}, + date = {2022-05-17}, + number = {arXiv:2205.08177}, + eprint = {2205.08177}, + eprinttype = {arxiv}, + primaryclass = {cond-mat}, + publisher = {{arXiv}}, + doi = {10.48550/arXiv.2205.08177}, + url = {http://arxiv.org/abs/2205.08177}, + urldate = {2022-05-21}, + abstract = {Traditionally, interatomic potentials assume local bond formation supplemented by long-range electrostatic interactions when necessary. This ignores intermediate range multi-atom interactions that arise from the relaxation of the electronic structure. Here, we present the multilayer atomic cluster expansion (ml-ACE) that includes collective, semi-local multi-atom interactions naturally within its remit. We demonstrate that ml-ACE significantly improves fit accuracy compared to a local expansion on selected examples and provide physical intuition to understand this improvement.}, + archiveprefix = {arXiv}, + keywords = {_tablet,ACE,descriptors,ML,ml-ACE}, + file = {/home/johannes/Nextcloud/Zotero/Bochkarev et al_2022_Multilayer atomic cluster expansion for semi-local interactions.pdf;/home/johannes/Zotero/storage/NQ2MH8V7/2205.html} +} + +@misc{bochkarevMultilayerAtomicCluster2022a, + title = {Multilayer Atomic Cluster Expansion for Semi-Local Interactions}, + author = {Bochkarev, Anton and Lysogorskiy, Yury and Ortner, Christoph and Csányi, Gábor and Drautz, Ralf}, + date = {2022-05-17}, + number = {arXiv:2205.08177}, + eprint = {2205.08177}, + eprinttype = {arxiv}, + primaryclass = {cond-mat}, + publisher = {{arXiv}}, + doi = {10.48550/arXiv.2205.08177}, + url = {http://arxiv.org/abs/2205.08177}, + urldate = {2022-09-29}, + abstract = {Traditionally, interatomic potentials assume local bond formation supplemented by long-range electrostatic interactions when necessary. This ignores intermediate range multi-atom interactions that arise from the relaxation of the electronic structure. Here, we present the multilayer atomic cluster expansion (ml-ACE) that includes collective, semi-local multi-atom interactions naturally within its remit. We demonstrate that ml-ACE significantly improves fit accuracy compared to a local expansion on selected examples and provide physical intuition to understand this improvement.}, + archiveprefix = {arXiv}, + keywords = {ACE,descriptors,ML,MLP,multilayer-ACE,semilocal interactions}, + file = {/home/johannes/Nextcloud/Zotero/Bochkarev et al_2022_Multilayer atomic cluster expansion for semi-local interactions2.pdf;/home/johannes/Zotero/storage/ZVU3IARD/2205.html} +} + +@article{bockHybridModellingMachine2021, + title = {Hybrid {{Modelling}} by {{Machine Learning Corrections}} of {{Analytical Model Predictions}} towards {{High-Fidelity Simulation Solutions}}}, + author = {Bock, Frederic E. and Keller, Sören and Huber, Norbert and Klusemann, Benjamin}, + date = {2021-01}, + journaltitle = {Materials}, + volume = {14}, + number = {8}, + pages = {1883}, + publisher = {{Multidisciplinary Digital Publishing Institute}}, + issn = {1996-1944}, + doi = {10.3390/ma14081883}, + url = {https://www.mdpi.com/1996-1944/14/8/1883}, + urldate = {2022-05-13}, + abstract = {Within the fields of materials mechanics, the consideration of physical laws in machine learning predictions besides the use of data can enable low prediction errors and robustness as opposed to predictions only based on data. On the one hand, exclusive utilization of fundamental physical relationships might show significant deviations in their predictions compared to reality, due to simplifications and assumptions. On the other hand, using only data and neglecting well-established physical laws can create the need for unreasonably large data sets that are required to exhibit low bias and are usually expensive to collect. However, fundamental but simplified physics in combination with a corrective model that compensates for possible deviations, e.g., to experimental data, can lead to physics-based predictions with low prediction errors, also despite scarce data. In this article, it is demonstrated that a hybrid model approach consisting of a physics-based model that is corrected via an artificial neural network represents an efficient prediction tool as opposed to a purely data-driven model. In particular, a semi-analytical model serves as an efficient low-fidelity model with noticeable prediction errors outside its calibration domain. An artificial neural network is used to correct the semi-analytical solution towards a desired reference solution provided by high-fidelity finite element simulations, while the efficiency of the semi-analytical model is maintained and the applicability range enhanced. We utilize residual stresses that are induced by laser shock peening as a use-case example. In addition, it is shown that non-unique relationships between model inputs and outputs lead to high prediction errors and the identification of salient input features via dimensionality analysis is highly beneficial to achieve low prediction errors. In a generalization task, predictions are also outside the process parameter space of the training region while remaining in the trained range of corrections. The corrective model predictions show substantially smaller errors than purely data-driven model predictions, which illustrates one of the benefits of the hybrid modelling approach. Ultimately, when the amount of samples in the data set is reduced, the generalization of the physics-related corrective model outperforms the purely data-driven model, which also demonstrates efficient applicability of the proposed hybrid modelling approach to problems where data is scarce.}, + issue = {8}, + langid = {english}, + keywords = {ANN,feature engineering,FEM,physics-informed ML}, + file = {/home/johannes/Nextcloud/Zotero/Bock et al_2021_Hybrid Modelling by Machine Learning Corrections of Analytical Model.pdf;/home/johannes/Zotero/storage/IN7CCMRJ/htm.html} +} + +@misc{bogojeskiEfficientPrediction3D2018, + title = {Efficient Prediction of {{3D}} Electron Densities Using Machine Learning}, + author = {Bogojeski, Mihail and Brockherde, Felix and Vogt-Maranto, Leslie and Li, Li and Tuckerman, Mark E. and Burke, Kieron and Müller, Klaus-Robert}, + date = {2018-11-15}, + number = {arXiv:1811.06255}, + eprint = {1811.06255}, + eprinttype = {arxiv}, + primaryclass = {physics}, + publisher = {{arXiv}}, + doi = {10.48550/arXiv.1811.06255}, + url = {http://arxiv.org/abs/1811.06255}, + urldate = {2022-07-08}, + abstract = {The Kohn-Sham scheme of density functional theory is one of the most widely used methods to solve electronic structure problems for a vast variety of atomistic systems across different scientific fields. While the method is fast relative to other first principles methods and widely successful, the computational time needed is still not negligible, making it difficult to perform calculations for very large systems or over long time-scales. In this submission, we revisit a machine learning model capable of learning the electron density and the corresponding energy functional based on a set of training examples. It allows us to bypass solving the Kohn-Sham equations, providing a significant decrease in computation time. We specifically focus on the machine learning formulation of the Hohenberg-Kohn map and its decomposability. We give results and discuss challenges, limits and future directions.}, + archiveprefix = {arXiv}, + keywords = {_tablet,DFT,HK map,ML,ML-DFT,ML-ESM,ML-HK map,molecules,prediction of electron density}, + file = {/home/johannes/Nextcloud/Zotero/Bogojeski et al_2018_Efficient prediction of 3D electron densities using machine learning.pdf;/home/johannes/Zotero/storage/MCBT39D4/1811.html} +} + +@article{bogojeskiQuantumChemicalAccuracy2020, + title = {Quantum Chemical Accuracy from Density Functional Approximations via Machine Learning}, + author = {Bogojeski, Mihail and Vogt-Maranto, Leslie and Tuckerman, Mark E. and Müller, Klaus-Robert and Burke, Kieron}, + date = {2020-10-16}, + journaltitle = {Nature Communications}, + shortjournal = {Nat Commun}, + volume = {11}, + number = {1}, + pages = {5223}, + publisher = {{Nature Publishing Group}}, + issn = {2041-1723}, + doi = {10.1038/s41467-020-19093-1}, + url = {https://www.nature.com/articles/s41467-020-19093-1}, + urldate = {2021-10-16}, + abstract = {Kohn-Sham density functional theory (DFT) is a standard tool in most branches of chemistry, but accuracies for many molecules are limited to 2-3 kcal~â‹…~mol−1 with presently-available functionals. Ab initio methods, such as coupled-cluster, routinely produce much higher accuracy, but computational costs limit their application to small molecules. In this paper, we leverage machine learning to calculate coupled-cluster energies from DFT densities, reaching quantum chemical accuracy (errors below 1 kcal~â‹…~mol−1) on test data. Moreover, density-based Δ-learning (learning only the correction to a standard DFT calculation, termed Δ-DFT~) significantly reduces the amount of training data required, particularly when molecular symmetries are included. The robustness of Δ-DFT~ is highlighted by correcting “on the fly†DFT-based molecular dynamics (MD) simulations of resorcinol (C6H4(OH)2) to obtain MD trajectories with coupled-cluster accuracy. We conclude, therefore, that Δ-DFT~ facilitates running gas-phase MD simulations with quantum chemical accuracy, even for strained geometries and conformer changes where standard DFT fails.}, + issue = {1}, + langid = {english}, + keywords = {_tablet,2-step model,CCSD(T),coupled cluster,Delta,delta learning,DFT,HK map,KKR,ML,ML-DFA,ML-DFT,ML-ESM,ML-HK map,molecules,prediction of electron density,with-code,Δ-machine learning}, + annotation = {Bandiera\_abtest: a Cc\_license\_type: cc\_by Cg\_type: Nature Research Journals Primary\_atype: Research Subject\_term: Computational chemistry;Computational science Subject\_term\_id: computational-chemistry;computational-science}, + file = {/home/johannes/Nextcloud/Zotero/Bogojeski et al_2020_Quantum chemical accuracy from density functional approximations via machine.pdf;/home/johannes/Nextcloud/Zotero/false} +} + +@article{borchaniSurveyMultioutputRegression2015, + title = {A Survey on Multi-Output Regression}, + author = {Borchani, Hanen and Varando, Gherardo and Bielza, Concha and Larrañaga, Pedro}, + date = {2015}, + journaltitle = {WIREs Data Mining and Knowledge Discovery}, + volume = {5}, + number = {5}, + pages = {216--233}, + issn = {1942-4795}, + doi = {10.1002/widm.1157}, + url = {https://onlinelibrary.wiley.com/doi/abs/10.1002/widm.1157}, + urldate = {2021-09-02}, + abstract = {In recent years, a plethora of approaches have been proposed to deal with the increasingly challenging task of multi-output regression. This study provides a survey on state-of-the-art multi-output regression methods, that are categorized as problem transformation and algorithm adaptation methods. In addition, we present the mostly used performance evaluation measures, publicly available data sets for multi-output regression real-world problems, as well as open-source software frameworks. WIREs Data Mining Knowl Discov 2015, 5:216–233. doi: 10.1002/widm.1157 This article is categorized under: Technologies {$>$} Machine Learning}, + langid = {english}, + keywords = {ML,multi-output learning,multi-target learning,regression}, + annotation = {\_eprint: https://wires.onlinelibrary.wiley.com/doi/pdf/10.1002/widm.1157}, + file = {/home/johannes/Nextcloud/Zotero/Borchani et al_2015_A survey on multi-output regression.pdf;/home/johannes/Zotero/storage/FKMFKWW3/widm.html} +} + +@thesis{bornemannLargescaleInvestigationsNontrivial2019, + title = {Large-Scale {{Investigations}} of {{Non-trivial Magnetic Textures}} in {{Chiral Magnets}} with {{Density Functional Theory}}}, + author = {Bornemann, Marcel}, + date = {2019}, + number = {FZJ-2019-02271}, + institution = {{Forschungszentrum Jülich GmbH Zentralbibliothek, Verlag}}, + url = {http://hdl.handle.net/2128/22015}, + urldate = {2022-08-12}, + abstract = {The large-scale Density Functional Theory (DFT) code KKRnano allows one to perform \$\textbackslash textit\{ab initio\}\$ simulations for thousands of atoms. In this thesis an extension of KKRnano is presented and utilized which facilitates the investigation of exotic non-collinear magnetic textures in bulk materials on huge length scales. Such an undertakinginevitably involves the utilization of High Performance Computing (HPC) which is itself a scientific field. The work in this context includes the adaptation of new coding paradigms and the optimization of codes on constantly changing hardware architectures. In KKRnano, the runtime of a simulation scales linearly with the number of atoms due to an advanced Korringa-Kohn-Rostoker (KKR) scheme that is applied, in which the sparsity of the matrices in the multiple-scattering equations is exploited. This enables us to investigate phenomena that occur on a length scale of nanometers involving thousands of atoms.The main purpose of this thesis was to generalize the KKR formalism in KKRnano in such a way that a non-collinear alignment of the atomic spins can be treated. In addition to this, the relativistic coupling of spin and orbital degrees of freedom, which arises from the Dirac equation, was introduced to the code. This coupling gives rise to the Dzyaloshinskii-Moriya interaction (DMI) from which the formation of non-collinear magnetic textures usually originates. Other methodological features that were added to KKRnano or were re-established in the context of this thesis are the Generalized Gradient Approximation (GGA), Lloyd’s formula and a semi-core energy contour integration. GGA is known to be a better approximation to the exchange-correlation energy in DFT than the still very popular Local Density Approximation (LDA), Lloyd’s formula allows to determine the charge density exactly, despite the angular momentum expansion of all quantities, and the semi-core energy contour integration facilitates the treatment of high-lying electronic core states. Furthermore, an experimental port of the multiple-scattering solver routine to Graphics Processing Unit (GPU) architectures is discussed and the large-scale capabilities of KKR nano are demonstrated by benchmark calculations on the supercomputer JUQUEEN that include more than 200.000 atoms. The new version of KKRnano is used to investigate the magnetic B20 compounds B20-MnGe and B20-FeGe as well as alloys of B20-Mn\$\_\{1−x\}\$Fe\$\_\{x\}\$Ge type with varied concentration of Mn and Ge. These compounds are well-known for exhibiting helicalstates. Recently reported observations of topologically protected magnetic particles, also known as skyrmions, make them promising candidates for future spintronic devices. Initially, the known pressure-induced transition from a high-spin to a low-spin state in B20-MnGe is reproduced with KKRnano and an examination of the magnetocrystalline anisotropy yields unexpected results. [...] Bornemann, Marcel}, + isbn = {9783958063945}, + langid = {english}, + keywords = {juKKR,KKR,PGI-1/IAS-1,thesis}, + file = {/home/johannes/Nextcloud/Zotero/Bornemann_2019_Large-scale Investigations of Non-trivial Magnetic Textures in Chiral Magnets.pdf;/home/johannes/Zotero/storage/BZP7D4IW/861845.html} +} + +@article{bouazizSpinDynamics3d2019, + title = {Spin Dynamics of 3d and 4d Impurities Embedded in Prototypical Topological Insulators}, + shorttitle = {Spin Dynamics Of}, + author = {Bouaziz, Juba}, + date = {2019}, + journaltitle = {Physical Review Materials}, + shortjournal = {Phys. Rev. Materials}, + volume = {3}, + number = {5}, + doi = {10.1103/PhysRevMaterials.3.054201}, + keywords = {_tablet,defects,Funsilab,impurity embedding,PGI-1/IAS-1,topological insulator}, + file = {/home/johannes/Nextcloud/Zotero/Bouaziz_2019_Spin dynamics of 3d and 4d impurities embedded in prototypical topological.pdf;/home/johannes/Zotero/storage/CW3GMSS2/PhysRevMaterials.3.html} +} + +@thesis{bouazizSpinorbitronicsNanoscaleAnalytical2019, + title = {Spin-Orbitronics at the Nanoscale: {{From}} Analytical Models to Real Materials}, + shorttitle = {Spin-Orbitronics at the Nanoscale}, + author = {Bouaziz, Juba}, + date = {2019}, + number = {FZJ-2019-05254}, + institution = {{Forschungszentrum Jülich GmbH Zentralbibliothek, Verlag}}, + url = {http://hdl.handle.net/2128/23183}, + urldate = {2022-08-12}, + abstract = {This thesis provides a theoretical description of magnetic nanostructures in inversion-asymmetric environments with strong spin-orbit interaction (SOI). The theoretical concepts introduced here can be applied in the field of spin-orbitronics, which consists ofexploiting the SOI to manipulate the electron spin without external magnetic fields. The investigated systems display a plethora of interesting phenomena ranging from chiral magnetic interactions to gapped magnetic excitations. In practice, we adopt two different approaches: First, a model-based one relying on the Rashba Hamiltonian, which is employed to demystify and understand magnetic and transport properties of magnetic nanostructures embedded in a Rashba electron gas. Second, we use a first-principles approach within the framework of the Korringa-Kohn-Rostoker (KKR) Green function method to investigate the ground state properties of magnetic impurities in topologically insulating hosts. This method is suitable to simulate nanostructures in real space. Then, we employed our newly developed code based on time-dependent density functional theory to compute the spin excitation spectra of these magnetic nanostructures embedded in topological insulators. Moreover, the KKR Green function method was used to simulate the electronic structure and ground state properties of large magnetic nanostructures, namely magnetic Skyrmions. In the first part, the analytical Rashba Green function and the scattering matrices modeling the magnetic impurities in the s-wave approximation are employed for the computation of the magnetic interaction tensor which contains: isotropic exchange, Dzyaloshinskii-Moriya (DM) and pseudo-dipolar interactions. The competition between these interactions leads to a rich phase diagram depending on the distance between the magnetic impurities. Next, we consider an external perturbing electric field and investigate the transport properties by computing the residual resistivity tensor within linear response theory. The contribution of SOI is explored. The investigation of arbitrary orientations of the impurity magnetic moment allowed a detailed analysis of contributions from the anisotropic magnetoresistance and planar Hall effect. Moreover, we calculate the impurity induced bound currents in the Rashba electron gas, which are used to compute the induced orbital magnetization. For a trimer of impurities with a non-vanishing spin chirality (SC) a finite orbital magnetization is observed when SOI is turned off. Since it emerges from the SC, it was named chiral orbital magnetization. [...] Bouaziz, Juba}, + isbn = {9783958064294}, + langid = {english}, + keywords = {Hall QHE,Hall QSHE,juKKR,KKR,PGI-1/IAS-1,skyrmions,thesis,topological insulator}, + file = {/home/johannes/Nextcloud/Zotero/Bouaziz_2019_Spin-orbitronics at the nanoscale.pdf;/home/johannes/Zotero/storage/YM28TKHA/865993.html} +} + +@article{brackTenSimpleRules2022, + title = {Ten Simple Rules for Making a Software Tool Workflow-Ready}, + author = {Brack, Paul and Crowther, Peter and Soiland-Reyes, Stian and Owen, Stuart and Lowe, Douglas and Williams, Alan R. and Groom, Quentin and Dillen, Mathias and Coppens, Frederik and Grüning, Björn and Eguinoa, Ignacio and Ewels, Philip and Goble, Carole}, + date = {2022-03-24}, + journaltitle = {PLOS Computational Biology}, + shortjournal = {PLOS Computational Biology}, + volume = {18}, + number = {3}, + pages = {e1009823}, + publisher = {{Public Library of Science}}, + issn = {1553-7358}, + doi = {10.1371/journal.pcbi.1009823}, + url = {https://journals.plos.org/ploscompbiol/article?id=10.1371/journal.pcbi.1009823}, + urldate = {2022-07-28}, + langid = {english}, + keywords = {best practices,RSE,scientific workflows}, + file = {/home/johannes/Nextcloud/Zotero/Brack et al_2022_Ten simple rules for making a software tool workflow-ready.pdf;/home/johannes/Zotero/storage/QPHMATCQ/article.html} +} + +@unpublished{brandstetterLiePointSymmetry2022, + title = {Lie {{Point Symmetry Data Augmentation}} for {{Neural PDE Solvers}}}, + author = {Brandstetter, Johannes and Welling, Max and Worrall, Daniel E.}, + date = {2022-05-29}, + number = {arXiv:2202.07643}, + eprint = {2202.07643}, + eprinttype = {arxiv}, + primaryclass = {cs}, + publisher = {{arXiv}}, + doi = {10.48550/arXiv.2202.07643}, + url = {http://arxiv.org/abs/2202.07643}, + urldate = {2022-06-09}, + abstract = {Neural networks are increasingly being used to solve partial differential equations (PDEs), replacing slower numerical solvers. However, a critical issue is that neural PDE solvers require high-quality ground truth data, which usually must come from the very solvers they are designed to replace. Thus, we are presented with a proverbial chicken-and-egg problem. In this paper, we present a method, which can partially alleviate this problem, by improving neural PDE solver sample complexity -- Lie point symmetry data augmentation (LPSDA). In the context of PDEs, it turns out that we are able to quantitatively derive an exhaustive list of data transformations, based on the Lie point symmetry group of the PDEs in question, something not possible in other application areas. We present this framework and demonstrate how it can easily be deployed to improve neural PDE solver sample complexity by an order of magnitude.}, + archiveprefix = {arXiv}, + keywords = {data augmentation,ML,neural PDE solver,PDE,PINN,symmetry,with-code}, + file = {/home/johannes/Nextcloud/Zotero/Brandstetter et al_2022_Lie Point Symmetry Data Augmentation for Neural PDE Solvers.pdf;/home/johannes/Zotero/storage/QUUR7MZV/2202.html} +} + +@article{brandtKadi4MatResearchData2021, + title = {{{Kadi4Mat}}: {{A Research Data Infrastructure}} for {{Materials Science}}}, + shorttitle = {{{Kadi4Mat}}}, + author = {Brandt, Nico and Griem, Lars and Herrmann, Christoph and Schoof, Ephraim and Tosato, Giovanna and Zhao, Yinghan and Zschumme, Philipp and Selzer, Michael}, + date = {2021-02-10}, + journaltitle = {Data Science Journal}, + volume = {20}, + number = {1}, + pages = {8}, + publisher = {{Ubiquity Press}}, + issn = {1683-1470}, + doi = {10.5334/dsj-2021-008}, + url = {http://datascience.codata.org/articles/10.5334/dsj-2021-008/}, + urldate = {2022-06-23}, + abstract = {The concepts and current developments of a research data infrastructure for materials science are presented, extending and combining the features of an electronic lab notebook and a repository. The objective of this infrastructure is to incorporate the possibility of structured data storage and data exchange with documented and reproducible data analysis and visualization, which finally leads to the publication of the data. This way, researchers can be supported throughout the entire research process. The software is being developed as a web-based and desktop-based system, offering both a graphical user interface and a programmatic interface. The focus of the development is on the integration of technologies and systems based on both established as well as new concepts. Due to the heterogeneous nature of materials science data, the current features are kept mostly generic, and the structuring of the data is largely left to the users. As a result, an extension of the research data infrastructure to other disciplines is possible in the future. The source code of the project is publicly available under a permissive Apache 2.0 license.}, + issue = {1}, + langid = {english}, + keywords = {ELN,RDM,repository,RSE,workflows}, + file = {/home/johannes/Nextcloud/Zotero/Brandt et al_2021_Kadi4Mat.pdf;/home/johannes/Zotero/storage/GMTCYYBY/dsj-2021-008.html} +} + +@article{braunImpactSpinOrbit2021, + title = {The {{Impact}} of {{Spin}}–{{Orbit Interaction}} on the {{Image States}} of {{High-Z Materials}}}, + author = {Braun, Jürgen and Ebert, Hubert}, + date = {2021}, + journaltitle = {physica status solidi (b)}, + volume = {258}, + number = {1}, + pages = {2000026}, + issn = {1521-3951}, + doi = {10.1002/pssb.202000026}, + url = {https://onlinelibrary.wiley.com/doi/abs/10.1002/pssb.202000026}, + urldate = {2022-05-30}, + abstract = {Due to many important technical developments over the past two decades angle-resolved (inverse) photoemission has become the method of choice to study experimentally the bulk and surface-related electronic states of solids in the most detailed way. Due to new powerful photon sources as well as efficient analyzers and detectors extremely high energy and angle resolution are achieved nowadays for spin-integrated and also for spin-resolved measurements. These developments allow in particular to explore the influence of spin–orbit coupling on image potential states of simple metals like Ir, Pt, or Au with a high atomic number as well as new types of materials as for example topological insulators. Herein, fully relativistic angle- and spin-resolved inverse photoemission calculations are presented that make use of the spin-density matrix formulation of the one-step model. This way a quantitative analysis of all occupied and unoccupied electronic features in the vicinity of the Fermi level is achieved for a wide range of excitation energies. Using this approach, in addition, it is possible to deal with arbitrarily ordered but also disordered systems. Because of these features, the one-step or spectral function approach to photoemission permits detailed theoretical studies on a large variety of interesting solid-state systems.}, + langid = {english}, + keywords = {image states,photoemission,spin–orbit interaction}, + annotation = {\_eprint: https://onlinelibrary.wiley.com/doi/pdf/10.1002/pssb.202000026}, + file = {/home/johannes/Nextcloud/Zotero/Braun_Ebert_2021_The Impact of Spin–Orbit Interaction on the Image States of High-Z Materials.pdf} +} + +@article{brockherdeBypassingKohnShamEquations2017, + title = {Bypassing the {{Kohn-Sham}} Equations with Machine Learning}, + author = {Brockherde, Felix and Vogt, Leslie and Li, Li and Tuckerman, Mark E. and Burke, Kieron and Müller, Klaus-Robert}, + date = {2017-10-11}, + journaltitle = {Nature Communications}, + shortjournal = {Nat Commun}, + volume = {8}, + number = {1}, + pages = {872}, + publisher = {{Nature Publishing Group}}, + issn = {2041-1723}, + doi = {10.1038/s41467-017-00839-3}, + url = {https://www.nature.com/articles/s41467-017-00839-3}, + urldate = {2022-06-15}, + abstract = {Last year, at least 30,000 scientific papers used the Kohn–Sham scheme of density functional theory to solve electronic structure problems in a wide variety of scientific fields. Machine learning holds the promise of learning the energy functional via examples, bypassing the need to solve the Kohn–Sham equations. This should yield substantial savings in computer time, allowing larger systems and/or longer time-scales to be tackled, but attempts to machine-learn this functional have been limited by the need to find its derivative. The present work overcomes this difficulty by directly learning the density-potential and energy-density maps for test systems and various molecules. We perform the first molecular dynamics simulation with a machine-learned density functional on malonaldehyde and are able to capture the intramolecular proton transfer process. Learning density models now allows the construction of accurate density functionals for realistic molecular systems.}, + issue = {1}, + langid = {english}, + keywords = {_tablet,DFT,HK map,KRR,ML,ML-DFT,ML-ESM,ML-HK map,ML-KS,ML-OF,prediction from potential,prediction of electron density}, + file = {/home/johannes/Nextcloud/Zotero/Brockherde et al_2017_Bypassing the Kohn-Sham equations with machine learning.pdf;/home/johannes/Zotero/storage/8X4ALINZ/s41467-017-00839-3.html} +} + +@unpublished{bronsteinGeometricDeepLearning2021, + title = {Geometric {{Deep Learning}}: {{Grids}}, {{Groups}}, {{Graphs}}, {{Geodesics}}, and {{Gauges}}}, + shorttitle = {Geometric {{Deep Learning}}}, + author = {Bronstein, Michael M. and Bruna, Joan and Cohen, Taco and VeliÄković, Petar}, + date = {2021-05-02}, + eprint = {2104.13478}, + eprinttype = {arxiv}, + primaryclass = {cs, stat}, + url = {http://arxiv.org/abs/2104.13478}, + urldate = {2022-04-14}, + abstract = {The last decade has witnessed an experimental revolution in data science and machine learning, epitomised by deep learning methods. Indeed, many high-dimensional learning tasks previously thought to be beyond reach -- such as computer vision, playing Go, or protein folding -- are in fact feasible with appropriate computational scale. Remarkably, the essence of deep learning is built from two simple algorithmic principles: first, the notion of representation or feature learning, whereby adapted, often hierarchical, features capture the appropriate notion of regularity for each task, and second, learning by local gradient-descent type methods, typically implemented as backpropagation. While learning generic functions in high dimensions is a cursed estimation problem, most tasks of interest are not generic, and come with essential pre-defined regularities arising from the underlying low-dimensionality and structure of the physical world. This text is concerned with exposing these regularities through unified geometric principles that can be applied throughout a wide spectrum of applications. Such a 'geometric unification' endeavour, in the spirit of Felix Klein's Erlangen Program, serves a dual purpose: on one hand, it provides a common mathematical framework to study the most successful neural network architectures, such as CNNs, RNNs, GNNs, and Transformers. On the other hand, it gives a constructive procedure to incorporate prior physical knowledge into neural architectures and provide principled way to build future architectures yet to be invented.}, + archiveprefix = {arXiv}, + keywords = {basics,CNN,Deep learning,equivariant,GCN,GDL,General ML,geometric deep learning,GNN,invariance,ML,MPNN,review,review-of-GDL,theory,tutorial}, + file = {/home/johannes/Nextcloud/Zotero/Bronstein et al_2021_Geometric Deep Learning.pdf;/home/johannes/Zotero/storage/6ZLIPHI5/2104.html} +} + +@article{burkeDFTNutshell2013, + title = {{{DFT}} in a Nutshell}, + author = {Burke, Kieron and Wagner, Lucas O.}, + date = {2013}, + journaltitle = {International Journal of Quantum Chemistry}, + volume = {113}, + number = {2}, + pages = {96--101}, + issn = {1097-461X}, + doi = {10.1002/qua.24259}, + url = {https://onlinelibrary.wiley.com/doi/abs/10.1002/qua.24259}, + urldate = {2022-07-10}, + abstract = {The purpose of this short essay is to introduce students and other newcomers to the basic ideas and uses of modern electronic density functional theory, including what kinds of approximations are in current use, and how well they work (or not). The complete newcomer should find it orients them well, while even longtime users and aficionados might find something new outside their area. Important questions varying in difficulty and effort are posed in the text, and are answered in the Supporting Information. © 2012 Wiley Periodicals, Inc.}, + langid = {english}, + keywords = {_tablet,density functional theory,electronic structure,local density approximation}, + annotation = {\_eprint: https://onlinelibrary.wiley.com/doi/pdf/10.1002/qua.24259}, + file = {/home/johannes/Nextcloud/Zotero/Burke_Wagner_2013_DFT in a nutshell.pdf;/home/johannes/Nextcloud/Zotero/false;/home/johannes/Zotero/storage/CCPHAAVK/qua.html} +} + +@misc{burkeLiesMyTeacher2021, + title = {Lies {{My Teacher Told Me About Density Functional Theory}}: {{Seeing Through Them}} with the {{Hubbard Dimer}}}, + shorttitle = {Lies {{My Teacher Told Me About Density Functional Theory}}}, + author = {Burke, Kieron and Kozlowski, John}, + date = {2021-10-18}, + number = {arXiv:2108.11534}, + eprint = {2108.11534}, + eprinttype = {arxiv}, + primaryclass = {cond-mat, physics:physics}, + publisher = {{arXiv}}, + doi = {10.48550/arXiv.2108.11534}, + url = {http://arxiv.org/abs/2108.11534}, + urldate = {2022-07-10}, + abstract = {Most realistic calculations of moderately correlated materials begin with a ground-state density functional theory (DFT) calculation. While Kohn-Sham DFT is used in about 40,000 scientific papers each year, the fundamental underpinnings are not widely appreciated. In this chapter, we analyze the inherent characteristics of DFT in their simplest form, using the asymmetric Hubbard dimer as an illustrative model. We begin by working through the core tenets of DFT, explaining what the exact ground-state density functional yields and does not yield. Given the relative simplicity of the system, almost all properties of the exact exchange-correlation functional are readily visualized and plotted. Key concepts include the Kohn-Sham scheme, the behavior of the XC potential as correlations become very strong, the derivative discontinuity and the difference between KS gaps and true charge gaps, and how to extract optical excitations using time-dependent DFT. By the end of this text and accompanying exercises, the reader will improve their ability to both explain and visualize the concepts of DFT, as well as better understand where others may go wrong.}, + archiveprefix = {arXiv}, + keywords = {DFT,physics,strongly correlated maeterials}, + file = {/home/johannes/Nextcloud/Zotero/Burke_Kozlowski_2021_Lies My Teacher Told Me About Density Functional Theory.pdf;/home/johannes/Zotero/storage/6EW6SVTP/2108.html} +} + +@article{bystromCIDERExpressiveNonlocal2022, + title = {{{CIDER}}: {{An Expressive}}, {{Nonlocal Feature Set}} for {{Machine Learning Density Functionals}} with {{Exact Constraints}}}, + shorttitle = {{{CIDER}}}, + author = {Bystrom, Kyle and Kozinsky, Boris}, + date = {2022-04-12}, + journaltitle = {Journal of Chemical Theory and Computation}, + shortjournal = {J. Chem. Theory Comput.}, + volume = {18}, + number = {4}, + pages = {2180--2192}, + publisher = {{American Chemical Society}}, + issn = {1549-9618}, + doi = {10.1021/acs.jctc.1c00904}, + url = {https://doi.org/10.1021/acs.jctc.1c00904}, + urldate = {2022-05-11}, + abstract = {Machine learning (ML) has recently gained attention as a means to develop more accurate exchange-correlation (XC) functionals for density functional theory, but functionals developed thus far need to be improved on several metrics, including accuracy, numerical stability, and transferability across chemical space. In this work, we introduce a set of nonlocal features of the density called the CIDER formalism, which we use to train a Gaussian process model for the exchange energy that obeys the critical uniform scaling rule for exchange. The resulting CIDER exchange functional is significantly more accurate than any semilocal functional tested here, and it has good transferability across main-group molecules. This work therefore serves as an initial step toward more accurate exchange functionals, and it also introduces useful techniques for developing robust, physics-informed XC models via ML.}, + keywords = {DFT,ML,ML-DFT,ML-ESM}, + file = {/home/johannes/Nextcloud/Zotero/Bystrom_Kozinsky_2022_CIDER.pdf} +} + +@article{calderonAFLOWStandardHighthroughput2015, + title = {The {{AFLOW}} Standard for High-Throughput Materials Science Calculations}, + author = {Calderon, Camilo E. and Plata, Jose J. and Toher, Cormac and Oses, Corey and Levy, Ohad and Fornari, Marco and Natan, Amir and Mehl, Michael J. and Hart, Gus and Buongiorno Nardelli, Marco and Curtarolo, Stefano}, + date = {2015-10-01}, + journaltitle = {Computational Materials Science}, + shortjournal = {Computational Materials Science}, + volume = {108}, + pages = {233--238}, + issn = {0927-0256}, + doi = {10.1016/j.commatsci.2015.07.019}, + url = {https://www.sciencedirect.com/science/article/pii/S0927025615004292}, + urldate = {2021-10-15}, + abstract = {The Automatic-Flow (AFLOW) standard for the high-throughput construction of materials science electronic structure databases is described. Electronic structure calculations of solid state materials depend on a large number of parameters which must be understood by researchers, and must be reported by originators to ensure reproducibility and enable collaborative database expansion. We therefore describe standard parameter values for k-point grid density, basis set plane wave kinetic energy cut-off, exchange–correlation functionals, pseudopotentials, DFT+U parameters, and convergence criteria used in AFLOW calculations.}, + langid = {english}, + keywords = {AFLOWLIB,High-throughput,Materials genomics,VASP}, + file = {/home/johannes/Nextcloud/Zotero/Calderon et al_2015_The AFLOW standard for high-throughput materials science calculations.pdf} +} + +@article{cangiPotentialFunctionalsDensity2013, + title = {Potential Functionals versus Density Functionals}, + author = {Cangi, Attila and Gross, E. K. U. and Burke, Kieron}, + date = {2013-12-04}, + journaltitle = {Physical Review A}, + shortjournal = {Phys. Rev. A}, + volume = {88}, + number = {6}, + pages = {062505}, + publisher = {{American Physical Society}}, + doi = {10.1103/PhysRevA.88.062505}, + url = {https://link.aps.org/doi/10.1103/PhysRevA.88.062505}, + urldate = {2022-07-08}, + abstract = {Potential functional approximations are an intriguing alternative to density functional approximations. The potential functional that is dual to the Lieb density functional is defined and its properties are reported. The relationship between the Thomas-Fermi theory as a density functional and the theory as a potential functional is derived. The properties of several recent semiclassical potential functionals are explored, especially regarding their approach to the large particle number and classical continuum limits. The lack of ambiguity in the energy density of potential functional approximations is demonstrated. The density-density response function of the semiclassical approximation is calculated and shown to violate a key symmetry condition.}, + keywords = {_tablet,density functional,density vs potential,DFT,potential}, + file = {/home/johannes/Nextcloud/Zotero/Cangi et al_2013_Potential functionals versus density functionals.pdf;/home/johannes/Zotero/storage/4U87YYPT/Cangi et al_2013_Potential functionals versus density functionals.pdf;/home/johannes/Zotero/storage/AJH43GTS/PhysRevA.88.html} +} + +@article{caoArtificialIntelligenceHighthroughput2020, + title = {Artificial Intelligence for High-Throughput Discovery of Topological Insulators: {{The}} Example of Alloyed Tetradymites}, + shorttitle = {Artificial Intelligence for High-Throughput Discovery of Topological Insulators}, + author = {Cao, Guohua and Ouyang, Runhai and Ghiringhelli, Luca M. and Scheffler, Matthias and Liu, Huijun and Carbogno, Christian and Zhang, Zhenyu}, + date = {2020-03-23}, + journaltitle = {Physical Review Materials}, + shortjournal = {Phys. Rev. Materials}, + volume = {4}, + number = {3}, + pages = {034204}, + publisher = {{American Physical Society}}, + doi = {10.1103/PhysRevMaterials.4.034204}, + url = {https://link.aps.org/doi/10.1103/PhysRevMaterials.4.034204}, + urldate = {2021-05-21}, + abstract = {Significant advances have been made in predicting new topological materials using high-throughput empirical descriptors or symmetry-based indicators. To date, these approaches have been applied to materials in existing databases, and are severely limited to systems with well-defined symmetries, leaving a much larger materials space unexplored. Using tetradymites as a prototypical class of examples, we uncover a two-dimensional descriptor by applying an artificial intelligence (AI)-based approach for fast and reliable identification of the topological characters of a drastically expanded range of materials, without prior determination of their specific symmetries and detailed band structures. By leveraging this descriptor that contains only the atomic number and electronegativity of the constituent species, we have readily scanned a huge number of alloys in the tetradymite family. Strikingly, nearly half of them are identified to be topological insulators, revealing a much larger territory of the topological materials world. The present work also attests to the increasingly important role of such AI-based approaches in modern materials discovery.}, + keywords = {2D descriptor,classification,descriptors,GW approximation,HTC,materials discovery,materials screening,ML,SISSO,SVM,topological insulator}, + file = {/home/johannes/Nextcloud/Zotero/Cao et al_2020_Artificial intelligence for high-throughput discovery of topological insulators2.pdf} +} + +@article{caoTestsAccuracyScalability2020, + title = {Tests on the {{Accuracy}} and {{Scalability}} of the {{Full-Potential DFT Method Based}} on {{Multiple Scattering Theory}}}, + author = {Cao, Peiyu and Fang, Jun and Gao, Xingyu and Tian, Fuyang and Song, Haifeng}, + date = {2020-01-01}, + journaltitle = {Frontiers in chemistry}, + shortjournal = {Front Chem}, + volume = {8}, + eprint = {33344416}, + eprinttype = {pmid}, + pages = {590047}, + issn = {2296-2646}, + doi = {10.3389/fchem.2020.590047}, + url = {https://europepmc.org/articles/PMC7746799}, + urldate = {2021-08-21}, + abstract = {We investigate a reduced scaling full-potential DFT method based on the multiple scattering theory (MST) code MuST, which is released online (https://github.com/mstsuite/MuST) very recently. First, we test the accuracy by calculating structural properties of typical body-centered cubic (BCC) metals (V, Nb, and Mo). It is shown that the calculated lattice parameters, bulk moduli, and elastic constants agree with those obtained from the VASP, WIEN2k, EMTO, and Elk codes. Second, we test the locally self-consistent multiple scattering (LSMS) mode, which achieves reduced scaling by neglecting the multiple scattering processes beyond a cut-off radius. In the case of Nb, the accuracy of 0.5 mRy/atom can be achieved with a cut-off radius of 20 Bohr, even when small deformations are imposed on the lattice. Despite that the calculation of valence states based on MST exhibits linear scaling, the whole computational procedure has an overall scaling of about O(N1.6) , due to the fact that the updating of Coulomb potential scales almost as O(N2) . Nevertheless, it can be still expected that MuST would provide a reliable and accessible way to large-scale first-principles simulations of metals and alloys.}, + langid = {english}, + pmcid = {PMC7746799}, + keywords = {KKR}, + file = {/home/johannes/Nextcloud/Zotero/Cao et al_2020_Tests on the Accuracy and Scalability of the Full-Potential DFT Method Based on.pdf} +} + +@unpublished{capelleBirdSeyeView2006, + title = {A Bird's-Eye View of Density-Functional Theory}, + author = {Capelle, Klaus}, + date = {2006-11-18}, + eprint = {cond-mat/0211443}, + eprinttype = {arxiv}, + url = {http://arxiv.org/abs/cond-mat/0211443}, + urldate = {2021-08-31}, + abstract = {This paper is the outgrowth of lectures the author gave at the Physics Institute and the Chemistry Institute of the University of Sao Paulo at Sao Carlos, Brazil, and at the VIII'th Summer School on Electronic Structure of the Brazilian Physical Society. It is an attempt to introduce density-functional theory (DFT) in a language accessible for students entering the field or researchers from other fields. It is not meant to be a scholarly review of DFT, but rather an informal guide to its conceptual basis and some recent developments and advances. The Hohenberg-Kohn theorem and the Kohn-Sham equations are discussed in some detail. Approximate density functionals, selected aspects of applications of DFT, and a variety of extensions of standard DFT are also discussed, albeit in less detail. Throughout it is attempted to provide a balanced treatment of aspects that are relevant for chemistry and aspects relevant for physics, but with a strong bias towards conceptual foundations. The paper is intended to be read before (or in parallel with) one of the many excellent more technical reviews available in the literature.}, + archiveprefix = {arXiv}, + keywords = {_tablet,DFT,learn DFT,review}, + file = {/home/johannes/Nextcloud/Zotero/Capelle_2006_A bird's-eye view of density-functional theory.pdf;/home/johannes/Zotero/storage/8TLEU4M3/0211443.html} +} + +@article{carleoMachineLearningPhysical2019, + title = {Machine Learning and the Physical Sciences}, + author = {Carleo, Giuseppe}, + date = {2019}, + journaltitle = {Reviews of Modern Physics}, + shortjournal = {Rev. Mod. Phys.}, + volume = {91}, + number = {4}, + doi = {10.1103/RevModPhys.91.045002}, + keywords = {Many-body theory,ML,review,science}, + file = {/home/johannes/Nextcloud/Zotero/Carleo_2019_Machine learning and the physical sciences.pdf;/home/johannes/Zotero/storage/9YE6JEBD/RevModPhys.91.html} +} + +@article{carleoSolvingQuantumManybody2017, + title = {Solving the Quantum Many-Body Problem with Artificial Neural Networks}, + author = {Carleo, Giuseppe and Troyer, Matthias}, + date = {2017-02-10}, + journaltitle = {Science}, + volume = {355}, + number = {6325}, + pages = {602--606}, + publisher = {{American Association for the Advancement of Science}}, + doi = {10.1126/science.aag2302}, + url = {https://www.science.org/doi/10.1126/science.aag2302}, + urldate = {2022-03-29}, + keywords = {ML,ML-QM,NN,rec-by-bluegel}, + file = {/home/johannes/Nextcloud/Zotero/Carleo_Troyer_2017_Solving the quantum many-body problem with artificial neural networks.pdf} +} + +@article{caroOptimizingManybodyAtomic2019, + title = {Optimizing Many-Body Atomic Descriptors for Enhanced Computational Performance of Machine Learning Based Interatomic Potentials}, + author = {Caro, Miguel A.}, + date = {2019-07-30}, + journaltitle = {Physical Review B}, + shortjournal = {Phys. Rev. B}, + volume = {100}, + number = {2}, + pages = {024112}, + publisher = {{American Physical Society}}, + doi = {10.1103/PhysRevB.100.024112}, + url = {https://link.aps.org/doi/10.1103/PhysRevB.100.024112}, + urldate = {2021-05-13}, + abstract = {We explore different ways to simplify the evaluation of the smooth overlap of atomic positions (SOAP) many-body atomic descriptor [Bartók et al., Phys. Rev. B 87, 184115 (2013).]. Our aim is to improve the computational efficiency of SOAP-based similarity kernel construction. While these improved atomic descriptors can be used for general characterization and interpolation of atomic properties, their main target application is accelerated evaluation of machine-learning-based interatomic potentials within the Gaussian approximation potential (GAP) framework [Bartók et al., Phys. Rev. Lett. 104, 136403 (2010)]. We achieve this objective by expressing the atomic densities in an approximate separable form, which decouples the radial and angular channels. We then express the elements of the SOAP descriptor (i.e., the expansion coefficients for the atomic densities) in analytical form given a particular choice of radial basis set. Finally, we derive recursion formulas for the expansion coefficients. This new SOAP-based descriptor allows for tenfold speedups compared to previous implementations, while improving the stability of the radial expansion for distant atomic neighbors, without degradation of the interpolation power of GAP models.}, + keywords = {_tablet,descriptor dimred,descriptors,descriptors analysis,dimensionality reduction,GAP,ML,MLP,SOAP}, + file = {/home/johannes/Nextcloud/Zotero/Caro_2019_Optimizing many-body atomic descriptors for enhanced computational performance.pdf;/home/johannes/Zotero/storage/FDHHHJTR/PhysRevB.100.html} +} + +@article{carvalhoRealspaceMappingTopological2018, + title = {Real-Space Mapping of Topological Invariants Using Artificial Neural Networks}, + author = {Carvalho, D. and GarcÃa-MartÃnez, N. A. and Lado, J. L. and Fernández-Rossier, J.}, + date = {2018-03-28}, + journaltitle = {Physical Review B}, + shortjournal = {Phys. Rev. B}, + volume = {97}, + number = {11}, + pages = {115453}, + publisher = {{American Physical Society}}, + doi = {10.1103/PhysRevB.97.115453}, + url = {https://link.aps.org/doi/10.1103/PhysRevB.97.115453}, + urldate = {2021-05-21}, + abstract = {Topological invariants allow one to characterize Hamiltonians, predicting the existence of topologically protected in-gap modes. Those invariants can be computed by tracing the evolution of the occupied wave functions under twisted boundary conditions. However, those procedures do not allow one to calculate a topological invariant by evaluating the system locally, and thus require information about the wave functions in the whole system. Here we show that artificial neural networks can be trained to identify the topological order by evaluating a local projection of the density matrix. We demonstrate this for two different models, a one-dimensional topological superconductor and a two-dimensional quantum anomalous Hall state, both with spatially modulated parameters. Our neural network correctly identifies the different topological domains in real space, predicting the location of in-gap states. By combining a neural network with a calculation of the electronic states that uses the kernel polynomial method, we show that the local evaluation of the invariant can be carried out by evaluating a local quantity, in particular for systems without translational symmetry consisting of tens of thousands of atoms. Our results show that supervised learning is an efficient methodology to characterize the local topology of a system.}, + keywords = {ANN,kernel methods,ML,superconductor,topological insulator,topological phase}, + file = {/home/johannes/Nextcloud/Zotero/Carvalho et al_2018_Real-space mapping of topological invariants using artificial neural networks.pdf;/home/johannes/Zotero/storage/ZNNT2KFN/PhysRevB.97.html} +} + +@article{cavaIntroductionQuantumMaterials2021, + title = {Introduction: {{Quantum Materials}}}, + shorttitle = {Introduction}, + author = {Cava, Robert and de Leon, Nathalie and Xie, Weiwei}, + options = {useprefix=true}, + date = {2021-03-10}, + journaltitle = {Chemical Reviews}, + shortjournal = {Chem. Rev.}, + volume = {121}, + number = {5}, + pages = {2777--2779}, + publisher = {{American Chemical Society}}, + issn = {0009-2665}, + doi = {10.1021/acs.chemrev.0c01322}, + url = {https://doi.org/10.1021/acs.chemrev.0c01322}, + urldate = {2021-08-23}, + file = {/home/johannes/Nextcloud/Zotero/Cava et al_2021_Introduction.pdf;/home/johannes/Zotero/storage/EDPNL6EQ/acs.chemrev.html} +} + +@article{ceriottiMachineLearningMeets2021, + title = {Machine Learning Meets Chemical Physics}, + author = {Ceriotti, Michele and Clementi, Cecilia and Anatole von Lilienfeld, O.}, + date = {2021-04-23}, + journaltitle = {The Journal of Chemical Physics}, + shortjournal = {J. Chem. Phys.}, + volume = {154}, + number = {16}, + pages = {160401}, + publisher = {{American Institute of Physics}}, + issn = {0021-9606}, + doi = {10.1063/5.0051418}, + url = {https://aip.scitation.org/doi/10.1063/5.0051418}, + urldate = {2021-05-13}, + abstract = {Over recent years, the use of statistical learning techniques applied to chemical problems has gained substantial momentum. This is particularly apparent in the realm of physical chemistry, where the balance between empiricism and physics-based theory has traditionally been rather in favor of the latter. In this guest Editorial for the special topic issue on “Machine Learning Meets Chemical Physics,†a brief rationale is provided, followed by an overview of the topics covered. We conclude by making some general remarks.}, + keywords = {ANN,Gaussian process,MD,ML,models,QC,review}, + file = {/home/johannes/Nextcloud/Zotero/Ceriotti et al_2021_Machine learning meets chemical physics.pdf;/home/johannes/Zotero/storage/6YB95LVA/5.html} +} + +@article{ceriottiPotentialsIntegratedMachine2022, + title = {Beyond Potentials: {{Integrated}} Machine~Learning Models for Materials}, + shorttitle = {Beyond Potentials}, + author = {Ceriotti, Michele}, + date = {2022-12-06}, + journaltitle = {MRS Bulletin}, + shortjournal = {MRS Bulletin}, + issn = {1938-1425}, + doi = {10.1557/s43577-022-00440-0}, + url = {https://doi.org/10.1557/s43577-022-00440-0}, + urldate = {2022-12-29}, + abstract = {Over the past decade, interatomic potentials based on machine~learning (ML) techniques have become an indispensable tool in the atomic-scale modeling of materials. Trained on energies and forces obtained from electronic-structure calculations, they inherit their predictive accuracy, and extend greatly the length and time scales that are accessible to explicit atomistic simulations. Inexpensive predictions of the energetics of individual configurations have facilitated greatly the calculation of the thermodynamics of materials, including finite-temperature effects and disorder. More recently, ML models have been closing the gap with first-principles calculations in another area: the prediction of arbitrarily complicated functional properties, from vibrational and optical spectroscopies to electronic excitations. The implementation of integrated ML models that combine energetic and functional predictions with statistical and dynamical sampling of atomic-scale properties is bringing the promise of predictive, uncompromising simulations of existing and novel materials closer to its full realization.}, + langid = {english}, + keywords = {equivariant,Gibbs free energy,integrated models,MLP,multiscale,prediction of DOS,prediction of polarizability,review,symmetry,tensorial target,thermodynamics}, + file = {/home/johannes/Nextcloud/Zotero/Ceriotti_2022_Beyond potentials.pdf} +} + +@article{ceriottiSimplifyingRepresentationComplex2011, + title = {Simplifying the Representation of Complex Free-Energy Landscapes Using Sketch-Map}, + author = {Ceriotti, Michele and Tribello, Gareth A. and Parrinello, Michele}, + date = {2011-08-09}, + journaltitle = {Proceedings of the National Academy of Sciences}, + shortjournal = {PNAS}, + volume = {108}, + number = {32}, + eprint = {21730167}, + eprinttype = {pmid}, + pages = {13023--13028}, + publisher = {{National Academy of Sciences}}, + issn = {0027-8424, 1091-6490}, + doi = {10.1073/pnas.1108486108}, + url = {https://www.pnas.org/content/108/32/13023}, + urldate = {2021-07-20}, + abstract = {A new scheme, sketch-map, for obtaining a low-dimensional representation of the region of phase space explored during an enhanced dynamics simulation is proposed. We show evidence, from an examination of the distribution of pairwise distances between frames, that some features of the free-energy surface are inherently high-dimensional. This makes dimensionality reduction problematic because the data does not satisfy the assumptions made in conventional manifold learning algorithms We therefore propose that when dimensionality reduction is performed on trajectory data one should think of the resultant embedding as a quickly sketched set of directions rather than a road map. In other words, the embedding tells one about the connectivity between states but does not provide the vectors that correspond to the slow degrees of freedom. This realization informs the development of sketch-map, which endeavors to reproduce the proximity information from the high-dimensionality description in a space of lower dimensionality even when a faithful embedding is not possible.}, + langid = {english}, + keywords = {data exploration,dimensionality reduction,library,MD,ML,sketchmap,unsupervised learning,visualization,with-code}, + file = {/home/johannes/Nextcloud/Zotero/Ceriotti et al_2011_Simplifying the representation of complex free-energy landscapes using.pdf;/home/johannes/Zotero/storage/GN7K44Z7/13023.html} +} + +@article{ceriottiUnsupervisedMachineLearning2019, + title = {Unsupervised Machine Learning in Atomistic Simulations, between Predictions and Understanding}, + author = {Ceriotti, Michele}, + date = {2019-04-19}, + journaltitle = {The Journal of Chemical Physics}, + volume = {150}, + number = {15}, + pages = {150901}, + publisher = {{AIP Publishing LLCAIP Publishing}}, + issn = {0021-9606}, + doi = {10.1063/1.5091842}, + url = {https://aip.scitation.org/doi/abs/10.1063/1.5091842}, + urldate = {2022-10-03}, + abstract = {Automated analyses of the outcome of a simulation have been an important part of atomistic modeling since the early days, addressing the need of linking the behavior of individual atoms and the collective properties that are usually the final quantity of interest. Methods such as clustering and dimensionality reduction have been used to provide a simplified, coarse-grained representation of the structure and dynamics of complex systems from proteins to nanoparticles. In recent years, the rise of machine learning has led to an even more widespread use of these algorithms in atomistic modeling and to consider different classification and inference techniques as part of a coherent toolbox of data-driven approaches. This perspective briefly reviews some of the unsupervised machine-learning methods—that are geared toward classification and coarse-graining of molecular simulations—seen in relation to the fundamental mathematical concepts that underlie all machine-learning techniques. It discusses the importance of using concise yet complete representations of atomic structures as the starting point of the analyses and highlights the risk of introducing preconceived biases when using machine learning to rationalize and understand structure-property relations. Supervised machine-learning techniques that explicitly attempt to predict the properties of a material given its structure are less susceptible to such biases. Current developments in the field suggest that using these two classes of approaches side-by-side and in a fully integrated mode, while keeping in mind the relations between the data analysis framework and the fundamental physical principles, will be key to realizing the full potential of machine learning to help understand the behavior of complex molecules and materials.}, + langid = {english}, + keywords = {clustering,data exploration,DBSCAN,dimensionality reduction,Gaussian mixture,ISOMAP,kernel methods,ML,multidimensional scaling,PCA,t-SNE,unsupervised learning}, + file = {/home/johannes/Nextcloud/Zotero/Ceriotti_2019_Unsupervised machine learning in atomistic simulations, between predictions and.pdf;/home/johannes/Zotero/storage/VYBXXJL4/1.html} +} + +@article{cersonskyImprovingSampleFeature2021, + title = {Improving Sample and Feature Selection with Principal Covariates Regression}, + author = {Cersonsky, Rose K. and Helfrecht, Benjamin A. and Engel, Edgar A. and Kliavinek, Sergei and Ceriotti, Michele}, + date = {2021-07}, + journaltitle = {Machine Learning: Science and Technology}, + shortjournal = {Mach. Learn.: Sci. Technol.}, + volume = {2}, + number = {3}, + pages = {035038}, + publisher = {{IOP Publishing}}, + issn = {2632-2153}, + doi = {10.1088/2632-2153/abfe7c}, + url = {https://doi.org/10.1088/2632-2153/abfe7c}, + urldate = {2022-08-10}, + abstract = {Selecting the most relevant features and samples out of a large set of candidates is a task that occurs very often in the context of automated data analysis, where it improves the computational performance and often the transferability of a model. Here we focus on two popular subselection schemes applied to this end: CUR decomposition, derived from a low-rank approximation of the feature matrix, and farthest point sampling (FPS), which relies on the iterative identification of the most diverse samples and discriminating features. We modify these unsupervised approaches, incorporating a supervised component following the same spirit as the principal covariates (PCov) regression method. We show how this results in selections that perform better in supervised tasks, demonstrating with models of increasing complexity, from ridge regression to kernel ridge regression and finally feed-forward neural networks. We also present adjustments to minimise the impact of any subselection when performing unsupervised tasks. We demonstrate the significant improvements associated with PCov-CUR and PCov-FPS selections for applications to chemistry and materials science, typically reducing by a factor of two the number of features and samples required to achieve a given level of regression accuracy.}, + langid = {english}, + keywords = {CUR decomposition,dimensionality reduction,feature selection,FPS,KRR,PCovR}, + file = {/home/johannes/Nextcloud/Zotero/Cersonsky et al_2021_Improving sample and feature selection with principal covariates regression.pdf;/home/johannes/Nextcloud/Zotero/false} +} + +@article{chandrasekaranSolvingElectronicStructure2019, + title = {Solving the Electronic Structure Problem with Machine Learning}, + author = {Chandrasekaran, Anand and Kamal, Deepak and Batra, Rohit and Kim, Chiho and Chen, Lihua and Ramprasad, Rampi}, + date = {2019-02-18}, + journaltitle = {npj Computational Materials}, + shortjournal = {npj Comput Mater}, + volume = {5}, + number = {1}, + pages = {1--7}, + publisher = {{Nature Publishing Group}}, + issn = {2057-3960}, + doi = {10.1038/s41524-019-0162-7}, + url = {https://www.nature.com/articles/s41524-019-0162-7}, + urldate = {2021-08-21}, + abstract = {Simulations based on solving the Kohn-Sham (KS) equation of density functional theory (DFT) have become a vital component of modern materials and chemical sciences research and development portfolios. Despite its versatility, routine DFT calculations are usually limited to a few hundred atoms due to the computational bottleneck posed by the KS equation. Here we introduce a machine-learning-based scheme to efficiently assimilate the function of the KS equation, and by-pass it to directly, rapidly, and accurately predict the electronic structure of a material or a molecule, given just its atomic configuration. A new rotationally invariant representation is utilized to map the atomic environment around a grid-point to the electron density and local density of states at that grid-point. This mapping is learned using a neural network trained on previously generated reference DFT results at millions of grid-points. The proposed paradigm allows for the high-fidelity emulation of KS DFT, but orders of magnitude faster than the direct solution. Moreover, the machine learning prediction scheme is strictly linear-scaling with system size.}, + issue = {1}, + langid = {english}, + keywords = {_tablet,custom structural descriptors,descriptors,DFT,FCNN,grid-based descriptors,LDOS,ML,ML-DFT,ML-ESM,models,NN,prediction from structure,prediction of electron density,prediction of LDOS,RNN}, + annotation = {Bandiera\_abtest: a Cc\_license\_type: cc\_by Cg\_type: Nature Research Journals Primary\_atype: Research Subject\_term: Computational methods;Electronic structure;Theory and computation Subject\_term\_id: computational-methods;electronic-structure;theory-and-computation}, + file = {/home/johannes/Nextcloud/Zotero/Chandrasekaran et al_2019_Solving the electronic structure problem with machine learning.pdf;/home/johannes/Nextcloud/Zotero/false;/home/johannes/Zotero/storage/TL92B668/s41524-019-0162-7.html} +} + +@article{changExperimentalObservationQuantum2013, + title = {Experimental {{Observation}} of the {{Quantum Anomalous Hall Effect}} in a {{Magnetic Topological Insulator}}}, + author = {Chang, Cui-Zu and Zhang, Jinsong and Feng, Xiao and Shen, Jie and Zhang, Zuocheng and Guo, Minghua and Li, Kang and Ou, Yunbo and Wei, Pang and Wang, Li-Li and Ji, Zhong-Qing and Feng, Yang and Ji, Shuaihua and Chen, Xi and Jia, Jinfeng and Dai, Xi and Fang, Zhong and Zhang, Shou-Cheng and He, Ke and Wang, Yayu and Lu, Li and Ma, Xu-Cun and Xue, Qi-Kun}, + date = {2013-04-12}, + journaltitle = {Science}, + volume = {340}, + number = {6129}, + pages = {167--170}, + publisher = {{American Association for the Advancement of Science}}, + doi = {10.1126/science.1234414}, + url = {https://www.science.org/doi/10.1126/science.1234414}, + urldate = {2022-05-13}, + file = {/home/johannes/Nextcloud/Zotero/Chang et al_2013_Experimental Observation of the Quantum Anomalous Hall Effect in a Magnetic.pdf} +} + +@unpublished{chardDLHubModelData2018, + title = {{{DLHub}}: {{Model}} and {{Data Serving}} for {{Science}}}, + shorttitle = {{{DLHub}}}, + author = {Chard, Ryan and Li, Zhuozhao and Chard, Kyle and Ward, Logan and Babuji, Yadu and Woodard, Anna and Tuecke, Steve and Blaiszik, Ben and Franklin, Michael J. and Foster, Ian}, + date = {2018-11-27}, + eprint = {1811.11213}, + eprinttype = {arxiv}, + primaryclass = {cs, stat}, + url = {http://arxiv.org/abs/1811.11213}, + urldate = {2022-01-03}, + abstract = {While the Machine Learning (ML) landscape is evolving rapidly, there has been a relative lag in the development of the "learning systems" needed to enable broad adoption. Furthermore, few such systems are designed to support the specialized requirements of scientific ML. Here we present the Data and Learning Hub for science (DLHub), a multi-tenant system that provides both model repository and serving capabilities with a focus on science applications. DLHub addresses two significant shortcomings in current systems. First, its selfservice model repository allows users to share, publish, verify, reproduce, and reuse models, and addresses concerns related to model reproducibility by packaging and distributing models and all constituent components. Second, it implements scalable and low-latency serving capabilities that can leverage parallel and distributed computing resources to democratize access to published models through a simple web interface. Unlike other model serving frameworks, DLHub can store and serve any Python 3-compatible model or processing function, plus multiple-function pipelines. We show that relative to other model serving systems including TensorFlow Serving, SageMaker, and Clipper, DLHub provides greater capabilities, comparable performance without memoization and batching, and significantly better performance when the latter two techniques can be employed. We also describe early uses of DLHub for scientific applications.}, + archiveprefix = {arXiv}, + keywords = {Computer Science - Distributed; Parallel; and Cluster Computing,Computer Science - Machine Learning,Statistics - Machine Learning}, + file = {/home/johannes/Nextcloud/Zotero/Chard et al_2018_DLHub.pdf;/home/johannes/Zotero/storage/VT5H6PP6/1811.html} +} + +@article{chenGraphNetworksUniversal2019, + title = {Graph {{Networks}} as a {{Universal Machine Learning Framework}} for {{Molecules}} and {{Crystals}}}, + author = {Chen, Chi and Ye, Weike and Zuo, Yunxing and Zheng, Chen and Ong, Shyue Ping}, + date = {2019-05-14}, + journaltitle = {Chemistry of Materials}, + shortjournal = {Chem. Mater.}, + volume = {31}, + number = {9}, + pages = {3564--3572}, + publisher = {{American Chemical Society}}, + issn = {0897-4756}, + doi = {10.1021/acs.chemmater.9b01294}, + url = {https://doi.org/10.1021/acs.chemmater.9b01294}, + urldate = {2022-01-02}, + abstract = {Graph networks are a new machine learning (ML) paradigm that supports both relational reasoning and combinatorial generalization. Here, we develop universal MatErials Graph Network (MEGNet) models for accurate property prediction in both molecules and crystals. We demonstrate that the MEGNet models outperform prior ML models such as the SchNet in 11 out of 13 properties of the QM9 molecule data set. Similarly, we show that MEGNet models trained on ∼60 000 crystals in the Materials Project substantially outperform prior ML models in the prediction of the formation energies, band gaps, and elastic moduli of crystals, achieving better than density functional theory accuracy over a much larger data set. We present two new strategies to address data limitations common in materials science and chemistry. First, we demonstrate a physically intuitive approach to unify four separate molecular MEGNet models for the internal energy at 0 K and room temperature, enthalpy, and Gibbs free energy into a single free energy MEGNet model by incorporating the temperature, pressure, and entropy as global state inputs. Second, we show that the learned element embeddings in MEGNet models encode periodic chemical trends and can be transfer-learned from a property model trained on a larger data set (formation energies) to improve property models with smaller amounts of data (band gaps and elastic moduli).}, + keywords = {_tablet,GNN,library,MEGNet,molecules,solids,vectorial learning target,with-code}, + file = {/home/johannes/Nextcloud/Zotero/Chen et al_2019_Graph Networks as a Universal Machine Learning Framework for Molecules and.pdf} +} + +@unpublished{chenUniversalGraphDeep2022, + title = {A {{Universal Graph Deep Learning Interatomic Potential}} for the {{Periodic Table}}}, + author = {Chen, Chi and Ong, Shyue Ping}, + date = {2022-02-04}, + eprint = {2202.02450}, + eprinttype = {arxiv}, + primaryclass = {cond-mat, physics:physics}, + url = {http://arxiv.org/abs/2202.02450}, + urldate = {2022-03-28}, + abstract = {Interatomic potentials (IAPs), which describe the potential energy surface of a collection of atoms, are a fundamental input for atomistic simulations. However, existing IAPs are either fitted to narrow chemistries or too inaccurate for general applications. Here, we report a universal IAP for materials based on graph neural networks with three-body interactions (M3GNet). The M3GNet IAP was trained on the massive database of structural relaxations performed by the Materials Project over the past 10 years and has broad applications in structural relaxation, dynamic simulations and property prediction of materials across diverse chemical spaces. About 1.8 million potentially stable materials were identified from a screening of 31 million hypothetical crystal structures, demonstrating a machine learning-accelerated pathway to the discovery of synthesizable materials with exceptional properties.}, + archiveprefix = {arXiv}, + keywords = {_tablet,condensed matter,GNN,library,M3GNet,materials,materials database,materials project,matterverse,MEGNet,ML,MLP,molecules,periodic table,solids,tensorial target,with-code}, + file = {/home/johannes/Nextcloud/Zotero/Chen_Ong_2022_A Universal Graph Deep Learning Interatomic Potential for the Periodic Table.pdf;/home/johannes/Zotero/storage/H4FKVKUF/2202.html} +} + +@article{choudharyAtomisticLineGraph2021, + title = {Atomistic {{Line Graph Neural Network}} for Improved Materials Property Predictions}, + author = {Choudhary, Kamal and DeCost, Brian}, + date = {2021-11-15}, + journaltitle = {npj Computational Materials}, + shortjournal = {npj Comput Mater}, + volume = {7}, + number = {1}, + pages = {1--8}, + publisher = {{Nature Publishing Group}}, + issn = {2057-3960}, + doi = {10.1038/s41524-021-00650-1}, + url = {https://www.nature.com/articles/s41524-021-00650-1}, + urldate = {2022-07-06}, + abstract = {Graph neural networks (GNN) have been shown to provide substantial performance improvements for atomistic material representation and modeling compared with descriptor-based machine learning models. While most existing GNN models for atomistic predictions are based on atomic distance information, they do not explicitly incorporate bond angles, which are critical for distinguishing many atomic structures. Furthermore, many material properties are known to be sensitive to slight changes in bond angles. We present an Atomistic Line Graph Neural Network (ALIGNN), a GNN architecture that performs message passing on both the interatomic bond graph and its line graph corresponding to bond angles. We demonstrate that angle information can be explicitly and efficiently included, leading to improved performance on multiple atomistic prediction tasks. We ALIGNN models for predicting 52 solid-state and molecular properties available in the JARVIS-DFT, Materials project, and QM9 databases. ALIGNN can outperform some previously reported GNN models on atomistic prediction tasks by up to 85\% in accuracy with better or comparable model training speed.}, + issue = {1}, + langid = {english}, + keywords = {ALIGNN,GNN,ML,MPNN,original publication}, + file = {/home/johannes/Nextcloud/Zotero/Choudhary_DeCost_2021_Atomistic Line Graph Neural Network for improved materials property predictions.pdf;/home/johannes/Zotero/storage/F8XSYTPV/s41524-021-00650-1.html} +} + +@article{choudharyMachineLearningForcefieldinspired2018, + title = {Machine Learning with Force-Field-Inspired Descriptors for Materials: {{Fast}} Screening and Mapping Energy Landscape}, + shorttitle = {Machine Learning with Force-Field-Inspired Descriptors for Materials}, + author = {Choudhary, Kamal and DeCost, Brian and Tavazza, Francesca}, + date = {2018-08-03}, + journaltitle = {Physical Review Materials}, + shortjournal = {Phys. Rev. Materials}, + volume = {2}, + number = {8}, + pages = {083801}, + publisher = {{American Physical Society}}, + doi = {10.1103/PhysRevMaterials.2.083801}, + url = {https://link.aps.org/doi/10.1103/PhysRevMaterials.2.083801}, + urldate = {2021-06-26}, + abstract = {We present a complete set of chemo-structural descriptors to significantly extend the applicability of machine learning (ML) in material screening and mapping the energy landscape for multicomponent systems. These descriptors allow differentiating between structural prototypes, which is not possible using the commonly used chemical-only descriptors. Specifically, we demonstrate that the combination of pairwise radial, nearest-neighbor, bond-angle, dihedral-angle, and core-charge distributions plays an important role in predicting formation energies, band gaps, static refractive indices, magnetic properties, and modulus of elasticity for three-dimensional materials as well as exfoliation energies of two-dimensional (2D)-layered materials. The training data consist of 24 549 bulk and 616 monolayer materials taken from the JARVIS-DFT database. We obtained very accurate ML models using a gradient-boosting algorithm. Then we use the trained models to discover exfoliable 2D-layered materials satisfying specific property requirements. Additionally, we integrate our formation-energy ML model with a genetic algorithm for structure search to verify if the ML model reproduces the density-functional-theory convex hull. This verification establishes a more stringent evaluation metric for the ML model than what is commonly used in data sciences. Our learned model is publicly available on the JARVIS-ML website (https://www.ctcms.nist.gov/jarvisml), property predictions of generalized materials.}, + keywords = {A3MDNet,CFID,classification of magnetic/nonmagnetic,classification of metal/insulator,descriptors,DFT,GBDT,JARVIS,ML,models}, + file = {/home/johannes/Nextcloud/Zotero/Choudhary et al_2018_Machine learning with force-field-inspired descriptors for materials Author's Manuscript.pdf;/home/johannes/Nextcloud/Zotero/Choudhary et al_2018_Machine learning with force-field-inspired descriptors for materials Suppl JARVIS-ML.pdf;/home/johannes/Nextcloud/Zotero/Choudhary et al_2018_Machine learning with force-field-inspired descriptors for materials.pdf;/home/johannes/Zotero/storage/88LWP9IL/Choudhary et al_2018_Machine learning with force-field-inspired descriptors for materials Suppl Feature Importance.xlsx;/home/johannes/Zotero/storage/8U5VA8X6/Choudhary et al_2018_Machine learning with force-field-inspired descriptors for materials Suppl Feature Importance.xlsx;/home/johannes/Zotero/storage/NYHDKNR3/PhysRevMaterials.2.html} +} + +@unpublished{chouhanImprovingScalabilityReliability2021, + title = {Improving Scalability and Reliability of {{MPI-agnostic}} Transparent Checkpointing for Production Workloads at {{NERSC}}}, + author = {Chouhan, Prashant Singh and Khetawat, Harsh and Resnik, Neil and Jain, Twinkle and Garg, Rohan and Cooperman, Gene and Hartman-Baker, Rebecca and Zhao, Zhengji}, + date = {2021-03-16}, + eprint = {2103.08546}, + eprinttype = {arxiv}, + primaryclass = {cs}, + url = {http://arxiv.org/abs/2103.08546}, + urldate = {2021-10-20}, + abstract = {Checkpoint/restart (C/R) provides fault-tolerant computing capability, enables long running applications, and provides scheduling flexibility for computing centers to support diverse workloads with different priority. It is therefore vital to get transparent C/R capability working at NERSC. MANA, by Garg et. al., is a transparent checkpointing tool that has been selected due to its MPI-agnostic and network-agnostic approach. However, originally written as a proof-of-concept code, MANA was not ready to use with NERSC's diverse production workloads, which are dominated by MPI and hybrid MPI+OpenMP applications. In this talk, we present ongoing work at NERSC to enable MANA for NERSC's production workloads, including fixing bugs that were exposed by the top applications at NERSC, adding new features to address system changes, evaluating C/R overhead at scale, etc. The lessons learned from making MANA production-ready for HPC applications will be useful for C/R tool developers, supercomputing centers and HPC end-users alike.}, + archiveprefix = {arXiv}, + keywords = {DFT,for introductions,master-thesis,supercomputing}, + file = {/home/johannes/Nextcloud/Zotero/Chouhan et al_2021_Improving scalability and reliability of MPI-agnostic transparent checkpointing.pdf;/home/johannes/Zotero/storage/BTL4HHE6/2103.html} +} + +@article{clementBenchmarkAFLOWData2020, + title = {Benchmark {{AFLOW Data Sets}} for {{Machine Learning}}}, + author = {Clement, Conrad L. and Kauwe, Steven K. and Sparks, Taylor D.}, + date = {2020-06-01}, + journaltitle = {Integrating Materials and Manufacturing Innovation}, + shortjournal = {Integr Mater Manuf Innov}, + volume = {9}, + number = {2}, + pages = {153--156}, + issn = {2193-9772}, + doi = {10.1007/s40192-020-00174-4}, + url = {https://doi.org/10.1007/s40192-020-00174-4}, + urldate = {2021-10-15}, + abstract = {Materials informatics is increasingly finding ways to exploit machine learning algorithms. Techniques such as decision trees, ensemble methods, support vector machines, and a variety of neural network architectures are used to predict likely material characteristics and property values. Supplemented with laboratory synthesis, applications of machine learning to compound discovery and characterization represent one of the most promising research directions in materials informatics. A shortcoming of this trend, in its current form, is a lack of standardized materials data sets on which to train, validate, and test model effectiveness. Applied machine learning research depends on benchmark data to make sense of its results. Fixed, predetermined data sets allow for rigorous model assessment and comparison. Machine learning publications that do not refer to benchmarks are often hard to contextualize and reproduce. In this data descriptor article, we present a collection of data sets of different material properties taken from the AFLOW database. We describe them, the procedures that generated them, and their use as potential benchmarks. We provide a compressed ZIP file containing the data sets and a GitHub repository of associated Python code. Finally, we discuss opportunities for future work incorporating the data sets and creating similar benchmark collections.}, + langid = {english}, + file = {/home/johannes/Nextcloud/Zotero/Clement et al_2020_Benchmark AFLOW Data Sets for Machine Learning.pdf} +} + +@unpublished{cobelliInversionChemicalEnvironment2022, + title = {Inversion of the Chemical Environment Representations}, + author = {Cobelli, Matteo and Cahalane, Paddy and Sanvito, Stefano}, + date = {2022-01-27}, + eprint = {2201.11591}, + eprinttype = {arxiv}, + primaryclass = {cond-mat}, + url = {http://arxiv.org/abs/2201.11591}, + urldate = {2022-03-23}, + abstract = {Machine-learning generative methods for material design are constructed by representing a given chemical structure, either a solid or a molecule, over appropriate atomic features, generally called structural descriptors. These must be fully descriptive of the system, must facilitate the training process and must be invertible, so that one can extract the atomic configurations corresponding to the output of the model. In general, this last requirement is not automatically satisfied by the most efficient structural descriptors, namely the representation is not directly invertible. Such drawback severely limits our freedom of choice in selecting the most appropriate descriptors for the problem, and thus our flexibility to construct generative models. In this work, we present a general optimization method capable of inverting any local many-body descriptor of the chemical environment, back to a cartesian representation. The algorithm is then implemented together with the bispectrum representation of the local structure and demonstrated for a number of molecules. The scheme presented here, thus, represents a general approach to the inversion of structural descriptors, enabling the construction of efficient structural generative models.}, + archiveprefix = {arXiv}, + keywords = {descriptors,generative models,inversion,ML}, + file = {/home/johannes/Nextcloud/Zotero/Cobelli et al_2022_Inversion of the chemical environment representations.pdf;/home/johannes/Zotero/storage/A6MH6ZIG/2201.html} +} + +@article{collinsHumanGenomeProject2003, + title = {The {{Human Genome Project}}: {{Lessons}} from {{Large-Scale Biology}}}, + shorttitle = {The {{Human Genome Project}}}, + author = {Collins, Francis S. and Morgan, Michael and Patrinos, Aristides}, + date = {2003-04-11}, + journaltitle = {Science}, + volume = {300}, + number = {5617}, + pages = {286--290}, + publisher = {{American Association for the Advancement of Science}}, + doi = {10.1126/science.1084564}, + url = {https://www.science.org/doi/full/10.1126/science.1084564}, + urldate = {2021-10-15}, + file = {/home/johannes/Nextcloud/Zotero/Collins et al_2003_The Human Genome Project.pdf} +} + +@article{cuevas-zuviriaMachineLearningAnalytical2021, + title = {Machine {{Learning}} of {{Analytical Electron Density}} in {{Large Molecules Through Message-Passing}}}, + author = {Cuevas-ZuvirÃa, Bruno and Pacios, Luis F.}, + date = {2021-06-28}, + journaltitle = {Journal of Chemical Information and Modeling}, + shortjournal = {J. Chem. Inf. Model.}, + volume = {61}, + number = {6}, + pages = {2658--2666}, + publisher = {{American Chemical Society}}, + issn = {1549-9596}, + doi = {10.1021/acs.jcim.1c00227}, + url = {https://doi.org/10.1021/acs.jcim.1c00227}, + urldate = {2022-07-10}, + abstract = {Machine learning milestones in computational chemistry are overshadowed by their unaccountability and the overwhelming zoo of tools for each specific task. A promising path to tackle these problems is using machine learning to reproduce physical magnitudes as a basis to derive many other properties. By using a model of the electron density consisting of an analytical expansion on a linear set of isotropic and anisotropic functions, we implemented in this work a message-passing neural network able to reproduce electron density in molecules with just a 2.5\% absolute error in complex cases. We also adapted our methodology to describe electron density in large biomolecules (proteins) and to obtain atomic charges, interaction energies, and DFT energies. We show that electron density learning is a new promising avenue with a variety of forthcoming applications.}, + keywords = {analytical model,GCN,GNN,ML,molecules,MPNN,prediction of electron density}, + file = {/home/johannes/Nextcloud/Zotero/Cuevas-ZuvirÃa_Pacios_2021_Machine Learning of Analytical Electron Density in Large Molecules Through.pdf} +} + +@article{curtaroloAFLOWAutomaticFramework2012, + title = {{{AFLOW}}: {{An}} Automatic Framework for High-Throughput Materials Discovery}, + shorttitle = {{{AFLOW}}}, + author = {Curtarolo, Stefano and Setyawan, Wahyu and Hart, Gus L. W. and Jahnatek, Michal and Chepulskii, Roman V. and Taylor, Richard H. and Wang, Shidong and Xue, Junkai and Yang, Kesong and Levy, Ohad and Mehl, Michael J. and Stokes, Harold T. and Demchenko, Denis O. and Morgan, Dane}, + date = {2012-06-01}, + journaltitle = {Computational Materials Science}, + shortjournal = {Computational Materials Science}, + volume = {58}, + pages = {218--226}, + issn = {0927-0256}, + doi = {10.1016/j.commatsci.2012.02.005}, + url = {https://www.sciencedirect.com/science/article/pii/S0927025612000717}, + urldate = {2021-10-17}, + abstract = {Recent advances in computational materials science present novel opportunities for structure discovery and optimization, including uncovering of unsuspected compounds and metastable structures, electronic structure, surface, and nano-particle properties. The practical realization of these opportunities requires systematic generation and classification of the relevant computational data by high-throughput methods. In this paper we present Aflow (Automatic Flow), a software framework for high-throughput calculation of crystal structure properties of alloys, intermetallics and inorganic compounds. The Aflow software is available for the scientific community on the website of the materials research consortium, aflowlib.org. Its geometric and electronic structure analysis and manipulation tools are additionally available for online operation at the same website. The combination of automatic methods and user online interfaces provide a powerful tool for efficient quantum computational materials discovery and characterization.}, + langid = {english}, + keywords = {Ab initio,AFLOW,Combinatorial materials science,High-throughput}, + file = {/home/johannes/Nextcloud/Zotero/Curtarolo et al_2012_AFLOW.pdf;/home/johannes/Zotero/storage/3ZKE8YHP/S0927025612000717.html} +} + +@article{curtaroloAFLOWLIBORGDistributed2012, + title = {{{AFLOWLIB}}.{{ORG}}: {{A}} Distributed Materials Properties Repository from High-Throughput Ab Initio Calculations}, + shorttitle = {{{AFLOWLIB}}.{{ORG}}}, + author = {Curtarolo, Stefano and Setyawan, Wahyu and Wang, Shidong and Xue, Junkai and Yang, Kesong and Taylor, Richard H. and Nelson, Lance J. and Hart, Gus L. W. and Sanvito, Stefano and Buongiorno-Nardelli, Marco and Mingo, Natalio and Levy, Ohad}, + date = {2012-06-01}, + journaltitle = {Computational Materials Science}, + shortjournal = {Computational Materials Science}, + volume = {58}, + pages = {227--235}, + issn = {0927-0256}, + doi = {10.1016/j.commatsci.2012.02.002}, + url = {https://www.sciencedirect.com/science/article/pii/S0927025612000687}, + urldate = {2021-10-17}, + abstract = {Empirical databases of crystal structures and thermodynamic properties are fundamental tools for materials research. Recent rapid proliferation of computational data on materials properties presents the possibility to complement and extend the databases where the experimental data is lacking or difficult to obtain. Enhanced repositories that integrate both computational and empirical approaches open novel opportunities for structure discovery and optimization, including uncovering of unsuspected compounds, metastable structures and correlations between various characteristics. The practical realization of these opportunities depends on a systematic compilation and classification of the generated data in addition to an accessible interface for the materials science community. In this paper we present an extensive repository, aflowlib.org, comprising phase-diagrams, electronic structure and magnetic properties, generated by the high-throughput framework AFLOW. This continuously updated compilation currently contains over 150,000 thermodynamic entries for alloys, covering the entire composition range of more than 650 binary systems, 13,000 electronic structure analyses of inorganic compounds, and 50,000 entries for novel potential magnetic and spintronics systems. The repository is available for the scientific community on the website of the materials research consortium, aflowlib.org.}, + langid = {english}, + keywords = {Ab initio,AFLOW,Combinatorial materials science,High-throughput,Materials genome initiative}, + file = {/home/johannes/Zotero/storage/5MTYTHXV/S0927025612000687.html} +} + +@unpublished{darbyCompressingLocalAtomic2021, + title = {Compressing Local Atomic Neighbourhood Descriptors}, + author = {Darby, James P. and Kermode, James R. and Csányi, Gábor}, + date = {2021-12-24}, + eprint = {2112.13055}, + eprinttype = {arxiv}, + primaryclass = {cond-mat}, + url = {http://arxiv.org/abs/2112.13055}, + urldate = {2022-01-03}, + abstract = {Many atomic descriptors are currently limited by their unfavourable scaling with the number of chemical elements \$S\$ e.g. the length of body-ordered descriptors, such as the Smooth Overlap of Atomic Positions (SOAP) power spectrum (3-body) and the Atomic Cluster Expansion (ACE) (multiple body-orders), scales as \$(NS)\^\textbackslash nu\$ where \$\textbackslash nu+1\$ is the body-order and \$N\$ is the number of radial basis functions used in the density expansion. We introduce two distinct approaches which can be used to overcome this scaling for the SOAP power spectrum. Firstly, we show that the power spectrum is amenable to lossless compression with respect to both \$S\$ and \$N\$, so that the descriptor length can be reduced from \$\textbackslash mathcal\{O\}(N\^2S\^2)\$ to \$\textbackslash mathcal\{O\}\textbackslash left(NS\textbackslash right)\$. Secondly, we introduce a generalized SOAP kernel, where compression is achieved through the use of the total, element agnostic density, in combination with radial projection. The ideas used in the generalized kernel are equally applicably to any other body-ordered descriptors and we demonstrate this for the Atom Centered Symmetry Functions (ACSF). Finally, both compression approaches are shown to offer comparable performance to the original descriptor across a variety of numerical tests.}, + archiveprefix = {arXiv}, + keywords = {_tablet,ACE,ACSF,chemical species scaling problem,descriptor dimred,descriptors,descriptors analysis,dimensionality reduction,library,ML,SOAP,with-code}, + file = {/home/johannes/Nextcloud/Zotero/Darby et al_2021_Compressing local atomic neighbourhood descriptors.pdf;/home/johannes/Zotero/storage/GXXQQPAA/2112.html} +} + +@article{darbyCompressingLocalAtomic2022, + title = {Compressing Local Atomic Neighbourhood Descriptors}, + author = {Darby, James P. and Kermode, James R. and Csányi, Gábor}, + date = {2022-08-11}, + journaltitle = {npj Computational Materials}, + shortjournal = {npj Comput Mater}, + volume = {8}, + number = {1}, + pages = {1--13}, + publisher = {{Nature Publishing Group}}, + issn = {2057-3960}, + doi = {10.1038/s41524-022-00847-y}, + url = {https://www.nature.com/articles/s41524-022-00847-y}, + urldate = {2022-09-27}, + abstract = {Many atomic descriptors are currently limited by their unfavourable scaling with the number of chemical elements S e.g. the length of body-ordered descriptors, such as the SOAP power spectrum (3-body) and the (ACE) (multiple body-orders), scales as (NS)ν where ν\,+\,1 is the body-order and N is the number of radial basis functions used in the density expansion. We introduce two distinct approaches which can be used to overcome this scaling for the SOAP power spectrum. Firstly, we show that the power spectrum is amenable to lossless compression with respect to both S and N, so that the descriptor length can be reduced from \$\$\{\{\{\textbackslash mathcal\{O\}\}\}\}(\{N\}\^\{2\}\{S\}\^\{2\})\$\$to \$\$\{\{\{\textbackslash mathcal\{O\}\}\}\}\textbackslash left(NS\textbackslash right)\$\$. Secondly, we introduce a generalised SOAP kernel, where compression is achieved through the use of the total, element agnostic density, in combination with radial projection. The ideas used in the generalised kernel are equally applicably to any other body-ordered descriptors and we demonstrate this for the (ACSF).}, + issue = {1}, + langid = {english}, + keywords = {ACE,ACSF,chemical species scaling problem,descriptor dimred,descriptors,descriptors analysis,dimensionality reduction,library,ML,SOAP,with-code}, + file = {/home/johannes/Nextcloud/Zotero/Darby et al_2022_Compressing local atomic neighbourhood descriptors.pdf;/home/johannes/Zotero/storage/WR6IJ7MC/s41524-022-00847-y.html} +} + +@misc{darbyTensorreducedAtomicDensity2022, + title = {Tensor-Reduced Atomic Density Representations}, + author = {Darby, James P. and Kovács, Dávid P. and Batatia, Ilyes and Caro, Miguel A. and Hart, Gus L. W. and Ortner, Christoph and Csányi, Gábor}, + date = {2022-10-01}, + number = {arXiv:2210.01705}, + eprint = {2210.01705}, + eprinttype = {arxiv}, + primaryclass = {cond-mat, physics:physics}, + publisher = {{arXiv}}, + doi = {10.48550/arXiv.2210.01705}, + url = {http://arxiv.org/abs/2210.01705}, + urldate = {2022-10-05}, + abstract = {Density based representations of atomic environments that are invariant under Euclidean symmetries have become a widely used tool in the machine learning of interatomic potentials, broader data-driven atomistic modelling and the visualisation and analysis of materials datasets.The standard mechanism used to incorporate chemical element information is to create separate densities for each element and form tensor products between them. This leads to a steep scaling in the size of the representation as the number of elements increases. Graph neural networks, which do not explicitly use density representations, escape this scaling by mapping the chemical element information into a fixed dimensional space in a learnable way. We recast this approach as tensor factorisation by exploiting the tensor structure of standard neighbour density based descriptors. In doing so, we form compact tensor-reduced representations whose size does not depend on the number of chemical elements, but remain systematically convergeable and are therefore applicable to a wide range of data analysis and regression tasks.}, + archiveprefix = {arXiv}, + keywords = {ACE,chemical species scaling problem,descriptor dimred,descriptors,dimensionality reduction,MACE,ML,Multi-ACE}, + file = {/home/johannes/Nextcloud/Zotero/Darby et al_2022_Tensor-reduced atomic density representations.pdf;/home/johannes/Zotero/storage/6XMXCLL4/2210.html} +} + +@article{dasCrysXPPExplainableProperty2022, + title = {{{CrysXPP}}: {{An}} Explainable Property Predictor for Crystalline Materials}, + shorttitle = {{{CrysXPP}}}, + author = {Das, Kishalay and Samanta, Bidisha and Goyal, Pawan and Lee, Seung-Cheol and Bhattacharjee, Satadeep and Ganguly, Niloy}, + date = {2022-03-18}, + journaltitle = {npj Computational Materials}, + shortjournal = {npj Comput Mater}, + volume = {8}, + number = {1}, + pages = {1--11}, + publisher = {{Nature Publishing Group}}, + issn = {2057-3960}, + doi = {10.1038/s41524-022-00716-8}, + url = {https://www.nature.com/articles/s41524-022-00716-8}, + urldate = {2022-09-27}, + abstract = {We present a deep-learning framework, CrysXPP, to allow rapid and accurate prediction of electronic, magnetic, and elastic properties of a wide range of materials. CrysXPP lowers the need for large property tagged datasets by intelligently designing an autoencoder, CrysAE. The important structural and chemical properties captured by CrysAE from a large amount of available crystal graphs data helped in achieving low prediction errors. Moreover, we design a feature selector that helps to interpret the model’s prediction. Most notably, when given a small amount of experimental data, CrysXPP is consistently able to outperform conventional DFT. A detailed ablation study establishes the importance of different design steps. We release the large pre-trained model CrysAE. We believe by fine-tuning the model with a small amount of property-tagged data, researchers can achieve superior performance on various applications with a restricted data source.}, + issue = {1}, + langid = {english}, + keywords = {autoencoder,CGCNN,CrysXPP,dimensionality reduction,feature selection,GCN,GNN,library,solids,with-code}, + file = {/home/johannes/Nextcloud/Zotero/Das et al_2022_CrysXPP.pdf;/home/johannes/Zotero/storage/SY9ITHM4/s41524-022-00716-8.html} +} + +@inproceedings{davidsonProvenanceScientificWorkflows2008, + title = {Provenance and Scientific Workflows: Challenges and Opportunities}, + shorttitle = {Provenance and Scientific Workflows}, + booktitle = {Proceedings of the 2008 {{ACM SIGMOD}} International Conference on {{Management}} of Data}, + author = {Davidson, Susan B. and Freire, Juliana}, + date = {2008-06-09}, + series = {{{SIGMOD}} '08}, + pages = {1345--1350}, + publisher = {{Association for Computing Machinery}}, + location = {{New York, NY, USA}}, + doi = {10.1145/1376616.1376772}, + url = {https://doi.org/10.1145/1376616.1376772}, + urldate = {2021-10-17}, + abstract = {Provenance in the context of workflows, both for the data they derive and for their specification, is an essential component to allow for result reproducibility, sharing, and knowledge re-use in the scientific community. Several workshops have been held on the topic, and it has been the focus of many research projects and prototype systems. This tutorial provides an overview of research issues in provenance for scientific workflows, with a focus on recent literature and technology in this area. It is aimed at a general database research audience and at people who work with scientific data and workflows. We will (1) provide a general overview of scientific workflows, (2) describe research on provenance for scientific workflows and show in detail how provenance is supported in existing systems; (3) discuss emerging applications that are enabled by provenance; and (4) outline open problems and new directions for database-related research.}, + isbn = {978-1-60558-102-6}, + keywords = {provenance,scientific workflows}, + file = {/home/johannes/Nextcloud/Zotero/Davidson_Freire_2008_Provenance and scientific workflows.pdf} +} + +@article{deComparingMoleculesSolids2016, + title = {Comparing Molecules and Solids across Structural and Alchemical Space}, + author = {De, Sandip and Bartók, Albert P. and Csányi, Gábor and Ceriotti, Michele}, + date = {2016-05-18}, + journaltitle = {Physical Chemistry Chemical Physics}, + shortjournal = {Phys. Chem. Chem. Phys.}, + volume = {18}, + number = {20}, + pages = {13754--13769}, + publisher = {{The Royal Society of Chemistry}}, + issn = {1463-9084}, + doi = {10.1039/C6CP00415F}, + url = {https://pubs.rsc.org/en/content/articlelanding/2016/cp/c6cp00415f}, + urldate = {2021-05-13}, + abstract = {Evaluating the (dis)similarity of crystalline, disordered and molecular compounds is a critical step in the development of algorithms to navigate automatically the configuration space of complex materials. For instance, a structural similarity metric is crucial for classifying structures, searching chemical space for better compounds and materials, and driving the next generation of machine-learning techniques for predicting the stability and properties of molecules and materials. In the last few years several strategies have been designed to compare atomic coordination environments. In particular, the smooth overlap of atomic positions (SOAPs) has emerged as an elegant framework to obtain translation, rotation and permutation-invariant descriptors of groups of atoms, underlying the development of various classes of machine-learned inter-atomic potentials. Here we discuss how one can combine such local descriptors using a regularized entropy match (REMatch) approach to describe the similarity of both whole molecular and bulk periodic structures, introducing powerful metrics that enable the navigation of alchemical and structural complexities within a unified framework. Furthermore, using this kernel and a ridge regression method we can predict atomization energies for a database of small organic molecules with a mean absolute error below 1 kcal mol−1, reaching an important milestone in the application of machine-learning techniques for the evaluation of molecular properties.}, + langid = {english}, + keywords = {classification,descriptors,kernel methods,ML,rec-by-tim-wuerger,REMatch,similarity analysis,SOAP}, + file = {/home/johannes/Nextcloud/Zotero/De et al_2016_Comparing molecules and solids across structural and alchemical space.pdf;/home/johannes/Zotero/storage/SA8QCH28/C6CP00415F.html} +} + +@article{degraveMagneticControlTokamak2022, + title = {Magnetic Control of Tokamak Plasmas through Deep Reinforcement Learning}, + author = {Degrave, Jonas and Felici, Federico and Buchli, Jonas and Neunert, Michael and Tracey, Brendan and Carpanese, Francesco and Ewalds, Timo and Hafner, Roland and Abdolmaleki, Abbas and de las Casas, Diego and Donner, Craig and Fritz, Leslie and Galperti, Cristian and Huber, Andrea and Keeling, James and Tsimpoukelli, Maria and Kay, Jackie and Merle, Antoine and Moret, Jean-Marc and Noury, Seb and Pesamosca, Federico and Pfau, David and Sauter, Olivier and Sommariva, Cristian and Coda, Stefano and Duval, Basil and Fasoli, Ambrogio and Kohli, Pushmeet and Kavukcuoglu, Koray and Hassabis, Demis and Riedmiller, Martin}, + options = {useprefix=true}, + date = {2022-02}, + journaltitle = {Nature}, + volume = {602}, + number = {7897}, + pages = {414--419}, + publisher = {{Nature Publishing Group}}, + issn = {1476-4687}, + doi = {10.1038/s41586-021-04301-9}, + url = {https://www.nature.com/articles/s41586-021-04301-9}, + urldate = {2022-03-29}, + abstract = {Nuclear fusion using magnetic confinement, in particular in the tokamak configuration, is a promising path towards sustainable energy. A core challenge is to shape and maintain a high-temperature plasma within the tokamak vessel. This requires high-dimensional, high-frequency, closed-loop control using magnetic actuator coils, further complicated by the diverse requirements across a wide range of plasma configurations. In this work, we introduce a previously undescribed architecture for tokamak magnetic controller design that autonomously learns to command the full set of control coils. This architecture meets control objectives specified at a high level, at the same time satisfying physical and operational constraints. This approach has unprecedented flexibility and generality in problem specification and yields a notable reduction in design effort to produce new plasma configurations. We successfully produce and control a diverse set of plasma configurations on the Tokamak à Configuration Variable1,2, including elongated, conventional shapes, as well as advanced configurations, such as negative triangularity and ‘snowflake’ configurations. Our approach achieves accurate tracking of the location, current and shape for these configurations. We also demonstrate sustained ‘droplets’ on TCV, in which two separate plasmas are maintained simultaneously within the vessel. This represents a notable advance for tokamak feedback control, showing the potential of reinforcement learning to accelerate research in the fusion domain, and is one of the most challenging real-world systems to which reinforcement learning has been applied.}, + issue = {7897}, + langid = {english}, + keywords = {Computer science,Magnetically confined plasmas,ML,Nuclear fusion and fission,rec-by-bluegel}, + file = {/home/johannes/Nextcloud/Zotero/Degrave et al_2022_Magnetic control of tokamak plasmas through deep reinforcement learning.pdf;/home/johannes/Zotero/storage/U6PRS6KM/s41586-021-04301-9.html} +} + +@article{dengImageNetLargescaleHierarchical2009, + title = {{{ImageNet}}: {{A}} Large-Scale Hierarchical Image Database}, + shorttitle = {{{ImageNet}}}, + author = {Deng, J. and Dong, Wei and Socher, R. and Li, Li-Jia and Li, K. and Fei-Fei, Li}, + date = {2009}, + journaltitle = {2009 IEEE Conference on Computer Vision and Pattern Recognition}, + doi = {10.1109/cvprw.2009.5206848}, + abstract = {The explosion of image data on the Internet has the potential to foster more sophisticated and robust models and algorithms to index, retrieve, organize and interact with images and multimedia data. But exactly how such data can be harnessed and organized remains a critical problem. We introduce here a new database called “ImageNetâ€, a large-scale ontology of images built upon the backbone of the WordNet structure. ImageNet aims to populate the majority of the 80,000 synsets of WordNet with an average of 500-1000 clean and full resolution images. This will result in tens of millions of annotated images organized by the semantic hierarchy of WordNet. This paper offers a detailed analysis of ImageNet in its current state: 12 subtrees with 5247 synsets and 3.2 million images in total. We show that ImageNet is much larger in scale and diversity and much more accurate than the current image datasets. Constructing such a large-scale database is a challenging task. We describe the data collection scheme with Amazon Mechanical Turk. Lastly, we illustrate the usefulness of ImageNet through three simple applications in object recognition, image classification and automatic object clustering. We hope that the scale, accuracy, diversity and hierarchical structure of ImageNet can offer unparalleled opportunities to researchers in the computer vision community and beyond.} +} + +@article{dengMachineLearningTopological2017, + title = {Machine Learning Topological States}, + author = {Deng, Dong-Ling and Li, Xiaopeng and Das Sarma, S.}, + date = {2017-11-22}, + journaltitle = {Physical Review B}, + shortjournal = {Phys. Rev. B}, + volume = {96}, + number = {19}, + pages = {195145}, + publisher = {{American Physical Society}}, + doi = {10.1103/PhysRevB.96.195145}, + url = {https://link.aps.org/doi/10.1103/PhysRevB.96.195145}, + urldate = {2021-05-21}, + abstract = {Artificial neural networks and machine learning have now reached a new era after several decades of improvement where applications are to explode in many fields of science, industry, and technology. Here, we use artificial neural networks to study an intriguing phenomenon in quantum physics—the topological phases of matter. We find that certain topological states, either symmetry-protected or with intrinsic topological order, can be represented with classical artificial neural networks. This is demonstrated by using three concrete spin systems, the one-dimensional (1D) symmetry-protected topological cluster state and the 2D and 3D toric code states with intrinsic topological orders. For all three cases, we show rigorously that the topological ground states can be represented by short-range neural networks in an exact and efficient fashion—the required number of hidden neurons is as small as the number of physical spins and the number of parameters scales only linearly with the system size. For the 2D toric-code model, we find that the proposed short-range neural networks can describe the excited states with Abelian anyons and their nontrivial mutual statistics as well. In addition, by using reinforcement learning we show that neural networks are capable of finding the topological ground states of nonintegrable Hamiltonians with strong interactions and studying their topological phase transitions. Our results demonstrate explicitly the exceptional power of neural networks in describing topological quantum states, and at the same time provide valuable guidance to machine learning of topological phases in generic lattice models.}, + keywords = {ANN,ML,reinforcement-learning,topological phase,topological phase transition}, + file = {/home/johannes/Nextcloud/Zotero/Deng et al_2017_Machine learning topological states.pdf;/home/johannes/Zotero/storage/QSGREI8E/PhysRevB.96.html} +} + +@article{dennerEfficientLearningOnedimensional2020, + title = {Efficient Learning of a One-Dimensional Density Functional Theory}, + author = {Denner, M. Michael}, + date = {2020}, + journaltitle = {Physical Review Research}, + shortjournal = {Phys. Rev. Research}, + volume = {2}, + number = {3}, + doi = {10.1103/PhysRevResearch.2.033388}, + keywords = {DFT,ML,ML-DFT,ML-ESM,prediction of ground-state properties,topological phase transition}, + file = {/home/johannes/Nextcloud/Zotero/Denner_2020_Efficient learning of a one-dimensional density functional theory.pdf;/home/johannes/Zotero/storage/UHHVADW4/PhysRevResearch.2.html} +} + +@article{depabloNewFrontiersMaterials2019, + title = {New Frontiers for the Materials Genome Initiative}, + author = {de Pablo, Juan J. and Jackson, Nicholas E. and Webb, Michael A. and Chen, Long-Qing and Moore, Joel E. and Morgan, Dane and Jacobs, Ryan and Pollock, Tresa and Schlom, Darrell G. and Toberer, Eric S. and Analytis, James and Dabo, Ismaila and DeLongchamp, Dean M. and Fiete, Gregory A. and Grason, Gregory M. and Hautier, Geoffroy and Mo, Yifei and Rajan, Krishna and Reed, Evan J. and Rodriguez, Efrain and Stevanovic, Vladan and Suntivich, Jin and Thornton, Katsuyo and Zhao, Ji-Cheng}, + options = {useprefix=true}, + date = {2019-04-05}, + journaltitle = {npj Computational Materials}, + shortjournal = {npj Comput Mater}, + volume = {5}, + number = {1}, + pages = {1--23}, + publisher = {{Nature Publishing Group}}, + issn = {2057-3960}, + doi = {10.1038/s41524-019-0173-4}, + url = {https://www.nature.com/articles/s41524-019-0173-4}, + urldate = {2021-10-15}, + abstract = {The Materials Genome Initiative (MGI) advanced a new paradigm for materials discovery and design, namely that the pace of new materials deployment could be accelerated through complementary efforts in theory, computation, and experiment. Along with numerous successes, new challenges are inviting researchers to refocus the efforts and approaches that were originally inspired by the MGI. In May 2017, the National Science Foundation sponsored the workshop “Advancing and Accelerating Materials Innovation Through the Synergistic Interaction among Computation, Experiment, and Theory: Opening New Frontiers†to review accomplishments that emerged from investments in science and infrastructure under the MGI, identify scientific opportunities in this new environment, examine how to effectively utilize new materials innovation infrastructure, and discuss challenges in achieving accelerated materials research through the seamless integration of experiment, computation, and theory. This article summarizes key findings from the workshop and provides perspectives that aim to guide the direction of future materials research and its translation into societal impacts.}, + issue = {1}, + langid = {english}, + annotation = {Bandiera\_abtest: a Cc\_license\_type: cc\_by Cg\_type: Nature Research Journals Primary\_atype: Reviews Subject\_term: Materials science;Nanoscience and technology Subject\_term\_id: materials-science;nanoscience-and-technology}, + file = {/home/johannes/Nextcloud/Zotero/de Pablo et al_2019_New frontiers for the materials genome initiative.pdf;/home/johannes/Zotero/storage/PY8DXX7D/s41524-019-0173-4.html} +} + +@article{deringerGaussianProcessRegression2021, + title = {Gaussian {{Process Regression}} for {{Materials}} and {{Molecules}}}, + author = {Deringer, Volker L. and Bartók, Albert P. and Bernstein, Noam and Wilkins, David M. and Ceriotti, Michele and Csányi, Gábor}, + date = {2021-08-25}, + journaltitle = {Chemical Reviews}, + shortjournal = {Chem. Rev.}, + volume = {121}, + number = {16}, + pages = {10073--10141}, + publisher = {{American Chemical Society}}, + issn = {0009-2665}, + doi = {10.1021/acs.chemrev.1c00022}, + url = {https://doi.org/10.1021/acs.chemrev.1c00022}, + urldate = {2022-06-03}, + abstract = {We provide an introduction to Gaussian process regression (GPR) machine-learning methods in computational materials science and chemistry. The focus of the present review is on the regression of atomistic properties: in particular, on the construction of interatomic potentials, or force fields, in the Gaussian Approximation Potential (GAP) framework; beyond this, we also discuss the fitting of arbitrary scalar, vectorial, and tensorial quantities. Methodological aspects of reference data generation, representation, and regression, as well as the question of how a data-driven model may be validated, are reviewed and critically discussed. A survey of applications to a variety of research questions in chemistry and materials science illustrates the rapid growth in the field. A vision is outlined for the development of the methodology in the years to come.}, + keywords = {_tablet,active learning,GAP,GPR,librascal,materials,ML,ML-DFT,ML-ESM,MLP,models,molecules,prediction of electron density,prediction of LDOS,review,SA-GPR,SOAP,structure prediction,structure search,tutorial,with-code}, + file = {/home/johannes/Nextcloud/Zotero/Deringer et al_2021_Gaussian Process Regression for Materials and Molecules.pdf;/home/johannes/Zotero/storage/LSTJST2A/acs.chemrev.html} +} + +@article{diceSignacDataManagement2021, + title = {Signac: {{Data Management}} and {{Workflows}} for {{Computational Researchers}}}, + shorttitle = {Signac}, + author = {Dice, Bradley D. and Butler, Brandon L. and Ramasubramani, Vyas and Travitz, Alyssa and Henry, Michael M. and Ojha, Hardik and Wang, Kelly L. and Adorf, Carl S. and Jankowski, Eric and Glotzer, Sharon C.}, + date = {2021}, + journaltitle = {Proceedings of the 20th Python in Science Conference}, + pages = {23--32}, + doi = {10.25080/majora-1b6fd038-003}, + url = {https://conference.scipy.org/proceedings/scipy2021/bradley_dice.html}, + urldate = {2022-08-24}, + eventtitle = {Proceedings of the 20th {{Python}} in {{Science Conference}}}, + keywords = {Data management,RDM}, + file = {/home/johannes/Nextcloud/Zotero/Dice et al_2021_signac.pdf;/home/johannes/Zotero/storage/YML5Z3T2/bradley_dice.html} +} + +@article{dickHighlyAccurateConstrained2021, + title = {Highly Accurate and Constrained Density Functional Obtained with Differentiable Programming}, + author = {Dick, Sebastian and Fernandez-Serra, Marivi}, + date = {2021-10-12}, + journaltitle = {Physical Review B}, + shortjournal = {Phys. Rev. B}, + volume = {104}, + number = {16}, + pages = {L161109}, + publisher = {{American Physical Society}}, + doi = {10.1103/PhysRevB.104.L161109}, + url = {https://link.aps.org/doi/10.1103/PhysRevB.104.L161109}, + urldate = {2021-12-14}, + abstract = {Using an end-to-end differentiable implementation of the Kohn-Sham self-consistent field equations, we obtain a highly accurate neural network–based exchange and correlation (XC) functional of the electronic density. The functional is optimized using information on both energy and density while exact constraints are enforced through an appropriate neural network architecture. We evaluate our model against different families of XC approximations and show that at the meta-GGA level our functional exhibits unprecedented accuracy for both energy and density predictions. For nonempirical functionals, there is a strong linear correlation between energy and density errors. We use this correlation to define an XC functional quality metric that includes both energy and density errors, leading to an improved way to rank different approximations.}, + keywords = {ANN,autodiff,DFT,library,ML,ML-DFA,molecules,prediction from density,prediction of Exc,prediction of vxc,with-code}, + file = {/home/johannes/Nextcloud/Zotero/Dick_Fernandez-Serra_2021_Highly accurate and constrained density functional obtained with differentiable.pdf;/home/johannes/Zotero/storage/C5259YA7/Dick and Fernandez-Serra - 2021 - Highly accurate and constrained density functional.pdf;/home/johannes/Zotero/storage/SDPZD88A/PhysRevB.104.html} +} + +@article{dickLearningDensityCorrect2019, + title = {Learning from the Density to Correct Total Energy and Forces in First Principle Simulations}, + author = {Dick, Sebastian and Fernandez-Serra, Marivi}, + date = {2019-10-14}, + journaltitle = {The Journal of Chemical Physics}, + shortjournal = {J. Chem. Phys.}, + volume = {151}, + number = {14}, + pages = {144102}, + publisher = {{American Institute of Physics}}, + issn = {0021-9606}, + doi = {10.1063/1.5114618}, + url = {https://aip.scitation.org/doi/full/10.1063/1.5114618}, + urldate = {2021-12-14}, + abstract = {We propose a new molecular simulation framework that combines the transferability, robustness, and chemical flexibility of an ab initio method with the accuracy and efficiency of a machine learning model. The key to achieve this mix is to use a standard density functional theory (DFT) simulation as a preprocessor for the atomic and molecular information, obtaining a good quality electronic density. General, symmetry preserving, atom-centered electronic descriptors are then built from this density to train a neural network to correct the baseline DFT energies and forces. These electronic descriptors encode much more information than local atomic environments, allowing a simple neural network to reach the accuracy required for the problem of study at a negligible additional cost. The balance between accuracy and efficiency is determined by the baseline simulation. This is shown in results where high level quantum chemical accuracy is obtained for simulations of liquid water at standard DFT cost or where high level DFT-accuracy is achieved in simulations with a low-level baseline DFT calculation at a significantly reduced cost.}, + keywords = {ACSF,BPNN,DFT,ML,ML-DFA,ML-DFT,ML-ESM,MLCF,molecules,prediction from density,prediction of energy correction}, + file = {/home/johannes/Nextcloud/Zotero/Dick_Fernandez-Serra_2019_Learning from the density to correct total energy and forces in first principle.pdf} +} + +@article{dickMachineLearningAccurate2020, + title = {Machine Learning Accurate Exchange and Correlation Functionals of the Electronic Density}, + author = {Dick, Sebastian and Fernandez-Serra, Marivi}, + date = {2020-07-14}, + journaltitle = {Nature Communications}, + shortjournal = {Nat Commun}, + volume = {11}, + number = {1}, + pages = {3509}, + publisher = {{Nature Publishing Group}}, + issn = {2041-1723}, + doi = {10.1038/s41467-020-17265-7}, + url = {https://www.nature.com/articles/s41467-020-17265-7}, + urldate = {2021-12-14}, + abstract = {Density functional theory (DFT) is the standard formalism to study the electronic structure of matter at the atomic scale. In Kohn–Sham DFT simulations, the balance between accuracy and computational cost depends on the choice of exchange and correlation functional, which only exists in approximate form. Here, we propose a framework to create density functionals using supervised machine learning, termed NeuralXC. These machine-learned functionals are designed to lift the accuracy of baseline functionals towards that provided by more accurate methods while maintaining their efficiency. We show that the functionals learn a meaningful representation of the physical information contained in the training data, making them transferable across systems. A NeuralXC functional optimized for water outperforms other methods characterizing bond breaking and excels when comparing against experimental results. This work demonstrates that NeuralXC is a first step towards the design of a universal, highly accurate functional valid for both molecules and solids.}, + issue = {1}, + langid = {english}, + keywords = {BPNN,DFT,library,ML,ML-DFA,MLCF,molecules,NeuralXC,prediction from density,prediction of Exc,prediction of vxc,with-code}, + annotation = {Bandiera\_abtest: a Cc\_license\_type: cc\_by Cg\_type: Nature Research Journals Primary\_atype: Research Subject\_term: Density functional theory;Electronic properties and materials;Molecular dynamics Subject\_term\_id: density-functional-theory;electronic-properties-and-materials;molecular-dynamics}, + file = {/home/johannes/Nextcloud/Zotero/Dick_Fernandez-Serra_2020_Machine learning accurate exchange and correlation functionals of the.pdf;/home/johannes/Zotero/storage/95GAG2CF/s41467-020-17265-7.html} +} + +@article{disanteDeepLearningFunctional2022, + title = {Deep {{Learning}} the {{Functional Renormalization Group}}}, + author = {Di Sante, Domenico and Medvidović, Matija and Toschi, Alessandro and Sangiovanni, Giorgio and Franchini, Cesare and Sengupta, Anirvan M. and Millis, Andrew J.}, + date = {2022-09-21}, + journaltitle = {Physical Review Letters}, + shortjournal = {Phys. Rev. Lett.}, + volume = {129}, + number = {13}, + pages = {136402}, + publisher = {{American Physical Society}}, + doi = {10.1103/PhysRevLett.129.136402}, + url = {https://link.aps.org/doi/10.1103/PhysRevLett.129.136402}, + urldate = {2022-10-19}, + abstract = {We perform a data-driven dimensionality reduction of the scale-dependent four-point vertex function characterizing the functional renormalization group (FRG) flow for the widely studied two-dimensional t−t′ Hubbard model on the square lattice. We demonstrate that a deep learning architecture based on a neural ordinary differential equation solver in a low-dimensional latent space efficiently learns the FRG dynamics that delineates the various magnetic and d-wave superconducting regimes of the Hubbard model. We further present a dynamic mode decomposition analysis that confirms that a small number of modes are indeed sufficient to capture the FRG dynamics. Our Letter demonstrates the possibility of using artificial intelligence to extract compact representations of the four-point vertex functions for correlated electrons, a goal of utmost importance for the success of cutting-edge quantum field theoretical methods for tackling the many-electron problem.}, + keywords = {Deep learning,FRG,Hubbard model,ML,rec-by-da-silva,renormalization group,superconductor}, + file = {/home/johannes/Nextcloud/Zotero/Di Sante et al_2022_Deep Learning the Functional Renormalization Group.pdf;/home/johannes/Zotero/storage/LKT2Z79L/Di Sante et al_2022_Deep Learning the Functional Renormalization Group-supp.pdf;/home/johannes/Zotero/storage/PGSNSHSM/PhysRevLett.129.html} +} + +@misc{dominaJacobiLegendrePotential2022, + title = {The {{Jacobi-Legendre}} Potential}, + author = {Domina, Michelangelo and Patil, Urvesh and Cobelli, Matteo and Sanvito, Stefano}, + date = {2022-08-22}, + number = {arXiv:2208.10292}, + eprint = {2208.10292}, + eprinttype = {arxiv}, + primaryclass = {cond-mat}, + publisher = {{arXiv}}, + doi = {10.48550/arXiv.2208.10292}, + url = {http://arxiv.org/abs/2208.10292}, + urldate = {2022-09-05}, + abstract = {Inspired by the cluster expansion method, we introduce a compact machine-learning potential constructed over Jacobi and Legendre polynomials. The constituent many-body contributions are separated, fully interpretable and adaptable to replicate the physical knowledge of the system, such as a repulsive behaviour at a small inter-atomic distance. Most importantly the potential requires a small number of features to achieve accuracy comparable to that of more numerically heavy and descriptor-rich alternatives. This is here tested for an organic molecule, a crystalline solid and an amorphous compound. Furthermore, we argue that the physical interpretability of the various terms is key to the selection and training of stable potentials.}, + archiveprefix = {arXiv}, + keywords = {ACE,descriptors,DFT,invariance,Jacobi-Legendre potential,JLP,linear regression,ML,ML-ESM,MLP,prediction of total energy,SNAP}, + file = {/home/johannes/Nextcloud/Zotero/Domina et al_2022_The Jacobi-Legendre potential.pdf;/home/johannes/Zotero/storage/DUUKR6TZ/2208.html} +} + +@article{dominaSpectralNeighborRepresentation2022, + title = {Spectral Neighbor Representation for Vector Fields: {{Machine}} Learning Potentials Including Spin}, + shorttitle = {Spectral Neighbor Representation for Vector Fields}, + author = {Domina, M. and Cobelli, M. and Sanvito, S.}, + date = {2022-06-30}, + journaltitle = {Physical Review B}, + shortjournal = {Phys. Rev. B}, + volume = {105}, + number = {21}, + pages = {214439}, + publisher = {{American Physical Society}}, + doi = {10.1103/PhysRevB.105.214439}, + url = {https://link.aps.org/doi/10.1103/PhysRevB.105.214439}, + urldate = {2022-09-05}, + abstract = {We introduce a translational and rotational invariant local representation for vector fields, which can be employed in the construction of machine learning energy models of solids and molecules. This allows us to describe, on the same footing, the energy fluctuations due to the atomic motion, the longitudinal and transverse excitations of the vector field, and their mutual interplay. The formalism can then be applied to physical systems where the total energy is determined by a vector density, as in the case of magnetism. Our representation is constructed over the power spectrum of the combined angular momentum describing the local atomic positions and the vector field, and it can be used in conjunction with different machine learning schemes and data taken from accurate ab initio electronic structure theories. We demonstrate the descriptive power of our representation for a range of classical spin Hamiltonian and machine learning algorithms. In particular, we construct energy models based on both linear Ridge regression, as in conventional spectral neighbor analysis potentials, and the Gaussian approximation. These are both built to represent a Heisenberg-type Hamiltonian including a longitudinal energy term and spin-lattice coupling.}, + keywords = {_tablet,descriptors,DFT,GPR,Heisenberg model,Jij,LRR,magnetism,ML,ML-DFT,ML-ESM,spin-dependent}, + file = {/home/johannes/Nextcloud/Zotero/Domina et al_2022_Spectral neighbor representation for vector fields.pdf;/home/johannes/Zotero/storage/F4KNYWPX/Domina et al_2022_Spectral neighbor representation for vector fields.pdf;/home/johannes/Zotero/storage/QX9ZENU5/PhysRevB.105.html} +} + +@unpublished{dominaSpectralneighbourRepresentationVector2022, + title = {A Spectral-Neighbour Representation for Vector Fields: Machine-Learning Potentials Including Spin}, + shorttitle = {A Spectral-Neighbour Representation for Vector Fields}, + author = {Domina, Michelangelo and Cobelli, Matteo and Sanvito, Stefano}, + date = {2022-02-23}, + eprint = {2202.13773}, + eprinttype = {arxiv}, + primaryclass = {cond-mat}, + url = {http://arxiv.org/abs/2202.13773}, + urldate = {2022-03-23}, + abstract = {We introduce a translational and rotational invariant local representation for vector fields, which can be employed in the construction of machine-learning energy models of solids and molecules. This allows us to describe, on the same footing, the energy fluctuations due to the atomic motion, the longitudinal and transverse excitations of the vector field, and their mutual interplay. The formalism can then be applied to physical systems where the total energy is determined by a vector density, as in the case of magnetism. Our representation is constructed over the power spectrum of the combined angular momentum describing the local atomic positions and the vector field, and can be used in conjunction with different machine-learning schemes and data taken from accurate ab initio electronic structure theories. We demonstrate the descriptive power of our representation for a range of classical spin Hamiltonian and machine-learning algorithms. In particular, we construct energy models based on both linear Ridge regression, as in conventional spectral neighbour analysis potentials, and gaussian approximation. These are both built to represent a Heisenberg-type Hamiltonian including a longitudinal energy term and spin-lattice coupling.}, + archiveprefix = {arXiv}, + keywords = {_tablet,descriptors,DFT,GPR,Heisenberg model,Jij,LRR,magnetism,ML,ML-DFT,ML-ESM,spin-dependent}, + file = {/home/johannes/Nextcloud/Zotero/Domina et al_2022_A spectral-neighbour representation for vector fields.pdf;/home/johannes/Zotero/storage/EB6UHPCQ/2202.html} +} + +@article{drautzAtomicClusterExpansion2019, + title = {Atomic Cluster Expansion for Accurate and Transferable Interatomic Potentials}, + author = {Drautz, Ralf}, + date = {2019-01-08}, + journaltitle = {Physical Review B}, + shortjournal = {Phys. Rev. B}, + volume = {99}, + number = {1}, + pages = {014104}, + publisher = {{American Physical Society}}, + doi = {10.1103/PhysRevB.99.014104}, + url = {https://link.aps.org/doi/10.1103/PhysRevB.99.014104}, + urldate = {2022-05-11}, + abstract = {The atomic cluster expansion is developed as a complete descriptor of the local atomic environment, including multicomponent materials, and its relation to a number of other descriptors and potentials is discussed. The effort for evaluating the atomic cluster expansion is shown to scale linearly with the number of neighbors, irrespective of the order of the expansion. Application to small Cu clusters demonstrates smooth convergence of the atomic cluster expansion to meV accuracy. By introducing nonlinear functions of the atomic cluster expansion an interatomic potential is obtained that is comparable in accuracy to state-of-the-art machine learning potentials. Because of the efficient convergence of the atomic cluster expansion relevant subspaces can be sampled uniformly and exhaustively. This is demonstrated by testing against a large database of density functional theory calculations for copper.}, + keywords = {_tablet,ACE,descriptors,original publication}, + file = {/home/johannes/Nextcloud/Zotero/Drautz_2019_Atomic cluster expansion for accurate and transferable interatomic potentials.pdf;/home/johannes/Zotero/storage/HNR9ZCLL/Drautz_2019_Atomic cluster expansion for accurate and transferable interatomic potentials.pdf;/home/johannes/Zotero/storage/NMAUF3NJ/PhysRevB.99.html} +} + +@article{drautzAtomicClusterExpansion2020, + title = {Atomic Cluster Expansion of Scalar, Vectorial, and Tensorial Properties Including Magnetism and Charge Transfer}, + author = {Drautz, Ralf}, + date = {2020}, + journaltitle = {Physical Review B}, + shortjournal = {Phys. Rev. B}, + volume = {102}, + number = {2}, + doi = {10.1103/PhysRevB.102.024104}, + keywords = {_tablet,ACE,descriptors,magnetism,ML,spin-dependent}, + file = {/home/johannes/Nextcloud/Zotero/Drautz_2020_Atomic cluster expansion of scalar, vectorial, and tensorial properties.pdf;/home/johannes/Zotero/storage/9W2WE4WX/PhysRevB.102.html} +} + +@misc{drautzAtomicClusterExpansion2022, + title = {Atomic Cluster Expansion and Wave Function Representations}, + author = {Drautz, Ralf and Ortner, Christoph}, + date = {2022-06-22}, + number = {arXiv:2206.11375}, + eprint = {2206.11375}, + eprinttype = {arxiv}, + primaryclass = {cond-mat}, + publisher = {{arXiv}}, + doi = {10.48550/arXiv.2206.11375}, + url = {http://arxiv.org/abs/2206.11375}, + urldate = {2022-06-28}, + abstract = {The atomic cluster expansion (ACE) has been highly successful for the parameterisation of symmetric (invariant or equivariant) properties of many-particle systems. Here, we generalize its derivation to anti-symmetric functions. We show how numerous well-known linear representations of wave functions naturally arise within this framework and we explore how recent successful nonlinear parameterisations can be further enhanced by employing ACE methodology. From this analysis we propose a wide design space of promising wave function representations.}, + archiveprefix = {arXiv}, + keywords = {_tablet,ACE,Backflow,Deep learning,ML-QM,prediction of wavefunction,representation of wavefunction,Slater-Jastrow}, + file = {/home/johannes/Nextcloud/Zotero/Drautz_Ortner_2022_Atomic cluster expansion and wave function representations.pdf;/home/johannes/Zotero/storage/6PTQT7NH/2206.html} +} + +@article{drautzSpinclusterExpansionParametrization2004, + title = {Spin-Cluster Expansion: {{Parametrization}} of the General Adiabatic Magnetic Energy Surface with {\emph{Ab Initio}} Accuracy}, + shorttitle = {Spin-Cluster Expansion}, + author = {Drautz, R.}, + date = {2004}, + journaltitle = {Physical Review B}, + shortjournal = {Phys. Rev. B}, + volume = {69}, + number = {10}, + doi = {10.1103/PhysRevB.69.104404}, + keywords = {ACE-related,descriptors}, + file = {/home/johannes/Zotero/storage/IICKDCFM/PhysRevB.69.html} +} + +@article{draxlNOMADFAIRConcept2018, + title = {{{NOMAD}}: {{The FAIR}} Concept for Big Data-Driven Materials Science}, + shorttitle = {{{NOMAD}}}, + author = {Draxl, Claudia and Scheffler, Matthias}, + date = {2018-09}, + journaltitle = {MRS Bulletin}, + volume = {43}, + number = {9}, + pages = {676--682}, + publisher = {{Cambridge University Press}}, + issn = {0883-7694, 1938-1425}, + doi = {10.1557/mrs.2018.208}, + url = {https://www.cambridge.org/core/journals/mrs-bulletin/article/nomad-the-fair-concept-for-big-datadriven-materials-science/1EEF321F62D41997CA16AD367B74C4B0}, + urldate = {2021-10-15}, + abstract = {, Data are a crucial raw material of this century. The amount of data that have been created in materials science thus far and that continues to be created every day is immense. Without a proper infrastructure that allows for collecting and sharing data, the envisioned success of big data-driven materials science will be hampered. For the field of computational materials science, the NOMAD (Novel Materials Discovery) Center of Excellence (CoE) has changed the scientific culture toward comprehensive and findable, accessible, interoperable, and reusable (FAIR) data, opening new avenues for mining materials science big data. Novel data-analytics concepts and tools turn data into knowledge and help in the prediction of new materials and in the identification of new properties of already known materials.}, + langid = {english}, + keywords = {artificial intelligence,data repositories,data sharing,machine learning,metadata}, + file = {/home/johannes/Nextcloud/Zotero/Draxl_Scheffler_2018_NOMAD.pdf;/home/johannes/Zotero/storage/3W2KJMWA/1EEF321F62D41997CA16AD367B74C4B0.html} +} + +@article{draxlNOMADLaboratoryData2019, + title = {The {{NOMAD}} Laboratory: From Data Sharing to Artificial Intelligence}, + shorttitle = {The {{NOMAD}} Laboratory}, + author = {Draxl, Claudia and Scheffler, Matthias}, + date = {2019-05}, + shortjournal = {J. Phys. Mater.}, + volume = {2}, + number = {3}, + pages = {036001}, + publisher = {{IOP Publishing}}, + issn = {2515-7639}, + doi = {10.1088/2515-7639/ab13bb}, + url = {https://doi.org/10.1088/2515-7639/ab13bb}, + urldate = {2021-10-15}, + abstract = {The Novel Materials Discovery (NOMAD) Laboratory is a user-driven platform for sharing and exploiting computational materials science data. It accounts for the various aspects of data being a crucial raw material and most relevant to accelerate materials research and engineering. NOMAD, with the NOMAD Repository, and its code-independent and normalized form, the NOMAD Archive, comprises the worldwide largest data collection of this field. Based on its findable accessible, interoperable, reusable data infrastructure, various services are offered, comprising advanced visualization, the NOMAD Encyclopedia, and artificial-intelligence tools. The latter are realized in the NOMAD Analytics Toolkit. Prerequisite for all this is the NOMAD metadata, a unique and thorough description of the data, that are produced by all important computer codes of the community. Uploaded data are tagged by a persistent identifier, and users can also request a digital object identifier to make data citable. Developments and advancements of parsers and metadata are organized jointly with users and code developers. In this work, we review the NOMAD concept and implementation, highlight its orthogonality to and synergistic interplay with other data collections, and provide an outlook regarding ongoing and future developments.}, + langid = {english}, + file = {/home/johannes/Nextcloud/Zotero/Draxl_Scheffler_2019_The NOMAD laboratory.pdf} +} + +@book{dresselhausGroupTheory2007, + title = {Group {{Theory}}}, + author = {Dresselhaus, Mildred S. and Dresselhaus, Gene and Jorio, Ado}, + date = {2007}, + publisher = {{Springer Berlin Heidelberg}}, + url = {https://doi.org/10.1007/978-3-540-32899-5}, + urldate = {2022-12-23}, + abstract = {Application to the Physics of Condensed Matter}, + isbn = {978-3-540-32897-1}, + langid = {english}, + keywords = {condensed matter,group theory,irreps,learning material,mathematics,rec-by-sabastian,textbook}, + file = {/home/johannes/Nextcloud/Zotero/Dresselhaus et al. - 2007 - Group Theory.pdf;/home/johannes/Zotero/storage/GGGVNLC4/978-3-540-32899-5.html} +} + +@article{drozdovConventionalSuperconductivity2032015, + title = {Conventional Superconductivity at 203 Kelvin at High Pressures in the Sulfur Hydride System}, + author = {Drozdov, A. P. and Eremets, M. I. and Troyan, I. A. and Ksenofontov, V. and Shylin, S. I.}, + date = {2015-09}, + journaltitle = {Nature}, + volume = {525}, + number = {7567}, + pages = {73--76}, + publisher = {{Nature Publishing Group}}, + issn = {1476-4687}, + doi = {10.1038/nature14964}, + url = {https://www.nature.com/articles/nature14964}, + urldate = {2021-10-21}, + abstract = {Conventional superconductivity is observed at 203 kelvin in the sulfur hydride system, well above the highest superconducting transition temperature obtained in the copper oxides, raising hopes that even higher transition temperatures will be discovered in other hydrogen-rich systems.}, + issue = {7567}, + langid = {english}, + keywords = {applications of DFT,DFT,master-thesis,superconductor}, + annotation = {Bandiera\_abtest: a Cg\_type: Nature Research Journals Primary\_atype: Research Subject\_term: Superconducting properties and materials Subject\_term\_id: superconducting-properties-and-materials}, + file = {/home/johannes/Nextcloud/Zotero/Drozdov et al_2015_Conventional superconductivity at 203 kelvin at high pressures in the sulfur.pdf;/home/johannes/Zotero/storage/CJIZLLVA/nature14964.html} +} + +@article{dunnBenchmarkingMaterialsProperty2020, + title = {Benchmarking Materials Property Prediction Methods: The {{Matbench}} Test Set and {{Automatminer}} Reference Algorithm}, + shorttitle = {Benchmarking Materials Property Prediction Methods}, + author = {Dunn, Alexander and Wang, Qi and Ganose, Alex and Dopp, Daniel and Jain, Anubhav}, + date = {2020-09-15}, + journaltitle = {npj Computational Materials}, + shortjournal = {npj Comput Mater}, + volume = {6}, + number = {1}, + pages = {1--10}, + publisher = {{Nature Publishing Group}}, + issn = {2057-3960}, + doi = {10.1038/s41524-020-00406-3}, + url = {https://www.nature.com/articles/s41524-020-00406-3}, + urldate = {2021-10-21}, + abstract = {We present a benchmark test suite and an automated machine learning procedure for evaluating supervised machine learning (ML) models for predicting properties of inorganic bulk materials. The test suite, Matbench, is a set of 13\,ML tasks that range in size from 312 to 132k samples and contain data from 10 density functional theory-derived and experimental sources. Tasks include predicting optical, thermal, electronic, thermodynamic, tensile, and elastic properties given a material’s composition and/or crystal structure. The reference algorithm, Automatminer, is a highly-extensible, fully automated ML pipeline for predicting materials properties from materials primitives (such as composition and crystal structure) without user intervention or hyperparameter tuning. We test Automatminer on the Matbench test suite and compare its predictive power with state-of-the-art crystal graph neural networks and a traditional descriptor-based Random Forest model. We find Automatminer achieves the best performance on 8 of 13 tasks in the benchmark. We also show our test suite is capable of exposing predictive advantages of each algorithm—namely, that crystal graph methods appear to outperform traditional machine learning methods given \textasciitilde 104 or greater data points. We encourage evaluating materials ML algorithms on the Matbench benchmark and comparing them against the latest version of Automatminer.}, + issue = {1}, + langid = {english}, + annotation = {Bandiera\_abtest: a Cc\_license\_type: cc\_by Cg\_type: Nature Research Journals Primary\_atype: Research Subject\_term: Atomistic models;Computational methods Subject\_term\_id: atomistic-models;computational-methods}, + file = {/home/johannes/Nextcloud/Zotero/Dunn et al_2020_Benchmarking materials property prediction methods.pdf;/home/johannes/Zotero/storage/N76WQWKL/s41524-020-00406-3.html} +} + +@unpublished{dussonAtomicClusterExpansion2021, + title = {Atomic {{Cluster Expansion}}: {{Completeness}}, {{Efficiency}} and {{Stability}}}, + shorttitle = {Atomic {{Cluster Expansion}}}, + author = {Dusson, Genevieve and Bachmayr, Markus and Csanyi, Gabor and Drautz, Ralf and Etter, Simon and van der Oord, Cas and Ortner, Christoph}, + options = {useprefix=true}, + date = {2021-05-12}, + eprint = {1911.03550}, + eprinttype = {arxiv}, + primaryclass = {cs, math}, + url = {http://arxiv.org/abs/1911.03550}, + urldate = {2022-05-11}, + abstract = {The Atomic Cluster Expansion (Drautz, Phys. Rev. B 99, 2019) provides a framework to systematically derive polynomial basis functions for approximating isometry and permutation invariant functions, particularly with an eye to modelling properties of atomistic systems. Our presentation extends the derivation by proposing a precomputation algorithm that yields immediate guarantees that a complete basis is obtained. We provide a fast recursive algorithm for efficient evaluation and illustrate its performance in numerical tests. Finally, we discuss generalisations and open challenges, particularly from a numerical stability perspective, around basis optimisation and parameter estimation, paving the way towards a comprehensive analysis of the convergence to a high-fidelity reference model.}, + archiveprefix = {arXiv}, + file = {/home/johannes/Nextcloud/Zotero/Dusson et al_2021_Atomic Cluster Expansion.pdf;/home/johannes/Zotero/storage/7WDUQE6K/1911.html} +} + +@unpublished{dymLowDimensionalInvariant2022, + title = {Low {{Dimensional Invariant Embeddings}} for {{Universal Geometric Learning}}}, + author = {Dym, Nadav and Gortler, Steven J.}, + date = {2022-05-05}, + number = {arXiv:2205.02956}, + eprint = {2205.02956}, + eprinttype = {arxiv}, + primaryclass = {cs, math}, + publisher = {{arXiv}}, + doi = {10.48550/arXiv.2205.02956}, + url = {http://arxiv.org/abs/2205.02956}, + urldate = {2022-05-18}, + abstract = {This paper studies separating invariants: mappings on \$d\$-dimensional semi-algebraic subsets of \$D\$ dimensional Euclidean domains which are invariant to semi-algebraic group actions and separate orbits. The motivation for this study comes from the usefulness of separating invariants in proving universality of equivariant neural network architectures. We observe that in several cases the cardinality of separating invariants proposed in the machine learning literature is much larger than the ambient dimension \$D\$. As a result, the theoretical universal constructions based on these separating invariants is unrealistically large. Our goal in this paper is to resolve this issue. We show that when a continuous family of semi-algebraic separating invariants is available, separation can be obtained by randomly selecting \$2d+1 \$ of these invariants. We apply this methodology to obtain an efficient scheme for computing separating invariants for several classical group actions which have been studied in the invariant learning literature. Examples include matrix multiplication actions on point clouds by permutations, rotations, and various other linear groups.}, + archiveprefix = {arXiv}, + keywords = {geometric deep learning,invariance}, + file = {/home/johannes/Nextcloud/Zotero/Dym_Gortler_2022_Low Dimensional Invariant Embeddings for Universal Geometric Learning.pdf;/home/johannes/Zotero/storage/I8BIG3VX/2205.html} +} + +@article{ebertCalculatingCondensedMatter2011, + title = {Calculating Condensed Matter Properties Using the {{KKR-Green}}'s Function Method—Recent Developments and Applications}, + author = {Ebert, H. and Ködderitzsch, D. and Minár, J.}, + date = {2011-08}, + journaltitle = {Reports on Progress in Physics}, + shortjournal = {Rep. Prog. Phys.}, + volume = {74}, + number = {9}, + pages = {096501}, + publisher = {{IOP Publishing}}, + issn = {0034-4885}, + doi = {10.1088/0034-4885/74/9/096501}, + url = {https://doi.org/10.1088/0034-4885/74/9/096501}, + urldate = {2021-10-04}, + abstract = {The modern version of the KKR (Korringa–Kohn–Rostoker) method represents the electronic structure of a system directly and efficiently in terms of its single-particle Green's function (GF). This is in contrast to its original version and many other traditional wave-function-based all-electron band structure methods dealing with periodically ordered solids. Direct access to the GF results in several appealing features. In addition, a wide applicability of the method is achieved by employing multiple scattering theory. The basic ideas behind the resulting KKR-GF method are outlined and the different techniques to deal with the underlying multiple scattering problem are reviewed. Furthermore, various applications of the KKR-GF method are reviewed in some detail to demonstrate the remarkable flexibility of the approach. Special attention is devoted to the numerous developments of the KKR-GF method, that have been contributed in recent years by a number of work groups, in particular in the following fields: embedding schemes for atoms, clusters and surfaces, magnetic response functions and anisotropy, electronic and spin-dependent transport, dynamical mean field theory, various kinds of spectroscopies, as well as first-principles determination of model parameters.}, + langid = {english}, + keywords = {_tablet,KKR,review}, + file = {/home/johannes/Nextcloud/Zotero/Ebert et al_2011_Calculating condensed matter properties using the KKR-Green's function.pdf} +} + +@unpublished{eckhoffHighDimensionalNeuralNetwork2021, + title = {High-{{Dimensional Neural Network Potentials}} for {{Magnetic Systems Using Spin-Dependent Atom-Centered Symmetry Functions}}}, + author = {Eckhoff, Marco and Behler, Jörg}, + date = {2021-04-29}, + eprint = {2104.14439}, + eprinttype = {arxiv}, + primaryclass = {physics}, + url = {http://arxiv.org/abs/2104.14439}, + urldate = {2021-05-18}, + abstract = {Machine learning potentials have emerged as a powerful tool to extend the time and length scales of first principles-quality simulations. Still, most machine learning potentials cannot distinguish different electronic spin orientations and thus are not applicable to materials in different magnetic states. Here, we propose spin-dependent atom-centered symmetry functions as a new type of descriptor taking the atomic spin degrees of freedom into account. When used as input for a high-dimensional neural network potential (HDNNP), accurate potential energy surfaces of multicomponent systems describing multiple magnetic states can be constructed. We demonstrate the performance of these magnetic HDNNPs for the case of manganese oxide, MnO. We show that the method predicts the magnetically distorted rhombohedral structure in excellent agreement with density functional theory and experiment. Its efficiency allows to determine the N\textbackslash '\{e\}el temperature considering structural fluctuations, entropic effects, and defects. The method is general and is expected to be useful also for other types of systems like oligonuclear transition metal complexes.}, + archiveprefix = {arXiv}, + keywords = {ACSF,ANN,descriptors,HDNNP,Heisenberg model,magnetism,ML,MLP,models,Physics - Computational Physics,spin-dependent,to read 2105}, + file = {/home/johannes/Nextcloud/Zotero/Eckhoff_Behler_2021_High-Dimensional Neural Network Potentials for Magnetic Systems Using.pdf;/home/johannes/Zotero/storage/KW8NBSDW/2104.html} +} + +@article{ellisAcceleratingFinitetemperatureKohnSham2021, + title = {Accelerating Finite-Temperature {{Kohn-Sham}} Density Functional Theory with Deep Neural Networks}, + author = {Ellis, J. A. and Fiedler, L. and Popoola, G. A. and Modine, N. A. and Stephens, J. A. and Thompson, A. P. and Cangi, A. and Rajamanickam, S.}, + date = {2021-07-08}, + journaltitle = {Physical Review B}, + shortjournal = {Phys. Rev. B}, + volume = {104}, + number = {3}, + pages = {035120}, + publisher = {{American Physical Society}}, + doi = {10.1103/PhysRevB.104.035120}, + url = {https://link.aps.org/doi/10.1103/PhysRevB.104.035120}, + urldate = {2021-12-05}, + abstract = {We present a numerical modeling workflow based on machine learning which reproduces the total energies produced by Kohn-Sham density functional theory (DFT) at finite electronic temperature to within chemical accuracy at negligible computational cost. Based on deep neural networks, our workflow yields the local density of states (LDOS) for a given atomic configuration. From the LDOS, spatially resolved, energy-resolved, and integrated quantities can be calculated, including the DFT total free energy, which serves as the Born-Oppenheimer potential energy surface for the atoms. We demonstrate the efficacy of this approach for both solid and liquid metals and compare results between independent and unified machine-learning models for solid and liquid aluminum. Our machine-learning density functional theory framework opens up the path towards multiscale materials modeling for matter under ambient and extreme conditions at a computational scale and cost that is unattainable with current algorithms.}, + keywords = {DFT,finite-temperature DFT,LAMMPS,library,MALA,ML,ML-DFT,ML-ESM,prediction of LDOS,quantum,Quantum ESPRESSO,SNAP,with-code}, + file = {/home/johannes/Nextcloud/Zotero/Ellis et al_2021_Accelerating finite-temperature Kohn-Sham density functional theory with deep.pdf;/home/johannes/Zotero/storage/AS2E35V9/PhysRevB.104.html} +} + +@article{evansGroupTheory2004, + title = {Group {{Theory}}}, + author = {Evans, Tim S. and Vvedensky, Dimitri D.}, + date = {2004-01-01}, + url = {https://www.academia.edu/2677753/Group_Theory}, + urldate = {2022-12-05}, + abstract = {Group Theory}, + keywords = {condensed matter,group theory,irreps,learning material,lecture notes,mathematics,rec-by-sabastian,textbook}, + file = {/home/johannes/Nextcloud/Zotero/Evans_2004_Group Theory.pdf;/home/johannes/Zotero/storage/K7KNP9VQ/Group_Theory.html} +} + +@article{faberCrystalStructureRepresentations2015, + title = {Crystal Structure Representations for Machine Learning Models of Formation Energies}, + author = {Faber, Felix and Lindmaa, Alexander and von Lilienfeld, O. Anatole and Armiento, Rickard}, + date = {2015}, + journaltitle = {International Journal of Quantum Chemistry}, + volume = {115}, + number = {16}, + pages = {1094--1101}, + issn = {1097-461X}, + doi = {10.1002/qua.24917}, + url = {https://onlinelibrary.wiley.com/doi/abs/10.1002/qua.24917}, + urldate = {2021-07-10}, + abstract = {We introduce and evaluate a set of feature vector representations of crystal structures for machine learning (ML) models of formation energies of solids. ML models of atomization energies of organic molecules have been successful using a Coulomb matrix representation of the molecule. We consider three ways to generalize such representations to periodic systems: (i) a matrix where each element is related to the Ewald sum of the electrostatic interaction between two different atoms in the unit cell repeated over the lattice; (ii) an extended Coulomb-like matrix that takes into account a number of neighboring unit cells; and (iii) an ansatz that mimics the periodicity and the basic features of the elements in the Ewald sum matrix using a sine function of the crystal coordinates of the atoms. The representations are compared for a Laplacian kernel with Manhattan norm, trained to reproduce formation energies using a dataset of 3938 crystal structures obtained from the Materials Project. For training sets consisting of 3000 crystals, the generalization error in predicting formation energies of new structures corresponds to (i) 0.49, (ii) 0.64, and (iii) for the respective representations. © 2015 Wiley Periodicals, Inc.}, + langid = {english}, + keywords = {descriptors,Ewald sum matrix,ML,original publication,Sine matrix}, + annotation = {\_eprint: https://onlinelibrary.wiley.com/doi/pdf/10.1002/qua.24917}, + file = {/home/johannes/Nextcloud/Zotero/Faber et al_2015_Crystal structure representations for machine learning models of formation.pdf} +} + +@article{fabianiInvestigatingUltrafastQuantum2019, + title = {Investigating Ultrafast Quantum Magnetism with Machine Learning}, + author = {Fabiani, Giammarco and Mentink, Johan}, + date = {2019-07-05}, + journaltitle = {SciPost Physics}, + volume = {7}, + number = {1}, + pages = {004}, + issn = {2542-4653}, + doi = {10.21468/SciPostPhys.7.1.004}, + url = {https://scipost.org/SciPostPhys.7.1.004}, + urldate = {2022-03-29}, + abstract = {SciPost Journals Publication Detail SciPost Phys. 7, 004 (2019) Investigating ultrafast quantum magnetism with machine learning}, + langid = {english}, + keywords = {Heisenberg model,magnetism,ML,NN,RBM,rec-by-bluegel,spin-dependent}, + file = {/home/johannes/Nextcloud/Zotero/Fabiani_Mentink_2019_Investigating ultrafast quantum magnetism with machine learning.pdf;/home/johannes/Zotero/storage/U8VR8E9L/SciPostPhys.7.1.html} +} + +@article{fabianiSupermagnonicPropagationTwoDimensional2021, + title = {Supermagnonic {{Propagation}} in {{Two-Dimensional Antiferromagnets}}}, + author = {Fabiani, G. and Bouman, M. D. and Mentink, J. H.}, + date = {2021-08-25}, + journaltitle = {Physical Review Letters}, + shortjournal = {Phys. Rev. Lett.}, + volume = {127}, + number = {9}, + pages = {097202}, + publisher = {{American Physical Society}}, + doi = {10.1103/PhysRevLett.127.097202}, + url = {https://link.aps.org/doi/10.1103/PhysRevLett.127.097202}, + urldate = {2022-03-29}, + abstract = {We investigate the propagation of magnons after ultrashort perturbations of the exchange interaction in the prototype two-dimensional Heisenberg antiferromagnet. Using the recently proposed neural quantum states, we predict highly anisotropic spreading in space constrained by the symmetry of the perturbation. Interestingly, the propagation speed at the shortest length scale and timescale is up to 40\% higher than the highest magnon velocity. We argue that the enhancement stems from extraordinary strong magnon-magnon interactions, suggesting new avenues for manipulating information transfer on ultrashort length scales and timescales.}, + keywords = {rec-by-bluegel}, + file = {/home/johannes/Nextcloud/Zotero/Fabiani et al_2021_Supermagnonic Propagation in Two-Dimensional Antiferromagnets.pdf;/home/johannes/Zotero/storage/VBV4L6RZ/Fabiani et al. - 2021 - Supermagnonic Propagation in Two-Dimensional Antif.pdf;/home/johannes/Zotero/storage/ZLEKQ276/PhysRevLett.127.html} +} + +@article{fabrizioElectronDensityLearning2019, + title = {Electron Density Learning of Non-Covalent Systems}, + author = {Fabrizio, Alberto and Grisafi, Andrea and Meyer, Benjamin and Ceriotti, Michele and Corminboeuf, Clemence}, + date = {2019}, + journaltitle = {Chemical Science}, + volume = {10}, + number = {41}, + pages = {9424--9432}, + publisher = {{Royal Society of Chemistry}}, + doi = {10.1039/C9SC02696G}, + url = {https://pubs.rsc.org/en/content/articlelanding/2019/sc/c9sc02696g}, + urldate = {2021-10-16}, + langid = {english}, + keywords = {ML,ML-DFT,ML-ESM,prediction of electron density}, + file = {/home/johannes/Nextcloud/Zotero/Fabrizio et al_2019_Electron density learning of non-covalent systems.pdf} +} + +@article{farajiHighAccuracyTransferability2017, + title = {High Accuracy and Transferability of a Neural Network Potential through Charge Equilibration for Calcium Fluoride}, + author = {Faraji, Somayeh and Ghasemi, S. Alireza and Rostami, Samare and Rasoulkhani, Robabe and Schaefer, Bastian and Goedecker, Stefan and Amsler, Maximilian}, + date = {2017-03-16}, + journaltitle = {Physical Review B}, + shortjournal = {Phys. Rev. B}, + volume = {95}, + number = {10}, + pages = {104105}, + publisher = {{American Physical Society}}, + doi = {10.1103/PhysRevB.95.104105}, + url = {https://link.aps.org/doi/10.1103/PhysRevB.95.104105}, + urldate = {2021-07-22}, + abstract = {We investigate the accuracy and transferability of a recently developed high-dimensional neural network (NN) method for calcium fluoride, fitted to a database of ab initio density functional theory (DFT) calculations based on the Perdew-Burke-Ernzerhof (PBE) exchange correlation functional. We call the method charge equilibration via neural network technique (CENT). Although the fitting database contains only clusters (i.e., nonperiodic structures), the NN scheme accurately describes a variety of bulk properties. In contrast to other available empirical methods the CENT potential has a much simpler functional form, nevertheless it correctly reproduces the PBE energetics of various crystalline phases both at ambient and high pressure. Surface energies and structures as well as dynamical properties derived from phonon calculations are also in good agreement with PBE results. Overall, the difference between the values obtained by the CENT potential and the PBE reference values is less than or equal to the difference between the values of local density approximation (LDA) and Born-Mayer-Huggins (BMH) with those calculated by the PBE exchange correlation functional.}, + keywords = {CENT,HDNNP,rec-by-bluegel}, + file = {/home/johannes/Nextcloud/Zotero/Faraji et al_2017_High accuracy and transferability of a neural network potential through charge.pdf;/home/johannes/Zotero/storage/GU7MU2BP/PhysRevB.95.html} +} + +@article{fernandez-delgadoWeNeedHundreds2014, + title = {Do We {{Need Hundreds}} of {{Classifiers}} to {{Solve Real World Classification Problems}}?}, + author = {Fernández-Delgado, Manuel and Cernadas, Eva and Barro, Senén and Amorim, Dinani}, + date = {2014}, + journaltitle = {Journal of Machine Learning Research}, + volume = {15}, + number = {90}, + pages = {3133--3181}, + issn = {1533-7928}, + url = {http://jmlr.org/papers/v15/delgado14a.html}, + urldate = {2022-01-02}, + abstract = {We evaluate 179 classifiers arising from 17 families (discriminant analysis, Bayesian, neural networks, support vector machines, decision trees, rule-based classifiers, boosting, bagging, stacking, random forests and other ensembles, generalized linear models, nearest-neighbors, partial least squares and principal component regression, logistic and multinomial regression, multiple adaptive regression splines and other methods), implemented in Weka, R (with and without the caret package), C and Matlab, including all the relevant classifiers available today. We use 121 data sets, which represent the whole UCI data base (excluding the large- scale problems) and other own real problems, in order to achieve significant conclusions about the classifier behavior, not dependent on the data set collection. The classifiers most likely to be the bests are the random forest (RF) versions, the best of which (implemented in R and accessed via caret) achieves 94.1\% of the maximum accuracy overcoming 90\% in the 84.3\% of the data sets. However, the difference is not statistically significant with the second best, the SVM with Gaussian kernel implemented in C using LibSVM, which achieves 92.3\% of the maximum accuracy. A few models are clearly better than the remaining ones: random forest, SVM with Gaussian and polynomial kernels, extreme learning machine with Gaussian kernel, C5.0 and avNNet (a committee of multi-layer perceptrons implemented in R with the caret package). The random forest is clearly the best family of classifiers (3 out of 5 bests classifiers are RF), followed by SVM (4 classifiers in the top-10), neural networks and boosting ensembles (5 and 3 members in the top-20, respectively).}, + file = {/home/johannes/Nextcloud/Zotero/Fernández-Delgado et al_2014_Do we Need Hundreds of Classifiers to Solve Real World Classification Problems.pdf} +} + +@article{feyPowerMessagePassing2022, + title = {On the Power of Message Passing for Learning on Graph-Structured Data}, + author = {Fey, Matthias}, + date = {2022}, + doi = {10.17877/DE290R-22906}, + url = {https://eldorado.tu-dortmund.de/handle/2003/41059}, + urldate = {2022-10-16}, + abstract = {This thesis proposes novel approaches for machine learning on irregularly structured input data such as graphs, point clouds and manifolds. Specifically, we are breaking up with the regularity restriction of conventional deep learning techniques, and propose solutions in designing, implementing and scaling up deep end-to-end representation learning on graph-structured data, known as Graph Neural Networks (GNNs). GNNs capture local graph structure and feature information by following a neural message passing scheme, in which node representations are recursively updated in a trainable and purely local fashion. In this thesis, we demonstrate the generality of message passing through a unified framework suitable for a wide range of operators and learning tasks. Specifically, we analyze the limitations and inherent weaknesses of GNNs and propose efficient solutions to overcome them, both theoretically and in practice, e.g., by conditioning messages via continuous B-spline kernels, by utilizing hierarchical message passing, or by leveraging positional encodings. In addition, we ensure that our proposed methods scale naturally to large input domains. In particular, we propose novel methods to fully eliminate the exponentially increasing dependency of nodes over layers inherent to message passing GNNs. Lastly, we introduce PyTorch Geometric, a deep learning library for implementing and working with graph-based neural network building blocks, built upon PyTorch.}, + langid = {english}, + keywords = {GDL,GNN,library,MPNN,PhD,PyG,pytorch,thesis}, + annotation = {Accepted: 2022-08-31T09:01:01Z}, + file = {/home/johannes/Nextcloud/Zotero/Fey_2022_On the power of message passing for learning on graph-structured data.pdf;/home/johannes/Zotero/storage/CEBUDJT7/41059.html} +} + +@unpublished{fiedlerDeepDiveMachine2021, + title = {A {{Deep Dive}} into {{Machine Learning Density Functional Theory}} for {{Materials Science}} and {{Chemistry}}}, + author = {Fiedler, Lenz and Shah, Karan and Bussmann, Michael and Cangi, Attila}, + date = {2021-10-03}, + eprint = {2110.00997}, + eprinttype = {arxiv}, + primaryclass = {cond-mat}, + url = {http://arxiv.org/abs/2110.00997}, + urldate = {2021-11-17}, + abstract = {With the growth of computational resources, the scope of electronic structure simulations has increased greatly. Artificial intelligence and robust data analysis hold the promise to accelerate large-scale simulations and their analysis to hitherto unattainable scales. Machine learning is a rapidly growing field for the processing of such complex datasets. It has recently gained traction in the domain of electronic structure simulations, where density functional theory takes the prominent role of the most widely used electronic structure method. Thus, DFT calculations represent one of the largest loads on academic high-performance computing systems across the world. Accelerating these with machine learning can reduce the resources required and enables simulations of larger systems. Hence, the combination of density functional theory and machine learning has the potential to rapidly advance electronic structure applications such as in-silico materials discovery and the search for new chemical reaction pathways. We provide the theoretical background of both density functional theory and machine learning on a generally accessible level. This serves as the basis of our comprehensive review including research articles up to December 2020 in chemistry and materials science that employ machine-learning techniques. In our analysis, we categorize the body of research into main threads and extract impactful results. We conclude our review with an outlook on exciting research directions in terms of a citation analysis.}, + archiveprefix = {arXiv}, + keywords = {_tablet,Condensed Matter - Materials Science,DFT,ML,ML-DFT,ML-ESM,review}, + file = {/home/johannes/Nextcloud/Zotero/Fiedler et al_2021_A Deep Dive into Machine Learning Density Functional Theory for Materials.pdf;/home/johannes/Zotero/storage/2XW6IGEA/2110.html} +} + +@article{flores-livasPredictionHotSuperconductivity2019, + title = {A {{Prediction}} for “{{Hot}}†{{Superconductivity}}}, + author = {Flores-Livas, José A. and Arita, Ryotaro}, + date = {2019-08-26}, + journaltitle = {Physics}, + volume = {12}, + pages = {96}, + publisher = {{American Physical Society}}, + doi = {10.1103/PhysRevLett.123.097001}, + url = {https://physics.aps.org/articles/v12/96}, + urldate = {2021-10-21}, + abstract = {A proposed hydrogen-rich solid would superconduct above the boiling point of water—though the material would need to be subjected to a colossal pressure.}, + langid = {english}, + keywords = {applications of DFT,DFT,master-thesis,superconductor}, + file = {/home/johannes/Nextcloud/Zotero/Flores-Livas_Arita_2019_A Prediction for “Hot†Superconductivity.pdf;/home/johannes/Zotero/storage/IEKIVRSH/96.html} +} + +@report{foulkesTopologyEntanglementStrong2020, + title = {Topology, {{Entanglement}}, and {{Strong Correlations}}}, + author = {Foulkes, W. M. C. and Drautz, Ralf}, + date = {2020}, + series = {Lecture {{Notes}} of the {{Autumn School}} on {{Correlated Electrons}}}, + number = {FZJ-2020-03083}, + institution = {{Forschungszentrum Jülich GmbH Zentralbibliothek, Verlag}}, + url = {https://juser.fz-juelich.de/record/884084/}, + urldate = {2022-06-28}, + abstract = {Topology and entanglement are key concepts in many-body physics. Understanding the as-sociated emergent phenomena beyond toy models – in the world of real strongly-correlatedmaterials – requires the mastery of a wealth of different methods. These encompass analytical tools such as group theory, first principles techniques based on density-functional theory, materials-specific model-building schemes, as well as advanced modern numerical approaches for solving realistic many-body models. This year’s school provides an overview of the state-of-the art of these methods, their successes and their limitations. After introducing the basics, lectures will present the core concepts of topology and entanglement in many-body systems. To make contact to real materials, strategies for building materials specific models and techniques for their solution will be introduced. Among the latter, the school will cover quantum Monte Carlo methods, construction and optimization of correlated wave-functions, recursion and renormalization group techniques, as well as dynamical mean-field theory. More advanced lectures will give a pedagogical overview ontopological materials and their physics: topological metals, semimetals, and superconductors. Towards the end of the school entanglement in quantum dynamics and perspectives in quantum computation will be discussed. The goal of the school is to introduce advanced graduate students and up to these modern approaches for the realistic modeling of strongly correlated materials. A school of this size and scope requires backing from many sources. This is even more truethis year. As everywhere, the Corona pandemics provided scores of new challenges. Plans had to be changed and real facilities had to be replaced with virtual ones. We are very grateful forall the practical and financial support we have received. The Institute for Advanced Simulationat the Forschungszentrum J ülich and the Jülich Supercomputer Centre provided the major part of the funding and were vital for the organization and reorganization of the school as well as for the production of this book. The Institute for Complex Adaptive Matter (ICAM) supplied additional funds and ideas for successful online formats. The nature of a school makes it desirable to have the lecture notes available when the lecturesare given. This way students get the chance to work through the lectures thoroughly while their memory is still fresh. We are therefore extremely grateful to the lecturers that, despite tight deadlines, provided their manuscripts in time for the production of this book. We are confident that the lecture notes collected here will not only serve the participants of the school but will also be useful for other students entering the exciting field of strongly correlated materials. We are grateful to Mrs. H. Lexis of the Verlag des Forschungszentrum Jülich and to Mrs. D. Mans of the Grafische Betriebe for providing their expert support in producing the present volume on a tight schedule. We heartily thank our students and postdocs who helped with proofreading the manuscripts, often on quite short notice: Elaheh Adibi, Julian Mußhoff, NedaSamani, and Xue-Jing Zhang. Finally, our special thanks go to Dipl.-Ing. R. Hölzle for his invaluable advice on the innu-merable questions concerning the organization of such an endeavor, and to Mrs. L. Snyders forexpertly handling all practical issues. Pavarini, Eva; Koch, Erik}, + editorb = {Pavarini, Eva and Koch, Erik}, + editorbtype = {redactor}, + eventtitle = {Autumn {{School}} on {{Correlated Electrons}}}, + isbn = {9783958064669}, + langid = {english}, + keywords = {_tablet,ACE,cusps,FermiNet,ML-ESM,ML-QMBP,prediction of wavefunction,QMC,Slater-Jastrow,VMC}, + file = {/home/johannes/Nextcloud/Zotero/Foulkes_Drautz_2020_Topology, Entanglement, and Strong Correlations.pdf;/home/johannes/Zotero/storage/WLIE37SZ/884084.html} +} + +@unpublished{frankDetectInteractionsThat2021, + title = {Detect the {{Interactions}} That {{Matter}} in {{Matter}}: {{Geometric Attention}} for {{Many-Body Systems}}}, + shorttitle = {Detect the {{Interactions}} That {{Matter}} in {{Matter}}}, + author = {Frank, Thorben and Chmiela, Stefan}, + date = {2021-09-06}, + number = {arXiv:2106.02549}, + eprint = {2106.02549}, + eprinttype = {arxiv}, + primaryclass = {physics}, + publisher = {{arXiv}}, + doi = {10.48550/arXiv.2106.02549}, + url = {http://arxiv.org/abs/2106.02549}, + urldate = {2022-05-18}, + abstract = {Attention mechanisms are developing into a viable alternative to convolutional layers as elementary building block of NNs. Their main advantage is that they are not restricted to capture local dependencies in the input, but can draw arbitrary connections. This unprecedented capability coincides with the long-standing problem of modeling global atomic interactions in molecular force fields and other many-body problems. In its original formulation, however, attention is not applicable to the continuous domains in which the atoms live. For this purpose we propose a variant to describe geometric relations for arbitrary atomic configurations in Euclidean space that also respects all relevant physical symmetries. We furthermore demonstrate, how the successive application of our learned attention matrices effectively translates the molecular geometry into a set of individual atomic contributions on-the-fly.}, + archiveprefix = {arXiv}, + keywords = {attention,invariance,ML}, + file = {/home/johannes/Nextcloud/Zotero/Frank_Chmiela_2021_Detect the Interactions that Matter in Matter.pdf;/home/johannes/Zotero/storage/7QC4UBJN/2106.html} +} + +@article{frauxChemiscopeInteractiveStructureproperty2020, + title = {Chemiscope: Interactive Structure-Property Explorer for Materials and Molecules}, + shorttitle = {Chemiscope}, + author = {Fraux, Guillaume and Cersonsky, Rose K. and Ceriotti, Michele}, + date = {2020-07-01}, + journaltitle = {Journal of Open Source Software}, + volume = {5}, + number = {51}, + pages = {2117}, + issn = {2475-9066}, + doi = {10.21105/joss.02117}, + url = {https://joss.theoj.org/papers/10.21105/joss.02117}, + urldate = {2022-10-03}, + abstract = {Fraux et al., (2020). Chemiscope: interactive structure-property explorer for materials and molecules. Journal of Open Source Software, 5(51), 2117, https://doi.org/10.21105/joss.02117}, + langid = {english}, + keywords = {data exploration,Database,ML,molecules,sketchmap,solids,unsupervised learning,visualization}, + file = {/home/johannes/Nextcloud/Zotero/Fraux et al_2020_Chemiscope.pdf;/home/johannes/Zotero/storage/TCQI9XE2/joss.html} +} + +@article{freyMachineLearningEnabledDesign2020, + title = {Machine {{Learning-Enabled Design}} of {{Point Defects}} in {{2D Materials}} for {{Quantum}} and {{Neuromorphic Information Processing}}}, + author = {Frey, Nathan C. and Akinwande, Deji and Jariwala, Deep and Shenoy, Vivek B.}, + date = {2020-10-27}, + journaltitle = {ACS Nano}, + shortjournal = {ACS Nano}, + volume = {14}, + number = {10}, + pages = {13406--13417}, + publisher = {{American Chemical Society}}, + issn = {1936-0851}, + doi = {10.1021/acsnano.0c05267}, + url = {https://doi.org/10.1021/acsnano.0c05267}, + urldate = {2021-05-20}, + abstract = {Engineered point defects in two-dimensional (2D) materials offer an attractive platform for solid-state devices that exploit tailored optoelectronic, quantum emission, and resistive properties. Naturally occurring defects are also unavoidably important contributors to material properties and performance. The immense variety and complexity of possible defects make it challenging to experimentally control, probe, or understand atomic-scale defect-property relationships. Here, we develop an approach based on deep transfer learning, machine learning, and first-principles calculations to rapidly predict key properties of point defects in 2D materials. We use physics-informed featurization to generate a minimal description of defect structures and present a general picture of defects across materials systems. We identify over one hundred promising, unexplored dopant defect structures in layered metal chalcogenides, hexagonal nitrides, and metal halides. These defects are prime candidates for quantum emission, resistive switching, and neuromorphic computing.}, + keywords = {2D material,classification,deep transfer learning,defects,DFT,materials discovery,materials for neuromorphic computing,materials for quantum computing,NN,random forest,regression,transfer learning}, + file = {/home/johannes/Nextcloud/Zotero/Frey et al_2020_Machine Learning-Enabled Design of Point Defects in 2D Materials for Quantum.pdf} +} + +@article{freyNeuralScalingDeep2022, + title = {Neural {{Scaling}} of {{Deep Chemical Models}}}, + author = {Frey, Nathan and Soklaski, Ryan and Axelrod, Simon and Samsi, Siddharth and Gomez-Bombarelli, Rafael and Coley, Connor and Gadepally, Vijay}, + date = {2022-05-16}, + doi = {10.26434/chemrxiv-2022-3s512}, + url = {https://chemrxiv.org/engage/chemrxiv/article-details/627bddd544bdd532395fb4b5}, + urldate = {2022-10-03}, + abstract = {Massive scale, both in terms of data availability and computation, enables significant breakthroughs in key application areas of deep learning such as natural language processing (NLP) and computer vision. There is emerging evidence that scale may be a key ingredient in scientific deep learning, but the importance of physical priors in scientific domains makes the strategies and benefits of scaling uncertain. Here, we investigate neural scaling behavior in large chemical models by varying model and dataset sizes over many orders of magnitude, studying models with over one billion parameters, pre-trained on datasets of up to ten million datapoints. We consider large language models for generative chemistry and graph neural networks for machine-learned interatomic potentials. To enable large-scale scientific deep learning studies under resource constraints, we develop the Training Performance Estimation (TPE) framework to reduce the costs of scalable hyperparameter optimization by up to 90\%. Using this framework, we discover empirical neural scaling relations for deep chemical models and investigate the interplay between physical priors and scale. Potential applications of large, pre-trained models for "prompt engineering" and unsupervised representation learning of molecules are shown.}, + langid = {english}, + keywords = {Allegro,ChemGPT,hyperparameters optimization,large language models,large models,MLP,models comparison,original publication,PAiNN,scaling,SchNet,SpookyNet,TODO,Training Performance Estimation,with-code}, + file = {/home/johannes/Nextcloud/Zotero/Frey et al_2022_Neural Scaling of Deep Chemical Models.pdf;/home/johannes/Zotero/storage/RLZJQP57/627bddd544bdd532395fb4b5.html} +} + +@article{freysoldtFirstprinciplesCalculationsPoint2014, + title = {First-Principles Calculations for Point Defects in Solids}, + author = {Freysoldt, Christoph}, + date = {2014}, + journaltitle = {Reviews of Modern Physics}, + shortjournal = {Rev. Mod. Phys.}, + volume = {86}, + number = {1}, + pages = {253--305}, + doi = {10.1103/RevModPhys.86.253}, + keywords = {basics,defects,DFT,impurity embedding,MPI Eisenforschung,review}, + file = {/home/johannes/Zotero/storage/26HPBYJN/RevModPhys.86.html} +} + +@article{friederichMachinelearnedPotentialsNextgeneration2021, + title = {Machine-Learned Potentials for next-Generation Matter Simulations}, + author = {Friederich, Pascal and Häse, Florian and Proppe, Jonny and Aspuru-Guzik, Alán}, + date = {2021-06}, + journaltitle = {Nature Materials}, + shortjournal = {Nat. Mater.}, + volume = {20}, + number = {6}, + pages = {750--761}, + publisher = {{Nature Publishing Group}}, + issn = {1476-4660}, + doi = {10.1038/s41563-020-0777-6}, + url = {https://www.nature.com/articles/s41563-020-0777-6}, + urldate = {2022-05-13}, + abstract = {The choice of simulation methods in computational materials science is driven by a fundamental trade-off: bridging large time- and length-scales with highly accurate simulations at an affordable computational cost. Venturing the investigation of complex phenomena on large scales requires fast yet accurate computational methods. We review the emerging field of machine-learned potentials, which promises to reach the accuracy of quantum mechanical computations at a substantially reduced computational cost. This Review will summarize the basic principles of the underlying machine learning methods, the data acquisition process and active learning procedures. We highlight multiple recent applications of machine-learned potentials in various fields, ranging from organic chemistry and biomolecules to inorganic crystal structure predictions and surface science. We furthermore discuss the developments required to promote a broader use of ML potentials, and the possibility of using them to help solve open questions in materials science and facilitate fully computational materials design.}, + issue = {6}, + langid = {english}, + keywords = {descriptors,MD,ML,MLP,review,review-of-MLP}, + file = {/home/johannes/Nextcloud/Zotero/Friederich et al_2021_Machine-learned potentials for next-generation matter simulations.pdf;/home/johannes/Zotero/storage/KBK5DLLS/s41563-020-0777-6.html} +} + +@article{frolovQuantumComputingReproducibility2021, + title = {Quantum Computing’s Reproducibility Crisis: {{Majorana}} Fermions}, + shorttitle = {Quantum Computing’s Reproducibility Crisis}, + author = {Frolov, Sergey}, + date = {2021-04}, + journaltitle = {Nature}, + volume = {592}, + number = {7854}, + pages = {350--352}, + publisher = {{Nature Publishing Group}}, + doi = {10.1038/d41586-021-00954-8}, + url = {https://www.nature.com/articles/d41586-021-00954-8}, + urldate = {2022-10-21}, + abstract = {The controversy over Majorana particles is eroding confidence in the field. More accountability and openness are needed — from authors, reviewers and journal editors.}, + issue = {7854}, + langid = {english}, + keywords = {Majorana,Peer review,quantum computing,rec-by-ghosh,reproducibility crisis,skeptics}, + annotation = {Bandiera\_abtest: a Cg\_type: Comment Subject\_term: Quantum physics, Publishing, Peer review}, + file = {/home/johannes/Nextcloud/Zotero/Frolov_2021_Quantum computing’s reproducibility crisis.pdf;/home/johannes/Zotero/storage/CLEGVGB5/d41586-021-00954-8.html} +} + +@misc{fuchsSETransformers3D2020, + title = {{{SE}}(3)-{{Transformers}}: {{3D Roto-Translation Equivariant Attention Networks}}}, + shorttitle = {{{SE}}(3)-{{Transformers}}}, + author = {Fuchs, Fabian B. and Worrall, Daniel E. and Fischer, Volker and Welling, Max}, + date = {2020-11-24}, + number = {arXiv:2006.10503}, + eprint = {2006.10503}, + eprinttype = {arxiv}, + primaryclass = {cs, stat}, + publisher = {{arXiv}}, + doi = {10.48550/arXiv.2006.10503}, + url = {http://arxiv.org/abs/2006.10503}, + urldate = {2022-10-03}, + abstract = {We introduce the SE(3)-Transformer, a variant of the self-attention module for 3D point clouds and graphs, which is equivariant under continuous 3D roto-translations. Equivariance is important to ensure stable and predictable performance in the presence of nuisance transformations of the data input. A positive corollary of equivariance is increased weight-tying within the model. The SE(3)-Transformer leverages the benefits of self-attention to operate on large point clouds and graphs with varying number of points, while guaranteeing SE(3)-equivariance for robustness. We evaluate our model on a toy N-body particle simulation dataset, showcasing the robustness of the predictions under rotations of the input. We further achieve competitive performance on two real-world datasets, ScanObjectNN and QM9. In all cases, our model outperforms a strong, non-equivariant attention baseline and an equivariant model without attention.}, + archiveprefix = {arXiv}, + keywords = {equivariant,GCN,GNN,library,ML,QM9,SchNet,SE(3),self-attention,transformer,with-code}, + file = {/home/johannes/Nextcloud/Zotero/Fuchs et al_2020_SE(3)-Transformers.pdf;/home/johannes/Zotero/storage/UMVV286P/2006.html} +} + +@article{fungPhysicallyInformedMachine2022, + title = {Physically {{Informed Machine Learning Prediction}} of {{Electronic Density}} of {{States}}}, + author = {Fung, Victor and Ganesh, P. and Sumpter, Bobby G.}, + date = {2022-06-14}, + journaltitle = {Chemistry of Materials}, + shortjournal = {Chem. Mater.}, + volume = {34}, + number = {11}, + pages = {4848--4855}, + publisher = {{American Chemical Society}}, + issn = {0897-4756}, + doi = {10.1021/acs.chemmater.1c04252}, + url = {https://doi.org/10.1021/acs.chemmater.1c04252}, + urldate = {2022-07-10}, + abstract = {The electronic structure of a material, such as its density of states (DOS), provides key insights into its physical and functional properties and serves as a valuable source of high-quality features for many materials screening and discovery workflows. However, the computational cost of calculating the DOS, most commonly with density functional theory (DFT), becomes prohibitive for meeting high-fidelity or high-throughput requirements, necessitating a cheaper but sufficiently accurate surrogate. To fulfill this demand, we develop a general machine learning method based on graph neural networks for predicting the DOS purely from atomic positions, six orders of magnitude faster than DFT. This approach can effectively use large materials databases and be applied generally across the entire periodic table to materials classes of arbitrary compositional and structural diversity. We furthermore devise a highly adaptable scheme for physically informed learning which encourages the DOS prediction to favor physically reasonable solutions defined by any set of desired constraints. This functionality provides a means for ensuring that the predicted DOS is reliable enough to be used as an input to downstream materials screening workflows to predict more complex functional properties, which rely on accurate physical features.}, + file = {/home/johannes/Nextcloud/Zotero/Fung et al_2022_Physically Informed Machine Learning Prediction of Electronic Density of States.pdf;/home/johannes/Zotero/storage/MFQH6849/acs.chemmater.html} +} + +@online{galkinGraphML20222021, + title = {Graph {{ML}} in 2022: {{Where Are We Now}}?}, + shorttitle = {Graph {{ML}} in 2022}, + author = {Galkin, Michael}, + date = {2021-12-30T06:00:52}, + url = {https://towardsdatascience.com/graph-ml-in-2022-where-are-we-now-f7f8242599e0}, + urldate = {2022-01-02}, + abstract = {Hot trends and major advancements}, + langid = {english}, + organization = {{Medium}}, + file = {/home/johannes/Zotero/storage/8ESSCXA2/graph-ml-in-2022-where-are-we-now-f7f8242599e0.html} +} + +@book{gammaDesignPatternsElements1995, + title = {Design Patterns: Elements of Reusable Object-Oriented Software}, + shorttitle = {Design Patterns}, + editor = {Gamma, Erich}, + date = {1995}, + series = {Addison-{{Wesley}} Professional Computing Series}, + publisher = {{Addison-Wesley}}, + location = {{Reading, Mass}}, + isbn = {978-0-201-63361-0}, + pagetotal = {395}, + keywords = {OO,Reusability,software engineering,Software patterns} +} + +@misc{gardnerSyntheticDataEnable2022, + title = {Synthetic Data Enable Experiments in Atomistic Machine Learning}, + author = {Gardner, John L. A. and Beaulieu, Zoé Faure and Deringer, Volker L.}, + date = {2022-11-29}, + number = {arXiv:2211.16443}, + eprint = {2211.16443}, + eprinttype = {arxiv}, + primaryclass = {physics}, + publisher = {{arXiv}}, + doi = {10.48550/arXiv.2211.16443}, + url = {http://arxiv.org/abs/2211.16443}, + urldate = {2022-12-29}, + abstract = {Machine-learning models are increasingly used to predict properties of atoms in chemical systems. There have been major advances in developing descriptors and regression frameworks for this task, typically starting from (relatively) small sets of quantum-mechanical reference data. Larger datasets of this kind are becoming available, but remain expensive to generate. Here we demonstrate the use of a large dataset that we have "synthetically" labelled with per-atom energies from an existing ML potential model. The cheapness of this process, compared to the quantum-mechanical ground truth, allows us to generate millions of datapoints, in turn enabling rapid experimentation with atomistic ML models from the small- to the large-data regime. This approach allows us here to compare regression frameworks in depth, and to explore visualisation based on learned representations. We also show that learning synthetic data labels can be a useful pre-training task for subsequent fine-tuning on small datasets. In the future, we expect that our open-sourced dataset, and similar ones, will be useful in rapidly exploring deep-learning models in the limit of abundant chemical data.}, + archiveprefix = {arXiv}, + keywords = {data augmentation,GAP,GPR,MD,ML,MLP,NN,prediction of potential energy,small data,SOAP,synthetic data}, + file = {/home/johannes/Nextcloud/Zotero/Gardner et al_2022_Synthetic data enable experiments in atomistic machine learning.pdf;/home/johannes/Zotero/storage/N3NP679J/2211.html} +} + +@article{garridotorresLowScalingAlgorithmNudged2019, + title = {Low-{{Scaling Algorithm}} for {{Nudged Elastic Band Calculations Using}} a {{Surrogate Machine Learning Model}}}, + author = {Garrido Torres, José A. and Jennings, Paul C. and Hansen, Martin H. and Boes, Jacob R. and Bligaard, Thomas}, + date = {2019-04-15}, + journaltitle = {Physical Review Letters}, + shortjournal = {Phys. Rev. Lett.}, + volume = {122}, + number = {15}, + pages = {156001}, + publisher = {{American Physical Society}}, + doi = {10.1103/PhysRevLett.122.156001}, + url = {https://link.aps.org/doi/10.1103/PhysRevLett.122.156001}, + urldate = {2021-08-21}, + abstract = {We present the incorporation of a surrogate Gaussian process regression (GPR) atomistic model to greatly accelerate the rate of convergence of classical nudged elastic band (NEB) calculations. In our surrogate model approach, the cost of converging the elastic band no longer scales with the number of moving images on the path. This provides a far more efficient and robust transition state search. In contrast to a conventional NEB calculation, the algorithm presented here eliminates any need for manipulating the number of images to obtain a converged result. This is achieved by inventing a new convergence criteria that exploits the probabilistic nature of the GPR to use uncertainty estimates of all images in combination with the force in the saddle point in the target model potential. Our method is an order of magnitude faster in terms of function evaluations than the conventional NEB method with no accuracy loss for the converged energy barrier values.}, + keywords = {DFT,GPR,ML,models,NEB,rec-by-ruess,surrogate model}, + file = {/home/johannes/Nextcloud/Zotero/Garrido Torres et al_2019_Low-Scaling Algorithm for Nudged Elastic Band Calculations Using a Surrogate.pdf;/home/johannes/Zotero/storage/TFSWCWBH/Garrido Torres et al. - 2019 - Low-Scaling Algorithm for Nudged Elastic Band Calc.pdf;/home/johannes/Zotero/storage/DWT7X58R/PhysRevLett.122.html} +} + +@misc{gasteigerDirectionalMessagePassing2022, + title = {Directional {{Message Passing}} for {{Molecular Graphs}}}, + author = {Gasteiger, Johannes and Groß, Janek and Günnemann, Stephan}, + date = {2022-04-05}, + number = {arXiv:2003.03123}, + eprint = {2003.03123}, + eprinttype = {arxiv}, + primaryclass = {physics, stat}, + publisher = {{arXiv}}, + doi = {10.48550/arXiv.2003.03123}, + url = {http://arxiv.org/abs/2003.03123}, + urldate = {2022-10-03}, + abstract = {Graph neural networks have recently achieved great successes in predicting quantum mechanical properties of molecules. These models represent a molecule as a graph using only the distance between atoms (nodes). They do not, however, consider the spatial direction from one atom to another, despite directional information playing a central role in empirical potentials for molecules, e.g. in angular potentials. To alleviate this limitation we propose directional message passing, in which we embed the messages passed between atoms instead of the atoms themselves. Each message is associated with a direction in coordinate space. These directional message embeddings are rotationally equivariant since the associated directions rotate with the molecule. We propose a message passing scheme analogous to belief propagation, which uses the directional information by transforming messages based on the angle between them. Additionally, we use spherical Bessel functions and spherical harmonics to construct theoretically well-founded, orthogonal representations that achieve better performance than the currently prevalent Gaussian radial basis representations while using fewer than 1/4 of the parameters. We leverage these innovations to construct the directional message passing neural network (DimeNet). DimeNet outperforms previous GNNs on average by 76\% on MD17 and by 31\% on QM9. Our implementation is available online.}, + archiveprefix = {arXiv}, + keywords = {DimeNet,GNN,MD,ML,MLP,molecules,MPNN,original publication,with-code}, + file = {/home/johannes/Nextcloud/Zotero/Gasteiger et al_2022_Directional Message Passing for Molecular Graphs.pdf;/home/johannes/Zotero/storage/G7KWBFCS/2003.html} +} + +@misc{gasteigerDirectionalMessagePassing2022a, + title = {Directional {{Message Passing}} on {{Molecular Graphs}} via {{Synthetic Coordinates}}}, + author = {Gasteiger, Johannes and Yeshwanth, Chandan and Günnemann, Stephan}, + date = {2022-04-05}, + number = {arXiv:2111.04718}, + eprint = {2111.04718}, + eprinttype = {arxiv}, + primaryclass = {physics, q-bio}, + publisher = {{arXiv}}, + doi = {10.48550/arXiv.2111.04718}, + url = {http://arxiv.org/abs/2111.04718}, + urldate = {2022-10-03}, + abstract = {Graph neural networks that leverage coordinates via directional message passing have recently set the state of the art on multiple molecular property prediction tasks. However, they rely on atom position information that is often unavailable, and obtaining it is usually prohibitively expensive or even impossible. In this paper we propose synthetic coordinates that enable the use of advanced GNNs without requiring the true molecular configuration. We propose two distances as synthetic coordinates: Distance bounds that specify the rough range of molecular configurations, and graph-based distances using a symmetric variant of personalized PageRank. To leverage both distance and angular information we propose a method of transforming normal graph neural networks into directional MPNNs. We show that with this transformation we can reduce the error of a normal graph neural network by 55\% on the ZINC benchmark. We furthermore set the state of the art on ZINC and coordinate-free QM9 by incorporating synthetic coordinates in the SMP and DimeNet++ models. Our implementation is available online.}, + archiveprefix = {arXiv}, + keywords = {DimeNet,DimeNet++,GNN,MD,MLP,molecules,MPNN,open-review}, + file = {/home/johannes/Nextcloud/Zotero/Gasteiger et al_2022_Directional Message Passing on Molecular Graphs via Synthetic Coordinates.pdf;/home/johannes/Zotero/storage/FEPN4JW4/2111.html} +} + +@misc{gasteigerFastUncertaintyAwareDirectional2022, + title = {Fast and {{Uncertainty-Aware Directional Message Passing}} for {{Non-Equilibrium Molecules}}}, + author = {Gasteiger, Johannes and Giri, Shankari and Margraf, Johannes T. and Günnemann, Stephan}, + date = {2022-04-05}, + number = {arXiv:2011.14115}, + eprint = {2011.14115}, + eprinttype = {arxiv}, + primaryclass = {physics}, + publisher = {{arXiv}}, + doi = {10.48550/arXiv.2011.14115}, + url = {http://arxiv.org/abs/2011.14115}, + urldate = {2022-10-03}, + abstract = {Many important tasks in chemistry revolve around molecules during reactions. This requires predictions far from the equilibrium, while most recent work in machine learning for molecules has been focused on equilibrium or near-equilibrium states. In this paper we aim to extend this scope in three ways. First, we propose the DimeNet++ model, which is 8x faster and 10\% more accurate than the original DimeNet on the QM9 benchmark of equilibrium molecules. Second, we validate DimeNet++ on highly reactive molecules by developing the challenging COLL dataset, which contains distorted configurations of small molecules during collisions. Finally, we investigate ensembling and mean-variance estimation for uncertainty quantification with the goal of accelerating the exploration of the vast space of non-equilibrium structures. Our DimeNet++ implementation as well as the COLL dataset are available online.}, + archiveprefix = {arXiv}, + keywords = {DimeNet,DimeNet++,GNN,MD,ML,MLP,molecules,MPNN,original publication,with-code}, + file = {/home/johannes/Nextcloud/Zotero/Gasteiger et al_2022_Fast and Uncertainty-Aware Directional Message Passing for Non-Equilibrium.pdf;/home/johannes/Zotero/storage/BVEXST79/2011.html} +} + +@misc{gasteigerGemNetUniversalDirectional2022, + title = {{{GemNet}}: {{Universal Directional Graph Neural Networks}} for {{Molecules}}}, + shorttitle = {{{GemNet}}}, + author = {Gasteiger, Johannes and Becker, Florian and Günnemann, Stephan}, + date = {2022-04-05}, + number = {arXiv:2106.08903}, + eprint = {2106.08903}, + eprinttype = {arxiv}, + primaryclass = {physics, stat}, + publisher = {{arXiv}}, + doi = {10.48550/arXiv.2106.08903}, + url = {http://arxiv.org/abs/2106.08903}, + urldate = {2022-10-03}, + abstract = {Effectively predicting molecular interactions has the potential to accelerate molecular dynamics by multiple orders of magnitude and thus revolutionize chemical simulations. Graph neural networks (GNNs) have recently shown great successes for this task, overtaking classical methods based on fixed molecular kernels. However, they still appear very limited from a theoretical perspective, since regular GNNs cannot distinguish certain types of graphs. In this work we close this gap between theory and practice. We show that GNNs with directed edge embeddings and two-hop message passing are indeed universal approximators for predictions that are invariant to translation, and equivariant to permutation and rotation. We then leverage these insights and multiple structural improvements to propose the geometric message passing neural network (GemNet). We demonstrate the benefits of the proposed changes in multiple ablation studies. GemNet outperforms previous models on the COLL, MD17, and OC20 datasets by 34\%, 41\%, and 20\%, respectively, and performs especially well on the most challenging molecules. Our implementation is available online.}, + archiveprefix = {arXiv}, + keywords = {DimeNet,DimeNet++,GemNet,MD,ML,MLP,molecules,MPNN,with-code}, + file = {/home/johannes/Nextcloud/Zotero/Gasteiger et al_2022_GemNet.pdf;/home/johannes/Zotero/storage/FE5R77B9/2106.html} +} + +@article{gebauerInverseDesign3d2022, + title = {Inverse Design of 3d Molecular Structures with Conditional Generative Neural Networks}, + author = {Gebauer, Niklas W. A. and Gastegger, Michael and Hessmann, Stefaan S. P. and Müller, Klaus-Robert and Schütt, Kristof T.}, + date = {2022-02-21}, + journaltitle = {Nature Communications}, + shortjournal = {Nat Commun}, + volume = {13}, + number = {1}, + pages = {973}, + publisher = {{Nature Publishing Group}}, + issn = {2041-1723}, + doi = {10.1038/s41467-022-28526-y}, + url = {https://www.nature.com/articles/s41467-022-28526-y}, + urldate = {2022-06-25}, + abstract = {The rational design of molecules with desired properties is a long-standing challenge in chemistry. Generative neural networks have emerged as a powerful approach to sample novel molecules from a learned distribution. Here, we propose a conditional generative neural network for 3d molecular structures with specified chemical and structural properties. This approach is agnostic to chemical bonding and enables targeted sampling of novel molecules from conditional distributions, even in domains where reference calculations are sparse. We demonstrate the utility of our method for inverse design by generating molecules with specified motifs or composition, discovering particularly stable molecules, and jointly targeting multiple electronic properties beyond the training regime.}, + issue = {1}, + langid = {english}, + keywords = {G-SchNet,generative models,inverse design,materials discovery,ML,prediction from properties,prediction of structure,SchNet}, + file = {/home/johannes/Nextcloud/Zotero/Gebauer et al_2022_Inverse design of 3d molecular structures with conditional generative neural.pdf;/home/johannes/Zotero/storage/XHLV2UHD/s41467-022-28526-y.html} +} + +@article{gedeonMachineLearningDerivative2021, + title = {Machine Learning the Derivative Discontinuity of Density-Functional Theory}, + author = {Gedeon, Johannes and Schmidt, Jonathan and Hodgson, Matthew J. P. and Wetherell, Jack and Benavides-Riveros, Carlos L. and Marques, Miguel A. L.}, + date = {2021-12}, + journaltitle = {Machine Learning: Science and Technology}, + shortjournal = {Mach. Learn.: Sci. Technol.}, + volume = {3}, + number = {1}, + pages = {015011}, + publisher = {{IOP Publishing}}, + issn = {2632-2153}, + doi = {10.1088/2632-2153/ac3149}, + url = {https://doi.org/10.1088/2632-2153/ac3149}, + urldate = {2022-01-10}, + abstract = {Machine learning is a powerful tool to design accurate, highly non-local, exchange-correlation functionals for density functional theory. So far, most of those machine learned functionals are trained for systems with an integer number of particles. As such, they are unable to reproduce some crucial and fundamental aspects, such as the explicit dependency of the functionals on the particle number or the infamous derivative discontinuity at integer particle numbers. Here we propose a solution to these problems by training a neural network as the universal functional of density-functional theory that (a) depends explicitly on the number of particles with a piece-wise linearity between the integer numbers and (b) reproduces the derivative discontinuity of the exchange-correlation energy. This is achieved by using an ensemble formalism, a training set containing fractional densities, and an explicitly discontinuous formulation.}, + langid = {english}, + file = {/home/johannes/Nextcloud/Zotero/Gedeon et al_2021_Machine learning the derivative discontinuity of density-functional theory.pdf} +} + +@misc{geigerE3nnEuclideanNeural2022, + title = {E3nn: {{Euclidean Neural Networks}}}, + shorttitle = {E3nn}, + author = {Geiger, Mario and Smidt, Tess}, + date = {2022-07-18}, + number = {arXiv:2207.09453}, + eprint = {2207.09453}, + eprinttype = {arxiv}, + primaryclass = {cs}, + publisher = {{arXiv}}, + doi = {10.48550/arXiv.2207.09453}, + url = {http://arxiv.org/abs/2207.09453}, + urldate = {2022-08-21}, + abstract = {We present e3nn, a generalized framework for creating E(3) equivariant trainable functions, also known as Euclidean neural networks. e3nn naturally operates on geometry and geometric tensors that describe systems in 3D and transform predictably under a change of coordinate system. The core of e3nn are equivariant operations such as the TensorProduct class or the spherical harmonics functions that can be composed to create more complex modules such as convolutions and attention mechanisms. These core operations of e3nn can be used to efficiently articulate Tensor Field Networks, 3D Steerable CNNs, Clebsch-Gordan Networks, SE(3) Transformers and other E(3) equivariant networks.}, + archiveprefix = {arXiv}, + keywords = {_tablet,e3nn,EGNN,ENN,equivariant,library,ML-ESM,prediction of electron density,with-code}, + file = {/home/johannes/Nextcloud/Zotero/Geiger_Smidt_2022_e3nn.pdf;/home/johannes/Zotero/storage/SJW8392C/2207.html} +} + +@article{genschComprehensiveDiscoveryPlatform2021, + title = {A {{Comprehensive Discovery Platform}} for {{Organophosphorus Ligands}} for {{Catalysis}}}, + author = {Gensch, Tobias and dos Passos Gomes, Gabriel and Friederich, Pascal and Peters, Ellyn and Gaudin, Theophile and Pollice, Robert and Jorner, Kjell and Nigam, AkshatKumar and Lindner D'Addario, Michael and Sigman, Matthew S. and Aspuru-Guzik, Alan}, + options = {useprefix=true}, + date = {2021-04-27}, + publisher = {{ChemRxiv}}, + doi = {10.26434/chemrxiv.12996665.v1}, + url = {/articles/preprint/A_Comprehensive_Discovery_Platform_for_Organophosphorus_Ligands_for_Catalysis/12996665/1}, + urldate = {2021-05-15}, + abstract = {The design of molecular catalysts typically involves reconciling multiple conflicting property requirements, largely relying on human intuition and local structural searches. However, the vast number of potential catalysts requires pruning of the candidate space by efficient property prediction with quantitative structure-property relationships. Data-driven workflows embedded in a library of potential catalysts can be used to build predictive models for catalyst performance and serve as a blueprint for novel catalyst designs. Herein we introduce kraken, a discovery platform covering monodentate organophosphorus(III) ligands providing comprehensive physicochemical descriptors based on representative conformer ensembles. Using quantum-mechanical methods, we calculated descriptors for 1,558 ligands, including commercially available examples, and trained machine learning models to predict properties of over 300,000 new ligands. We demonstrate the application of kraken to systematically explore the property space of organophosphorus ligands and how existing datasets in catalysis can be used to accelerate ligand selection during reaction optimization.}, + langid = {english}, + keywords = {chemistry,descriptors,kraken,materials database,ML,models,organic chemistry,visualization}, + file = {/home/johannes/Nextcloud/Zotero/Gensch et al_2021_A Comprehensive Discovery Platform for Organophosphorus Ligands for Catalysis.pdf;/home/johannes/Zotero/storage/ZSYETR3S/12996665.html} +} + +@misc{gerardGoldstandardSolutionsSchr2022, + title = {Gold-Standard Solutions to the {{Schr}}\textbackslash "odinger Equation Using Deep Learning: {{How}} Much Physics Do We Need?}, + shorttitle = {Gold-Standard Solutions to the {{Schr}}\textbackslash "odinger Equation Using Deep Learning}, + author = {Gerard, Leon and Scherbela, Michael and Marquetand, Philipp and Grohs, Philipp}, + date = {2022-05-31}, + number = {arXiv:2205.09438}, + eprint = {2205.09438}, + eprinttype = {arxiv}, + primaryclass = {physics}, + publisher = {{arXiv}}, + doi = {10.48550/arXiv.2205.09438}, + url = {http://arxiv.org/abs/2205.09438}, + urldate = {2022-08-16}, + abstract = {Finding accurate solutions to the Schr\textbackslash "odinger equation is the key unsolved challenge of computational chemistry. Given its importance for the development of new chemical compounds, decades of research have been dedicated to this problem, but due to the large dimensionality even the best available methods do not yet reach the desired accuracy. Recently the combination of deep learning with Monte Carlo methods has emerged as a promising way to obtain highly accurate energies and moderate scaling of computational cost. In this paper we significantly contribute towards this goal by introducing a novel deep-learning architecture that achieves 40-70\% lower energy error at 8x lower computational cost compared to previous approaches. Using our method we establish a new benchmark by calculating the most accurate variational ground state energies ever published for a number of different atoms and molecules. We systematically break down and measure our improvements, focusing in particular on the effect of increasing physical prior knowledge. We surprisingly find that increasing the prior knowledge given to the architecture can actually decrease accuracy.}, + archiveprefix = {arXiv}, + keywords = {DNN,FermiNet,ML-ESM,ML-QMBP,molecules,PauliNet,prediction of wavefunction,QMC,VMC}, + file = {/home/johannes/Nextcloud/Zotero/Gerard et al_2022_Gold-standard solutions to the Schr-odinger equation using deep learning.pdf;/home/johannes/Zotero/storage/DWVRHXZW/2205.html} +} + +@book{geronHandsonMachineLearning2019, + title = {Hands-on Machine Learning with {{Scikit-Learn}}, {{Keras}}, and {{TensorFlow}}: Concepts, Tools, and Techniques to Build Intelligent Systems}, + shorttitle = {Hands-on Machine Learning with {{Scikit-Learn}}, {{Keras}}, and {{TensorFlow}}}, + author = {Géron, Aurélien}, + date = {2019}, + edition = {Second edition}, + publisher = {{O'Reilly Media, Inc}}, + location = {{Sebastopol, CA}}, + abstract = {Through a series of recent breakthroughs, deep learning has boosted the entire field of machine learning. Now, even programmers who know close to nothing about this technology can use simple, efficient tools to implement programs capable of learning from data. The updated edition of this best-selling book uses concrete examples, minimal theory, and two production-ready Python frameworks-Scikit-Learn and TensorFlow 2-to help you gain an intuitive understanding of the concepts and tools for building intelligent systems. Practitioners will learn a range of techniques that they can quickly put to use on the job. Part 1 employs Scikit-Learn to introduce fundamental machine learning tasks, such as simple linear regression. Part 2, which has been significantly updated, employs Keras and TensorFlow 2 to guide the reader through more advanced machine learning methods using deep neural networks. With exercises in each chapter to help you apply what you've learned, all you need is programming experience to get started. NEW FOR THE SECOND EDITION:Updated all code to TensorFlow 2 ; Introduced the high-level Keras API ; New and expanded coverage including TensorFlow's Data API, Eager Execution, Estimators API, deploying on Google Cloud ML, handling time series, embeddings and more With Early Release ebooks, you get books in their earliest form-the author's raw and unedited content as he or she writes-so you can take advantage of these technologies long before the official release of these titles. You'll also receive updates when significant changes are made, new chapters are available, and the final ebook bundle is released}, + isbn = {978-1-4920-3264-9}, + pagetotal = {819}, + keywords = {general,ML,notebooks,practice,Python,TensorFlow,with-code}, + annotation = {OCLC: on1124925244}, + file = {/home/johannes/Books/machine_learning/general_practice/Géron_2019_Hands-on machine learning with Scikit-Learn, Keras, and TensorFlow.pdf} +} + +@article{ghiringhelliAItoolkitDevelopShare2021, + title = {An {{AI-toolkit}} to Develop and Share Research into New Materials}, + author = {Ghiringhelli, Luca M.}, + date = {2021-09-09}, + journaltitle = {Nature Reviews Physics}, + shortjournal = {Nat Rev Phys}, + pages = {1--1}, + publisher = {{Nature Publishing Group}}, + issn = {2522-5820}, + doi = {10.1038/s42254-021-00373-8}, + url = {https://www.nature.com/articles/s42254-021-00373-8}, + urldate = {2021-09-11}, + abstract = {Luca Ghiringhelli introduces an AI toolkit that can be used with materials databases to discover new materials, or new properties of known materials.}, + langid = {english}, + annotation = {Bandiera\_abtest: a Cg\_type: Nature Research Journals Primary\_atype: Research Highlights Subject\_term: Computational methods;Scientific data Subject\_term\_id: computational-methods;scientific-data}, + file = {/home/johannes/Nextcloud/Zotero/Ghiringhelli_2021_An AI-toolkit to develop and share research into new materials.pdf;/home/johannes/Zotero/storage/LTJNU3SG/s42254-021-00373-8.html} +} + +@article{ghiringhelliBigDataMaterials2015, + title = {Big {{Data}} of {{Materials Science}}: {{Critical Role}} of the {{Descriptor}}}, + shorttitle = {Big {{Data}} of {{Materials Science}}}, + author = {Ghiringhelli, Luca M. and Vybiral, Jan and Levchenko, Sergey V. and Draxl, Claudia and Scheffler, Matthias}, + date = {2015-03-10}, + journaltitle = {Physical Review Letters}, + shortjournal = {Phys. Rev. Lett.}, + volume = {114}, + number = {10}, + pages = {105503}, + publisher = {{American Physical Society}}, + doi = {10.1103/PhysRevLett.114.105503}, + url = {https://link.aps.org/doi/10.1103/PhysRevLett.114.105503}, + urldate = {2021-05-15}, + abstract = {Statistical learning of materials properties or functions so far starts with a largely silent, nonchallenged step: the choice of the set of descriptive parameters (termed descriptor). However, when the scientific connection between the descriptor and the actuating mechanisms is unclear, the causality of the learned descriptor-property relation is uncertain. Thus, a trustful prediction of new promising materials, identification of anomalies, and scientific advancement are doubtful. We analyze this issue and define requirements for a suitable descriptor. For a classic example, the energy difference of zinc blende or wurtzite and rocksalt semiconductors, we demonstrate how a meaningful descriptor can be found systematically.}, + keywords = {descriptors,ML}, + file = {/home/johannes/Nextcloud/Zotero/Ghiringhelli et al_2015_Big Data of Materials Science.pdf;/home/johannes/Zotero/storage/VMWGSVVB/PhysRevLett.114.html} +} + +@article{ghiringhelliEfficientDataExchange2017, + title = {Towards Efficient Data Exchange and Sharing for Big-Data Driven Materials Science: Metadata and Data Formats}, + shorttitle = {Towards Efficient Data Exchange and Sharing for Big-Data Driven Materials Science}, + author = {Ghiringhelli, Luca M. and Carbogno, Christian and Levchenko, Sergey and Mohamed, Fawzi and Huhs, Georg and Lüders, Martin and Oliveira, Micael and Scheffler, Matthias}, + date = {2017-11-06}, + journaltitle = {npj Computational Materials}, + shortjournal = {npj Comput Mater}, + volume = {3}, + number = {1}, + pages = {1--9}, + publisher = {{Nature Publishing Group}}, + issn = {2057-3960}, + doi = {10.1038/s41524-017-0048-5}, + url = {https://www.nature.com/articles/s41524-017-0048-5}, + urldate = {2021-10-15}, + abstract = {With big-data driven materials research, the new paradigm of materials science, sharing and wide accessibility of data are becoming crucial aspects. Obviously, a prerequisite for data exchange and big-data analytics is standardization, which means using consistent and unique conventions for, e.g., units, zero base lines, and file formats. There are two main strategies to achieve this goal. One accepts the heterogeneous nature of the community, which comprises scientists from physics, chemistry, bio-physics, and materials science, by complying with the diverse ecosystem of computer codes and thus develops “converters†for the input and output files of all important codes. These converters then translate the data of each code into a standardized, code-independent format. The other strategy is to provide standardized open libraries that code developers can adopt for shaping their inputs, outputs, and restart files, directly into the same code-independent format. In this perspective paper, we present both strategies and argue that they can and should be regarded as complementary, if not even synergetic. The represented appropriate format and conventions were agreed upon by two teams, the Electronic Structure Library (ESL) of the European Center for Atomic and Molecular Computations (CECAM) and the NOvel MAterials Discovery (NOMAD) Laboratory, a European Centre of Excellence (CoE). A key element of this work is the definition of hierarchical metadata describing state-of-the-art electronic-structure calculations.}, + issue = {1}, + langid = {english}, + keywords = {materials database,materials informatics,materials metadata,NOMAD}, + annotation = {Bandiera\_abtest: a Cc\_license\_type: cc\_by Cg\_type: Nature Research Journals Primary\_atype: Reviews Subject\_term: Condensed-matter physics;Theory and computation Subject\_term\_id: condensed-matter-physics;theory-and-computation}, + file = {/home/johannes/Nextcloud/Zotero/Ghiringhelli et al_2017_Towards efficient data exchange and sharing for big-data driven materials.pdf;/home/johannes/Zotero/storage/G3CTM9SN/s41524-017-0048-5.html} +} + +@misc{ghoshClassicalQuantumMachine2022, + title = {Classical and Quantum Machine Learning Applications in Spintronics}, + author = {Ghosh, Kumar and Ghosh, Sumit}, + date = {2022-07-26}, + number = {arXiv:2207.12837}, + eprint = {2207.12837}, + eprinttype = {arxiv}, + primaryclass = {cond-mat, physics:physics, physics:quant-ph}, + publisher = {{arXiv}}, + doi = {10.48550/arXiv.2207.12837}, + url = {http://arxiv.org/abs/2207.12837}, + urldate = {2022-10-28}, + abstract = {In this article we demonstrate the applications of classical and quantum machine learning in quantum transport and spintronics. With the help of a two terminal device with magnetic impurity we show how machine learning algorithms can predict the highly non-linear nature of conductance as well as the non-equilibrium spin response function for any random magnetic configuration. We finally describe the applicability of quantum machine learning which has the capability to handle a significantly large configuration space. Our approach is also applicable for molecular systems. These outcomes are crucial in predicting the behaviour of large scale systems where a quantum mechanical calculation is computationally challenging and therefore would play a crucial role in designing nano devices.}, + archiveprefix = {arXiv}, + keywords = {ML,PGI-1/IAS-1,QML,QSVM,quantum computing,quantum transport,random forest,rec-by-ghosh,spin dynamics,Spintronics,SVM,tight binding,transport properties}, + file = {/home/johannes/Nextcloud/Zotero/Ghosh_Ghosh_2022_Classical and quantum machine learning applications in spintronics.pdf;/home/johannes/Zotero/storage/FEUD8XZQ/2207.html} +} + +@misc{gilmerNeuralMessagePassing2017, + title = {Neural {{Message Passing}} for {{Quantum Chemistry}}}, + author = {Gilmer, Justin and Schoenholz, Samuel S. and Riley, Patrick F. and Vinyals, Oriol and Dahl, George E.}, + date = {2017-06-12}, + number = {arXiv:1704.01212}, + eprint = {1704.01212}, + eprinttype = {arxiv}, + primaryclass = {cs}, + publisher = {{arXiv}}, + doi = {10.48550/arXiv.1704.01212}, + url = {http://arxiv.org/abs/1704.01212}, + urldate = {2022-10-03}, + abstract = {Supervised learning on molecules has incredible potential to be useful in chemistry, drug discovery, and materials science. Luckily, several promising and closely related neural network models invariant to molecular symmetries have already been described in the literature. These models learn a message passing algorithm and aggregation procedure to compute a function of their entire input graph. At this point, the next step is to find a particularly effective variant of this general approach and apply it to chemical prediction benchmarks until we either solve them or reach the limits of the approach. In this paper, we reformulate existing models into a single common framework we call Message Passing Neural Networks (MPNNs) and explore additional novel variations within this framework. Using MPNNs we demonstrate state of the art results on an important molecular property prediction benchmark; these results are strong enough that we believe future work should focus on datasets with larger molecules or more accurate ground truth labels.}, + archiveprefix = {arXiv}, + keywords = {DeepMind,GCN,GNN,Google,ML,molecules,MPNN}, + file = {/home/johannes/Nextcloud/Zotero/Gilmer et al_2017_Neural Message Passing for Quantum Chemistry.pdf;/home/johannes/Zotero/storage/A2EV2Y8T/1704.html} +} + +@book{girvinModernCondensedMatter2019, + title = {Modern {{Condensed Matter Physics}}}, + author = {Girvin, Steven M. and Yang, Kun}, + date = {2019-02-28}, + publisher = {{Cambridge University Press}}, + doi = {10.1017/9781316480649}, + url = {https://www.cambridge.org/highereducation/books/modern-condensed-matter-physics/F0A27AC5DEA8A40EA6EA5D727ED8B14E}, + urldate = {2022-06-18}, + abstract = {Modern Condensed Matter Physics brings together the most important advances in the field of recent decades. It provides instructors teaching graduate-level condensed matter courses with a comprehensive and in-depth textbook that will prepare graduate students for research or further study as well as reading more advanced and specialized books and research literature in the field. This textbook covers the basics of crystalline solids as well as analogous optical lattices and photonic crystals, while discussing cutting-edge topics such as disordered systems, mesoscopic systems, many-body systems, quantum magnetism, Bose–Einstein condensates, quantum entanglement, and superconducting quantum bits. Students are provided with the appropriate mathematical background to understand the topological concepts that have been permeating the field, together with numerous physical examples ranging from the fractional quantum Hall effect to topological insulators, the toric code, and majorana fermions. Exercises, commentary boxes, and appendices afford guidance and feedback for beginners and experts alike.}, + isbn = {978-1-316-48064-9}, + langid = {english}, + keywords = {_tablet,condensed matter,graduate,magnetism,superconductor,textbook,topological insulator}, + file = {/home/johannes/Nextcloud/Zotero/Girvin_Yang_2019_Modern Condensed Matter Physics.pdf;/home/johannes/Zotero/storage/3FP65JQ3/F0A27AC5DEA8A40EA6EA5D727ED8B14E.html} +} + +@article{glielmoAccurateInteratomicForce2017, + title = {Accurate Interatomic Force Fields via Machine Learning with Covariant Kernels}, + author = {Glielmo, Aldo and Sollich, Peter and De Vita, Alessandro}, + date = {2017-06-08}, + journaltitle = {Physical Review B}, + shortjournal = {Phys. Rev. B}, + volume = {95}, + number = {21}, + pages = {214302}, + publisher = {{American Physical Society}}, + doi = {10.1103/PhysRevB.95.214302}, + url = {https://link.aps.org/doi/10.1103/PhysRevB.95.214302}, + urldate = {2021-10-19}, + abstract = {We present a novel scheme to accurately predict atomic forces as vector quantities, rather than sets of scalar components, by Gaussian process (GP) regression. This is based on matrix-valued kernel functions, on which we impose the requirements that the predicted force rotates with the target configuration and is independent of any rotations applied to the configuration database entries. We show that such covariant GP kernels can be obtained by integration over the elements of the rotation group SO(d) for the relevant dimensionality d. Remarkably, in specific cases the integration can be carried out analytically and yields a conservative force field that can be recast into a pair interaction form. Finally, we show that restricting the integration to a summation over the elements of a finite point group relevant to the target system is sufficient to recover an accurate GP. The accuracy of our kernels in predicting quantum-mechanical forces in real materials is investigated by tests on pure and defective Ni, Fe, and Si crystalline systems.}, + keywords = {ML,nonscalar learning target,vectorial learning target}, + file = {/home/johannes/Nextcloud/Zotero/Glielmo et al_2017_Accurate interatomic force fields via machine learning with covariant kernels.pdf;/home/johannes/Zotero/storage/LVBU2R8M/Glielmo et al. - 2017 - Accurate interatomic force fields via machine lear.pdf;/home/johannes/Zotero/storage/RVCFAL4C/PhysRevB.95.html} +} + +@unpublished{glielmoDADApyDistancebasedAnalysis2022, + title = {{{DADApy}}: {{Distance-based Analysis}} of {{DAta-manifolds}} in {{Python}}}, + shorttitle = {{{DADApy}}}, + author = {Glielmo, Aldo and Macocco, Iuri and Doimo, Diego and Carli, Matteo and Zeni, Claudio and Wild, Romina and d' Errico, Maria and Rodriguez, Alex and Laio, Alessandro}, + options = {useprefix=true}, + date = {2022-05-04}, + eprint = {2205.03373}, + eprinttype = {arxiv}, + primaryclass = {physics, stat}, + url = {http://arxiv.org/abs/2205.03373}, + urldate = {2022-05-11}, + abstract = {DADApy is a python software package for analysing and characterising high-dimensional data manifolds. It provides methods for estimating the intrinsic dimension and the probability density, for performing density-based clustering and for comparing different distance metrics. We review the main functionalities of the package and exemplify its usage in toy cases and in a real-world application. The package is freely available under the open-source Apache 2.0 license and can be downloaded from the Github page https://github.com/sissa-data-science/DADApy.}, + archiveprefix = {arXiv}, + keywords = {clustering,DADApy,data exploration,density estimation,feature importance,Information imbalance,intrinsic dimension,kNN,library,Manifolds,MD,SISSA,unsupervised learning,with-code}, + file = {/home/johannes/Nextcloud/Zotero/Glielmo et al_2022_DADApy.pdf;/home/johannes/Zotero/storage/K2ZVKUHA/2205.html} +} + +@unpublished{glielmoRankingInformationContent2021, + title = {Ranking the Information Content of Distance Measures}, + author = {Glielmo, Aldo and Zeni, Claudio and Cheng, Bingqing and Csanyi, Gabor and Laio, Alessandro}, + date = {2021-04-30}, + eprint = {2104.15079}, + eprinttype = {arxiv}, + primaryclass = {cs, math, stat}, + url = {http://arxiv.org/abs/2104.15079}, + urldate = {2021-05-08}, + abstract = {Real-world data typically contain a large number of features that are often heterogeneous in nature, relevance, and also units of measure. When assessing the similarity between data points, one can build various distance measures using subsets of these features. Using the fewest features but still retaining sufficient information about the system is crucial in many statistical learning approaches, particularly when data are sparse. We introduce a statistical test that can assess the relative information retained when using two different distance measures, and determine if they are equivalent, independent, or if one is more informative than the other. This in turn allows finding the most informative distance measure out of a pool of candidates. The approach is applied to find the most relevant policy variables for controlling the Covid-19 epidemic and to find compact yet informative representations of atomic structures, but its potential applications are wide ranging in many branches of science.}, + archiveprefix = {arXiv}, + keywords = {ACSF,descriptor dimred,descriptors,descriptors analysis,ML,SOAP}, + file = {/home/johannes/Nextcloud/Zotero/Glielmo et al_2021_Ranking the information content of distance measures.pdf;/home/johannes/Zotero/storage/IHJU7M5J/2104.html} +} + +@article{glielmoRankingInformationContent2022, + title = {Ranking the Information Content of Distance Measures}, + author = {Glielmo, Aldo and Zeni, Claudio and Cheng, Bingqing and Csányi, Gábor and Laio, Alessandro}, + date = {2022-06-24}, + journaltitle = {PNAS Nexus}, + volume = {1}, + number = {2}, + pages = {pgac039}, + publisher = {{Proceedings of the National Academy of Sciences}}, + doi = {10.1093/pnasnexus/pgac039}, + url = {https://www.pnas.org/doi/full/10.1093/pnasnexus/pgac039}, + urldate = {2022-07-02}, + keywords = {ACSF,descriptor comparison,descriptors,dimensionality reduction,GPR,information imbalance,MD,ML,SOAP}, + file = {/home/johannes/Nextcloud/Zotero/Glielmo et al_2022_Ranking the information content of distance measures.pdf} +} + +@online{GlobalMLOpsML, + title = {Global {{MLOps}} and {{ML}} Tools Landscape | {{MLReef}}}, + url = {https://about.mlreef.com/blog/global-mlops-and-ml-tools-landscape}, + urldate = {2021-05-13}, + keywords = {ML,MLOps}, + file = {/home/johannes/Zotero/storage/LGPLQHSH/global-mlops-and-ml-tools-landscape.html} +} + +@article{golzeGWCompendiumPractical2019, + title = {The {{GW Compendium}}: {{A Practical Guide}} to {{Theoretical Photoemission Spectroscopy}}}, + shorttitle = {The {{GW Compendium}}}, + author = {Golze, Dorothea and Dvorak, Marc and Rinke, Patrick}, + date = {2019}, + journaltitle = {Frontiers in Chemistry}, + shortjournal = {Front. Chem.}, + volume = {7}, + publisher = {{Frontiers}}, + issn = {2296-2646}, + doi = {10.3389/fchem.2019.00377}, + url = {https://www.frontiersin.org/articles/10.3389/fchem.2019.00377/full?utm_source=ad&utm_medium=fb&utm_campaign=ba_sci_fchem}, + urldate = {2021-05-13}, + abstract = {The GW approximation in electronic structure theory has become a widespread tool for predicting electronic excitations in chemical compounds and materials. In the realm of theoretical spectroscopy, the GW method provides access to charged excitations as measured in direct or inverse photoemission spectroscopy. The number of GW calculations in the past two decades has exploded with increased computing power and modern codes. The success of GW can be attributed to many factors: favorable scaling with respect to system size, a formal interpretation for charged excitation energies, the importance of dynamical screening in real systems, and its practical combination with other theories. In this review, we provide an overview of these formal and practical considerations. We expand, in detail, on the choices presented to the scientist performing GW calculations for the first time. We also give an introduction to the many-body theory behind GW, a review of modern applications like molecules and surfaces, and a perspective on methods which go beyond conventional GW calculations. This review addresses chemists, physicists and material scientists with an interest in theoretical spectroscopy. It is intended for newcomers to GW calculations but can also serve as an alternative perspective for experts and an up-to-date source of computational techniques.}, + langid = {english}, + keywords = {electronic structure theory,GW approximation,Many-body theory,PES,photoemission,physics,review}, + file = {/home/johannes/Nextcloud/Zotero/Golze et al_2019_The GW Compendium.pdf} +} + +@unpublished{goodallRapidDiscoveryStable2022, + title = {Rapid {{Discovery}} of {{Stable Materials}} by {{Coordinate-free Coarse Graining}}}, + author = {Goodall, Rhys E. A. and Parackal, Abhijith S. and Faber, Felix A. and Armiento, Rickard and Lee, Alpha A.}, + date = {2022-03-15}, + eprint = {2106.11132}, + eprinttype = {arxiv}, + primaryclass = {cond-mat, physics:physics}, + url = {http://arxiv.org/abs/2106.11132}, + urldate = {2022-05-09}, + abstract = {A fundamental challenge in materials science pertains to elucidating the relationship between stoichiometry, stability, structure, and property. Recent advances have shown that machine learning can be used to learn such relationships, allowing the stability and functional properties of materials to be accurately predicted. However, most of these approaches use atomic coordinates as input and are thus bottle-necked by crystal structure identification when investigating novel materials. Our approach solves this bottleneck by coarse-graining the infinite search space of atomic coordinates into a combinatorially enumerable search space. The key idea is to use Wyckoff representations -- coordinate-free sets of symmetry-related positions in a crystal -- as the input to a machine learning model. Our model demonstrates exceptionally high precision in discovering new theoretically stable materials, identifying 1,569 materials that lie below the known convex hull of previously calculated materials from just 5,675 ab-initio calculations. Our approach opens up fundamental advances in computational materials discovery.}, + archiveprefix = {arXiv}, + keywords = {Wyckoff representation}, + file = {/home/johannes/Nextcloud/Zotero/Goodall et al_2022_Rapid Discovery of Stable Materials by Coordinate-free Coarse Graining.pdf;/home/johannes/Zotero/storage/G7U8SY86/2106.html} +} + +@misc{goodallRapidDiscoveryStable2022a, + title = {Rapid {{Discovery}} of {{Stable Materials}} by {{Coordinate-free Coarse Graining}}}, + author = {Goodall, Rhys E. A. and Parackal, Abhijith S. and Faber, Felix A. and Armiento, Rickard and Lee, Alpha A.}, + date = {2022-03-15}, + number = {arXiv:2106.11132}, + eprint = {2106.11132}, + eprinttype = {arxiv}, + primaryclass = {cond-mat, physics:physics}, + publisher = {{arXiv}}, + doi = {10.48550/arXiv.2106.11132}, + url = {http://arxiv.org/abs/2106.11132}, + urldate = {2022-10-03}, + abstract = {A fundamental challenge in materials science pertains to elucidating the relationship between stoichiometry, stability, structure, and property. Recent advances have shown that machine learning can be used to learn such relationships, allowing the stability and functional properties of materials to be accurately predicted. However, most of these approaches use atomic coordinates as input and are thus bottle-necked by crystal structure identification when investigating novel materials. Our approach solves this bottleneck by coarse-graining the infinite search space of atomic coordinates into a combinatorially enumerable search space. The key idea is to use Wyckoff representations -- coordinate-free sets of symmetry-related positions in a crystal -- as the input to a machine learning model. Our model demonstrates exceptionally high precision in discovering new theoretically stable materials, identifying 1,569 materials that lie below the known convex hull of previously calculated materials from just 5,675 ab-initio calculations. Our approach opens up fundamental advances in computational materials discovery.}, + archiveprefix = {arXiv}, + keywords = {CGCNN,Condensed Matter - Materials Science,crystal structure,crystal symmetry,GNN,MPNN,original publication,Physics - Computational Physics,regression,Wren,Wyckoff positions,Wyckoff representation}, + file = {/home/johannes/Nextcloud/Zotero/Goodall et al_2022_Rapid Discovery of Stable Materials by Coordinate-free Coarse Graining2.pdf;/home/johannes/Zotero/storage/8I7WCWRJ/2106.html} +} + +@inproceedings{goodallWyckoffSetRegression2020, + title = {Wyckoff {{Set Regression}} for {{Materials Discovery}}}, + author = {Goodall, Rhys E A and Parackal, Abhijith S and Faber, Felix A and Armiento, Rickard}, + date = {2020}, + pages = {7}, + url = {https://ml4physicalsciences.github.io/2020/files/NeurIPS_ML4PS_2020_21.pdf}, + abstract = {In recent years machine learning has been shown to be able to approximate the accuracy and amortise the computational cost of ab-initio quantum mechanics calculations. This has opened up many exciting use cases in the study of materials in-silico. However, the majority of the these works make use of atomic positions as inputs which limits their application to novel material discovery applications where crystal structures are a-priori unknown. For a model to see useful application in materials discovery we need to be able to enumerate its inputs over a possible design space of new materials. In this work, we build upon a recent machine learning framework for material science that operates on the stoichiometry of materials and extend it to look at Wyckoff sets. We show that operating on Wyckoff sets allows the model to handle compositions with multiple polymorphs, therefore, overcoming one of the major limitations of composition-based models whilst maintaining the key benefit of having a combinatorially enumerable input space.}, + eventtitle = {{{NeurIPS}}}, + langid = {english}, + keywords = {CGCNN,crystal structure,crystal symmetry,GNN,MPNN,original publication,regression,Wren,Wyckoff positions,Wyckoff representation}, + file = {/home/johannes/Zotero/storage/A5G9FYMS/Goodall et al_Wyckoff Set Regression for Materials Discovery.pdf} +} + +@online{GooglePythonStyle, + title = {Google {{Python Style Guide}}}, + url = {https://google.github.io/styleguide/pyguide.html}, + urldate = {2021-09-23}, + abstract = {Style guides for Google-originated open-source projects}, + langid = {american}, + organization = {{styleguide}}, + keywords = {coding style guide,Python,software engineering}, + file = {/home/johannes/Zotero/storage/HRL7NEIR/pyguide.html} +} + +@book{gorelickHighPerformancePython2020, + title = {High Performance {{Python}}: Practical Performance Programming for Humans}, + shorttitle = {High Performance {{Python}}}, + author = {Gorelick, Micha and Ozsvald, Ian}, + date = {2020}, + edition = {Second edition}, + publisher = {{O'Reilly}}, + location = {{Beijing}}, + isbn = {978-1-4920-5502-0}, + langid = {english}, + pagetotal = {444}, + file = {/home/johannes/Books/software_engineering/performance_engineering/Gorelick_HighPerformancePython_2e-2020.pdf} +} + +@unpublished{goscinskiOptimalRadialBasis2021, + title = {Optimal Radial Basis for Density-Based Atomic Representations}, + author = {Goscinski, Alexander and Musil, Félix and Pozdnyakov, Sergey and Ceriotti, Michele}, + date = {2021-05-18}, + eprint = {2105.08717}, + eprinttype = {arxiv}, + primaryclass = {physics, stat}, + url = {http://arxiv.org/abs/2105.08717}, + urldate = {2021-05-30}, + abstract = {The input of almost every machine learning algorithm targeting the properties of matter at the atomic scale involves a transformation of the list of Cartesian atomic coordinates into a more symmetric representation. Many of these most popular representations can be seen as an expansion of the symmetrized correlations of the atom density, and differ mainly by the choice of basis. Here we discuss how to build an adaptive, optimal numerical basis that is chosen to represent most efficiently the structural diversity of the dataset at hand. For each training dataset, this optimal basis is unique, and can be computed at no additional cost with respect to the primitive basis by approximating it with splines. We demonstrate that this construction yields representations that are accurate and computationally efficient, presenting examples that involve both molecular and condensed-phase machine-learning models.}, + archiveprefix = {arXiv}, + keywords = {descriptors,descriptors analysis,ML,SOAP}, + file = {/home/johannes/Nextcloud/Zotero/false;/home/johannes/Nextcloud/Zotero/Goscinski et al_2021_Optimal radial basis for density-based atomic representations.pdf;/home/johannes/Zotero/storage/HHPU5HMP/2105.html} +} + +@article{goscinskiRoleFeatureSpace2021, + title = {The Role of Feature Space in Atomistic Learning}, + author = {Goscinski, Alexander and Fraux, Guillaume and Imbalzano, Giulio and Ceriotti, Michele}, + date = {2021-04}, + journaltitle = {Machine Learning: Science and Technology}, + shortjournal = {Mach. Learn.: Sci. Technol.}, + volume = {2}, + number = {2}, + pages = {025028}, + publisher = {{IOP Publishing}}, + issn = {2632-2153}, + doi = {10.1088/2632-2153/abdaf7}, + url = {https://doi.org/10.1088/2632-2153/abdaf7}, + urldate = {2021-05-13}, + abstract = {Efficient, physically-inspired descriptors of the structure and composition of molecules and materials play a key role in the application of machine-learning techniques to atomistic simulations. The proliferation of approaches, as well as the fact that each choice of features can lead to very different behavior depending on how they are used, e.g. by introducing non-linear kernels and non-Euclidean metrics to manipulate them, makes it difficult to objectively compare different methods, and to address fundamental questions on how one feature space is related to another. In this work we introduce a framework to compare different sets of descriptors, and different ways of transforming them by means of metrics and kernels, in terms of the structure of the feature space that they induce. We define diagnostic tools to determine whether alternative feature spaces contain equivalent amounts of information, and whether the common information is substantially distorted when going from one feature space to another. We compare, in particular, representations that are built in terms of n-body correlations of the atom density, quantitatively assessing the information loss associated with the use of low-order features. We also investigate the impact of different choices of basis functions and hyperparameters of the widely used SOAP and Behler–Parrinello features, and investigate how the use of non-linear kernels, and of a Wasserstein-type metric, change the structure of the feature space in comparison to a simpler linear feature space.}, + langid = {english}, + keywords = {BPSF,descriptor comparison,descriptors analysis,ML,SOAP}, + file = {/home/johannes/Nextcloud/Zotero/Goscinski et al_2021_The role of feature space in atomistic learning.pdf} +} + +@misc{grisafiElectronicstructurePropertiesAtomcentered2022, + title = {Electronic-Structure Properties from Atom-Centered Predictions of the Electron Density}, + author = {Grisafi, Andrea and Lewis, Alan M. and Rossi, Mariana and Ceriotti, Michele}, + date = {2022-06-28}, + number = {arXiv:2206.14087}, + eprint = {2206.14087}, + eprinttype = {arxiv}, + primaryclass = {cond-mat, physics:physics, stat}, + publisher = {{arXiv}}, + doi = {10.48550/arXiv.2206.14087}, + url = {http://arxiv.org/abs/2206.14087}, + urldate = {2022-07-02}, + abstract = {The electron density of a molecule or material has recently received major attention as a target quantity of machine-learning models. A natural choice to construct a model that yields transferable and linear-scaling predictions is to represent the scalar field using a multi-centered atomic basis analogous to that routinely used in density fitting approximations. However, the non-orthogonality of the basis poses challenges for the learning exercise, as it requires accounting for all the atomic density components at once. We devise a gradient-based approach to directly minimize the loss function of the regression problem in an optimized and highly sparse feature space. In so doing, we overcome the limitations associated with adopting an atom-centered model to learn the electron density over arbitrarily complex datasets, obtaining extremely accurate predictions. The enhanced framework is tested on 32-molecule periodic cells of liquid water, presenting enough complexity to require an optimal balance between accuracy and computational efficiency. We show that starting from the predicted density a single Kohn-Sham diagonalization step can be performed to access total energy components that carry an error of just 0.1 meV/atom with respect to the reference density functional calculations. Finally, we test our method on the highly heterogeneous QM9 benchmark dataset, showing that a small fraction of the training data is enough to derive ground-state total energies within chemical accuracy.}, + archiveprefix = {arXiv}, + keywords = {_tablet,DFT,dimensionality reduction,ML-DFT,ML-ESM,molecules,molecules & solids,prediction of electron density,QM9,SALTED,SOAP,solids}, + file = {/home/johannes/Nextcloud/Zotero/Grisafi et al_2022_Electronic-structure properties from atom-centered predictions of the electron.pdf;/home/johannes/Zotero/storage/QPHBS33I/2206.html} +} + +@article{grisafiIncorporatingLongrangePhysics2019, + title = {Incorporating Long-Range Physics in Atomic-Scale Machine Learning}, + author = {Grisafi, Andrea and Ceriotti, Michele}, + date = {2019-11-28}, + journaltitle = {The Journal of Chemical Physics}, + shortjournal = {J. Chem. Phys.}, + volume = {151}, + number = {20}, + pages = {204105}, + publisher = {{American Institute of Physics}}, + issn = {0021-9606}, + doi = {10.1063/1.5128375}, + url = {https://aip.scitation.org/doi/full/10.1063/1.5128375}, + urldate = {2022-08-16}, + abstract = {The most successful and popular machine learning models of atomic-scale properties derive their transferability from a locality ansatz. The properties of a large molecule or a bulk material are written as a sum over contributions that depend on the configurations within finite atom-centered environments. The obvious downside of this approach is that it cannot capture nonlocal, nonadditive effects such as those arising due to long-range electrostatics or quantum interference. We propose a solution to this problem by introducing nonlocal representations of the system, which are remapped as feature vectors that are defined locally and are equivariant in O(3). We consider, in particular, one form that has the same asymptotic behavior as the electrostatic potential. We demonstrate that this framework can capture nonlocal, long-range physics by building a model for the electrostatic energy of randomly distributed point-charges, for the unrelaxed binding curves of charged organic molecular dimers, and for the electronic dielectric response of liquid water. By combining a representation of the system that is sensitive to long-range correlations with the transferability of an atom-centered additive model, this method outperforms current state-of-the-art machine-learning schemes and provides a conceptual framework to incorporate nonlocal physics into atomistic machine learning.}, + keywords = {ML-ESM,SA-GPR}, + file = {/home/johannes/Nextcloud/Zotero/Grisafi_Ceriotti_2019_Incorporating long-range physics in atomic-scale machine learning.pdf} +} + +@article{grisafiMultiscaleApproachPrediction2021, + title = {Multi-Scale Approach for the Prediction of Atomic Scale Properties}, + author = {Grisafi, Andrea and Nigam, Jigyasa and Ceriotti, Michele}, + date = {2021-02-18}, + journaltitle = {Chemical Science}, + shortjournal = {Chem. Sci.}, + volume = {12}, + number = {6}, + pages = {2078--2090}, + publisher = {{The Royal Society of Chemistry}}, + issn = {2041-6539}, + doi = {10.1039/D0SC04934D}, + url = {https://pubs.rsc.org/en/content/articlelanding/2021/sc/d0sc04934d}, + urldate = {2022-08-16}, + abstract = {Electronic nearsightedness is one of the fundamental principles that governs the behavior of condensed matter and supports its description in terms of local entities such as chemical bonds. Locality also underlies the tremendous success of machine-learning schemes that predict quantum mechanical observables – such as the cohesive energy, the electron density, or a variety of response properties – as a sum of atom-centred contributions, based on a short-range representation of atomic environments. One of the main shortcomings of these approaches is their inability to capture physical effects ranging from electrostatic interactions to quantum delocalization, which have a long-range nature. Here we show how to build a multi-scale scheme that combines in the same framework local and non-local information, overcoming such limitations. We show that the simplest version of such features can be put in formal correspondence with a multipole expansion of permanent electrostatics. The data-driven nature of the model construction, however, makes this simple form suitable to tackle also different types of delocalized and collective effects. We present several examples that range from molecular physics to surface science and biophysics, demonstrating the ability of this multi-scale approach to model interactions driven by electrostatics, polarization and dispersion, as well as the cooperative behavior of dielectric response functions.}, + langid = {english}, + keywords = {ML-ESM,SA-GPR}, + file = {/home/johannes/Nextcloud/Zotero/Grisafi et al_2021_Multi-scale approach for the prediction of atomic scale properties.pdf;/home/johannes/Zotero/storage/A6YSFTUU/Grisafi et al. - 2021 - Multi-scale approach for the prediction of atomic .pdf;/home/johannes/Zotero/storage/SG2EBGXV/d0sc04934d.html} +} + +@article{grisafiSymmetryAdaptedMachineLearning2018, + title = {Symmetry-{{Adapted Machine Learning}} for {{Tensorial Properties}} of {{Atomistic Systems}}}, + author = {Grisafi, Andrea and Wilkins, David M. and Csányi, Gábor and Ceriotti, Michele}, + date = {2018-01-19}, + journaltitle = {Physical Review Letters}, + shortjournal = {Phys. Rev. Lett.}, + volume = {120}, + number = {3}, + pages = {036002}, + publisher = {{American Physical Society}}, + doi = {10.1103/PhysRevLett.120.036002}, + url = {https://link.aps.org/doi/10.1103/PhysRevLett.120.036002}, + urldate = {2021-10-19}, + abstract = {Statistical learning methods show great promise in providing an accurate prediction of materials and molecular properties, while minimizing the need for computationally demanding electronic structure calculations. The accuracy and transferability of these models are increased significantly by encoding into the learning procedure the fundamental symmetries of rotational and permutational invariance of scalar properties. However, the prediction of tensorial properties requires that the model respects the appropriate geometric transformations, rather than invariance, when the reference frame is rotated. We introduce a formalism that extends existing schemes and makes it possible to perform machine learning of tensorial properties of arbitrary rank, and for general molecular geometries. To demonstrate it, we derive a tensor kernel adapted to rotational symmetry, which is the natural generalization of the smooth overlap of atomic positions kernel commonly used for the prediction of scalar properties at the atomic scale. The performance and generality of the approach is demonstrated by learning the instantaneous response to an external electric field of water oligomers of increasing complexity, from the isolated molecule to the condensed phase.}, + keywords = {GPR,lambda-SOAP,library,ML,models,nonscalar learning target,original publication,SA-GPR,with-code}, + file = {/home/johannes/Nextcloud/Zotero/Grisafi et al_2018_Symmetry-Adapted Machine Learning for Tensorial Properties of Atomistic Systems.pdf;/home/johannes/Zotero/storage/JGVW5QVD/Grisafi et al. - 2018 - Symmetry-Adapted Machine Learning for Tensorial Pr.pdf;/home/johannes/Zotero/storage/KLCEZH25/PhysRevLett.120.html} +} + +@article{grisafiTransferableMachineLearningModel2019, + title = {Transferable {{Machine-Learning Model}} of the {{Electron Density}}}, + author = {Grisafi, Andrea and Fabrizio, Alberto and Meyer, Benjamin and Wilkins, David M. and Corminboeuf, Clemence and Ceriotti, Michele}, + date = {2019-01-23}, + journaltitle = {ACS Central Science}, + shortjournal = {ACS Cent. Sci.}, + volume = {5}, + number = {1}, + pages = {57--64}, + publisher = {{American Chemical Society}}, + issn = {2374-7943}, + doi = {10.1021/acscentsci.8b00551}, + url = {https://doi.org/10.1021/acscentsci.8b00551}, + urldate = {2021-10-19}, + abstract = {The electronic charge density plays a central role in determining the behavior of matter at the atomic scale, but its computational evaluation requires demanding electronic-structure calculations. We introduce an atom-centered, symmetry-adapted framework to machine-learn the valence charge density based on a small number of reference calculations. The model is highly transferable, meaning it can be trained on electronic-structure data of small molecules and used to predict the charge density of larger compounds with low, linear-scaling cost. Applications are shown for various hydrocarbon molecules of increasing complexity and flexibility, and demonstrate the accuracy of the model when predicting the density on octane and octatetraene after training exclusively on butane and butadiene. This transferable, data-driven model can be used to interpret experiments, accelerate electronic structure calculations, and compute electrostatic interactions in molecules and condensed-phase systems.}, + keywords = {_tablet,DFT,GPR,lambda-SOAP,library,ML,ML-DFT,ML-ESM,models,molecules,prediction of electron density,SA-GPR,SALTED,SOAP,with-code}, + file = {/home/johannes/Nextcloud/Zotero/Grisafi et al_2019_Transferable Machine-Learning Model of the Electron Density.pdf;/home/johannes/Zotero/storage/HMBCGARZ/acscentsci.html} +} + +@book{grusDataScienceScratch2019, + title = {Data Science from Scratch: First Principles with {{Python}}}, + shorttitle = {Data Science from Scratch}, + author = {Grus, Joel}, + date = {2019}, + edition = {Second edition}, + publisher = {{O'Reilly Media}}, + location = {{Sebastopol, CA}}, + isbn = {978-1-4920-4113-9}, + pagetotal = {384}, + keywords = {data science,general,practice,python}, + annotation = {OCLC: on1060198620}, + file = {/home/johannes/Books/data_science/general_practice/Grus_DataScienceFromScratchPython_2e-2019.epub} +} + +@inproceedings{gundersenStateArtReproducibility2018, + title = {State of the {{Art}}: {{Reproducibility}} in {{Artificial Intelligence}}}, + shorttitle = {State of the {{Art}}}, + booktitle = {Thirty-{{Second AAAI Conference}} on {{Artificial Intelligence}}}, + author = {Gundersen, Odd Erik and Kjensmo, Sigbjørn}, + date = {2018-04-25}, + url = {https://www.aaai.org/ocs/index.php/AAAI/AAAI18/paper/view/17248}, + urldate = {2021-10-23}, + abstract = {Background: Research results in artificial intelligence (AI) are criticized for not being reproducible. Objective: To quantify the state of reproducibility of empirical AI research using six reproducibility metrics measuring three different degrees of reproducibility. Hypotheses: 1) AI research is not documented well enough to reproduce the reported results. 2) Documentation practices have improved over time. Method: The literature is reviewed and a set of variables that should be documented to enable reproducibility are grouped into three factors: Experiment, Data and Method. The metrics describe how well the factors have been documented for a paper. A total of 400 research papers from the conference series IJCAI and AAAI have been surveyed using the metrics. Findings: None of the papers document all of the variables. The metrics show that between 20\% and 30\% of the variables for each factor are documented. One of the metrics show statistically significant increase over time while the others show no change. Interpretation: The reproducibility scores decrease with in- creased documentation requirements. Improvement over time is found. Conclusion: Both hypotheses are supported.}, + eventtitle = {Thirty-{{Second AAAI Conference}} on {{Artificial Intelligence}}}, + langid = {english}, + file = {/home/johannes/Nextcloud/Zotero/Gundersen_Kjensmo_2018_State of the Art.pdf;/home/johannes/Zotero/storage/PXBAF83Q/17248.html} +} + +@misc{gutmannPenPaperExercises2022, + title = {Pen and {{Paper Exercises}} in {{Machine Learning}}}, + author = {Gutmann, Michael U.}, + date = {2022-06-27}, + number = {arXiv:2206.13446}, + eprint = {2206.13446}, + eprinttype = {arxiv}, + primaryclass = {cs, stat}, + publisher = {{arXiv}}, + doi = {10.48550/arXiv.2206.13446}, + url = {http://arxiv.org/abs/2206.13446}, + urldate = {2022-06-29}, + abstract = {This is a collection of (mostly) pen-and-paper exercises in machine learning. The exercises are on the following topics: linear algebra, optimisation, directed graphical models, undirected graphical models, expressive power of graphical models, factor graphs and message passing, inference for hidden Markov models, model-based learning (including ICA and unnormalised models), sampling and Monte-Carlo integration, and variational inference.}, + archiveprefix = {arXiv}, + keywords = {course material,exercises,General ML,graphical model,hidden Markov model,MC integration,ML,sampling,teaching}, + file = {/home/johannes/Nextcloud/Zotero/Gutmann_2022_Pen and Paper Exercises in Machine Learning.pdf;/home/johannes/Zotero/storage/KMSFX6RY/2206.html} +} + +@article{hafizHighthroughputDataAnalysis2018, + title = {A High-Throughput Data Analysis and Materials Discovery Tool for Strongly Correlated Materials}, + author = {Hafiz, Hasnain and Khair, Adnan Ibne and Choi, Hongchul and Mueen, Abdullah and Bansil, Arun and Eidenbenz, Stephan and Wills, John and Zhu, Jian-Xin and Balatsky, Alexander V. and Ahmed, Towfiq}, + date = {2018-11-22}, + journaltitle = {npj Computational Materials}, + volume = {4}, + number = {1}, + pages = {1--9}, + publisher = {{Nature Publishing Group}}, + issn = {2057-3960}, + doi = {10.1038/s41524-018-0120-9}, + url = {https://www.nature.com/articles/s41524-018-0120-9}, + urldate = {2021-05-19}, + abstract = {Modeling of f-electron systems is challenging due to the complex interplay of the effects of spin–orbit coupling, electron–electron interactions, and the hybridization of the localized f-electrons with itinerant conduction electrons. This complexity drives not only the richness of electronic properties but also makes these materials suitable for diverse technological applications. In this context, we propose and implement a data-driven approach to aid the materials discovery process. By deploying state-of-the-art algorithms and query tools, we train our learning models using a large, simulated dataset based on existing actinide and lanthanide compounds. The machine-learned models so obtained can then be used to search for new classes of stable materials with desired electronic and physical properties. We discuss the basic structure of our f-electron database, and our approach towards cleaning and correcting the structure data files. Illustrative examples of the applications of our database include successful prediction of stable superstructures of double perovskites and identification of a number of physically-relevant trends in strongly correlated features of f-electron based materials.}, + issue = {1}, + langid = {english}, + keywords = {HTC,materials informatics,strongly correlated maeterials}, + file = {/home/johannes/Nextcloud/Zotero/Hafiz et al_2018_A high-throughput data analysis and materials discovery tool for strongly.pdf;/home/johannes/Zotero/storage/VAW7UFBH/s41524-018-0120-9.html} +} + +@article{handleyNextGenerationInteratomic2014, + title = {Next Generation Interatomic Potentials for Condensed Systems}, + author = {Handley, Christopher Michael and Behler, Jörg}, + date = {2014-07-07}, + journaltitle = {The European Physical Journal B}, + shortjournal = {Eur. Phys. J. B}, + volume = {87}, + number = {7}, + pages = {152}, + issn = {1434-6036}, + doi = {10.1140/epjb/e2014-50070-0}, + url = {https://doi.org/10.1140/epjb/e2014-50070-0}, + urldate = {2021-05-18}, + abstract = {The computer simulation of condensed systems is a challenging task. While electronic structure methods like density-functional theory (DFT) usually provide a good compromise between accuracy and efficiency, they are computationally very demanding and thus applicable only to systems containing up to a few hundred atoms. Unfortunately, many interesting problems require simulations to be performed on much larger systems involving thousands of atoms or more. Consequently, more efficient methods are urgently needed, and a lot of effort has been spent on the development of a large variety of potentials enabling simulations with significantly extended time and length scales. Most commonly, these potentials are based on physically motivated functional forms and thus perform very well for the applications they have been designed for. On the other hand, they are often highly system-specific and thus cannot easily be transferred from one system to another. Moreover, their numerical accuracy is restricted by the intrinsic limitations of the imposed functional forms. In recent years, several novel types of potentials have emerged, which are not based on physical considerations. Instead, they aim to reproduce a set of reference electronic structure data as accurately as possible by using very general and flexible functional forms. In this review we will survey a number of these methods. While they differ in the choice of the employed mathematical functions, they all have in common that they provide high-quality potential-energy surfaces, while the efficiency is comparable to conventional empirical potentials. It has been demonstrated that in many cases these potentials now offer a very interesting new approach to study complex systems with hitherto unreached accuracy.}, + langid = {english}, + keywords = {condensed,ML,MLP,models}, + file = {/home/johannes/Nextcloud/Zotero/Handley_Behler_2014_Next generation interatomic potentials for condensed systems.pdf} +} + +@article{hartlNationaleForschungsdateninfrastrukturNFDI2021, + title = {Nationale Forschungsdateninfrastruktur (NFDI)}, + author = {Hartl, Nathalie and Wössner, Elena and Sure-Vetter, York}, + date = {2021-10-01}, + journaltitle = {Informatik Spektrum}, + shortjournal = {Informatik Spektrum}, + volume = {44}, + number = {5}, + pages = {370--373}, + issn = {1432-122X}, + doi = {10.1007/s00287-021-01392-6}, + url = {https://doi.org/10.1007/s00287-021-01392-6}, + urldate = {2021-10-15}, + abstract = {In der Nationalen Forschungsdateninfrastruktur (NFDI) werden wertvolle Forschungsdaten für das gesamte deutsche Wissenschaftssystem systematisch erschlossen, vernetzt und nachhaltig nutzbar gemacht. Bislang sind diese meist dezentral, projektbezogen oder nur zeitlich begrenzt verfügbar. Mit der NFDI soll ein digitaler Wissensspeicher unter Berücksichtigung der FAIR-Prinzipien (Findable, Accessible, Interoperable, Reusable) geschaffen werden. Bereits vorhandene Daten können zur Bearbeitung weiterer Forschungsfragen genutzt werden und neue Erkenntnisse und Innovationen ermöglichen.Bis zu 30 NFDI-Konsortien, Zusammenschlüsse verschiedener Einrichtungen innerhalb eines Forschungsfeldes, arbeiten zusammen interdisziplinär an der Zielumsetzung. Um die Aktivitäten zum Aufbau einer Nationalen Forschungsdateninfrastruktur zu koordinieren, wurde der gemeinnützige Verein Nationale Forschungsdateninfrastruktur (NFDI) e.V. mit Sitz in Karlsruhe im Oktober 2020 gegründet. Gemeinsam gestalten Verein und NFDI-Konsortien die Zukunft des Forschungsdatenmanagements in Deutschland. Darüber hinaus soll NFDI auch am Aufbau internationaler Initiativen, beispielsweise der European Open Science Cloud (EOSC), mitwirken.}, + langid = {ngerman}, + file = {/home/johannes/Nextcloud/Zotero/Hartl et al_2021_Nationale Forschungsdateninfrastruktur (NFDI).pdf} +} + +@article{hartmaierDataOrientedConstitutiveModeling2020, + title = {Data-{{Oriented Constitutive Modeling}} of {{Plasticity}} in {{Metals}}}, + author = {Hartmaier, Alexander}, + date = {2020-01}, + journaltitle = {Materials}, + volume = {13}, + number = {7}, + pages = {1600}, + publisher = {{Multidisciplinary Digital Publishing Institute}}, + issn = {1996-1944}, + doi = {10.3390/ma13071600}, + url = {https://www.mdpi.com/1996-1944/13/7/1600}, + urldate = {2022-05-13}, + abstract = {Constitutive models for plastic deformation of metals are typically based on flow rules determining the transition from elastic to plastic response of a material as function of the applied mechanical load. These flow rules are commonly formulated as a yield function, based on the equivalent stress and the yield strength of the material, and its derivatives. In this work, a novel mathematical formulation is developed that allows the efficient use of machine learning algorithms describing the elastic-plastic deformation of a solid under arbitrary mechanical loads and that can replace the standard yield functions with more flexible algorithms. By exploiting basic physical principles of elastic-plastic deformation, the dimensionality of the problem is reduced without loss of generality. The data-oriented approach inherently offers a great flexibility to handle different kinds of material anisotropy without the need for explicitly calculating a large number of model parameters. The applicability of this formulation in finite element analysis is demonstrated, and the results are compared to formulations based on Hill-like anisotropic plasticity as reference model. In future applications, the machine learning algorithm can be trained by hybrid experimental and numerical data, as for example obtained from fundamental micromechanical simulations based on crystal plasticity models. In this way, data-oriented constitutive modeling will also provide a new way to homogenize numerical results in a scale-bridging approach.}, + issue = {7}, + langid = {english}, + keywords = {constitutive modeling,FEM,ML}, + file = {/home/johannes/Nextcloud/Zotero/Hartmaier_2020_Data-Oriented Constitutive Modeling of Plasticity in Metals.pdf;/home/johannes/Zotero/storage/94LTSN79/htm.html} +} + +@article{hasnipDensityFunctionalTheory2014, + title = {Density Functional Theory in the Solid State}, + author = {Hasnip, Philip J. and Refson, Keith and Probert, Matt I. J. and Yates, Jonathan R. and Clark, Stewart J. and Pickard, Chris J.}, + date = {2014-03-13}, + journaltitle = {Philosophical Transactions of the Royal Society A: Mathematical, Physical and Engineering Sciences}, + volume = {372}, + number = {2011}, + pages = {20130270}, + publisher = {{Royal Society}}, + doi = {10.1098/rsta.2013.0270}, + url = {https://royalsocietypublishing.org/doi/10.1098/rsta.2013.0270}, + urldate = {2022-05-18}, + abstract = {Density functional theory (DFT) has been used in many fields of the physical sciences, but none so successfully as in the solid state. From its origins in condensed matter physics, it has expanded into materials science, high-pressure physics and mineralogy, solid-state chemistry and more, powering entire computational subdisciplines. Modern DFT simulation codes can calculate a vast range of structural, chemical, optical, spectroscopic, elastic, vibrational and thermodynamic phenomena. The ability to predict structure–property relationships has revolutionized experimental fields, such as vibrational and solid-state NMR spectroscopy, where it is the primary method to analyse and interpret experimental spectra. In semiconductor physics, great progress has been made in the electronic structure of bulk and defect states despite the severe challenges presented by the description of excited states. Studies are no longer restricted to known crystallographic structures. DFT is increasingly used as an exploratory tool for materials discovery and computational experiments, culminating in ex nihilo crystal structure prediction, which addresses the long-standing difficult problem of how to predict crystal structure polymorphs from nothing but a specified chemical composition. We present an overview of the capabilities of solid-state DFT simulations in all of these topics, illustrated with recent examples using the CASTEP computer program.}, + keywords = {_tablet,condensed matter,DFT,review}, + file = {/home/johannes/Nextcloud/Zotero/Hasnip et al_2014_Density functional theory in the solid state.pdf} +} + +@article{helfrechtStructurepropertyMapsKernel2020, + title = {Structure-Property Maps with {{Kernel}} Principal Covariates Regression}, + author = {Helfrecht, Benjamin A. and Cersonsky, Rose K. and Fraux, Guillaume and Ceriotti, Michele}, + date = {2020-11}, + shortjournal = {Mach. Learn.: Sci. Technol.}, + volume = {1}, + number = {4}, + pages = {045021}, + publisher = {{IOP Publishing}}, + issn = {2632-2153}, + doi = {10.1088/2632-2153/aba9ef}, + url = {https://doi.org/10.1088/2632-2153/aba9ef}, + urldate = {2021-10-19}, + abstract = {Data analyses based on linear methods constitute the simplest, most robust, and transparent approaches to the automatic processing of large amounts of data for building supervised or unsupervised machine learning models. Principal covariates regression (PCovR) is an underappreciated method that interpolates between principal component analysis and linear regression and can be used conveniently to reveal structure-property relations in terms of simple-to-interpret, low-dimensional maps. Here we provide a pedagogic overview of these data analysis schemes, including the use of the kernel trick to introduce an element of non-linearity while maintaining most of the convenience and the simplicity of linear approaches. We then introduce a kernelized version of PCovR and a sparsified extension, and demonstrate the performance of this approach in revealing and predicting structure-property relations in chemistry and materials science, showing a variety of examples including elemental carbon, porous silicate frameworks, organic molecules, amino acid conformers, and molecular materials.}, + langid = {english}, + keywords = {KPCovR,models,original publication,PCovR}, + file = {/home/johannes/Nextcloud/Zotero/Helfrecht et al_2020_Structure-property maps with Kernel principal covariates regression.pdf} +} + +@article{henkTopologicalCharacterMagnetism2012, + title = {Topological {{Character}} and {{Magnetism}} of the {{Dirac State}} in {{Mn-Doped}} \$\{\textbackslash mathrm\{\vphantom{\}\}}{{Bi}}\vphantom\{\}\vphantom\{\}\_\{2\}\{\textbackslash mathrm\{\vphantom{\}\}}{{Te}}\vphantom\{\}\vphantom\{\}\_\{3\}\$}, + author = {Henk, J. and Flieger, M. and Maznichenko, I. V. and Mertig, I. and Ernst, A. and Eremeev, S. V. and Chulkov, E. V.}, + date = {2012-08-16}, + journaltitle = {Physical Review Letters}, + shortjournal = {Phys. Rev. Lett.}, + volume = {109}, + number = {7}, + pages = {076801}, + publisher = {{American Physical Society}}, + doi = {10.1103/PhysRevLett.109.076801}, + url = {https://link.aps.org/doi/10.1103/PhysRevLett.109.076801}, + urldate = {2022-05-13}, + abstract = {First-principles and model calculations show that the Dirac surface state of the topological insulator Bi2Te3 survives upon moderate Mn doping of the surface layers but can lose its topological character as a function of magnetization direction. The dispersion depends considerably on the direction of the Mn magnetization: for perpendicular magnetization, a gap of 16 meV opens up at the Dirac point; for in-plane magnetization, a tiny gap can be opened or closed in dependence on the magnetization azimuth. The ground state is ferromagnetic, with a critical temperature of 12 K. The results provide a path towards a magnetic control of the topological character of the Dirac surface state and its consequences to spin-dependent transport properties.}, + file = {/home/johannes/Nextcloud/Zotero/Henk et al_2012_Topological Character and Magnetism of the Dirac State in Mn-Doped.pdf;/home/johannes/Zotero/storage/W6BV33VI/Henk et al. - 2012 - Topological Character and Magnetism of the Dirac S.pdf;/home/johannes/Zotero/storage/ZTFJBVIM/PhysRevLett.109.html} +} + +@article{herbstDFTKJulianApproach2021, + title = {{{DFTK}}: {{A Julian}} Approach for Simulating Electrons in Solids}, + shorttitle = {{{DFTK}}}, + author = {Herbst, Michael F. and Levitt, A. and Cancès, E.}, + date = {2021}, + journaltitle = {JuliaCon Proceedings}, + doi = {10.21105/JCON.00069}, + abstract = {Density-functional theory (DFT) is a widespread method for simulating the quantum-chemical behaviour of electrons in matter. It provides a first-principles description of many optical, mechanical and chemical properties at an acceptable computational cost [16, 2, 3]. For a wide range of systems the obtained predictions are accurate and shortcomings of the theory are by now wellunderstood [2, 3]. The desire to tackle even bigger systems and more involved materials, however, keeps posing novel challenges that require methods to constantly improve. One example are socalled high-throughput screening approaches, which are becoming prominent in recent years. In these techniques one wishes to systematically scan over huge design spaces of compounds in order to identify promising novel materials for targeted follow-up investigation. This has already lead to many success stories [14], such as the discovery of novel earth-abundant semiconductors [11], novel light-absorbing materials [20], electrocatalysts [8], materials for hydrogen storage [13] or for Li-ion batteries [1]. Keeping in mind the large range of physics that needs to be covered in these studies as well as the typical number of calculations (up to the order of millions), a bottleneck in these studies is the reliability and performance of the underlying DFT codes.}, + file = {/home/johannes/Nextcloud/Zotero/Herbst et al_2021_DFTK.pdf} +} + +@unpublished{herbstSurrogateModelsQuantum2021, + title = {Surrogate Models for Quantum Spin Systems Based on Reduced Order Modeling}, + author = {Herbst, Michael F. and Wessel, Stefan and Rizzi, Matteo and Stamm, Benjamin}, + date = {2021-11-24}, + eprint = {2110.15665}, + eprinttype = {arxiv}, + primaryclass = {cond-mat, physics:quant-ph}, + url = {http://arxiv.org/abs/2110.15665}, + urldate = {2022-01-11}, + abstract = {We present a methodology to investigate phase-diagrams of quantum models based on the principle of the reduced basis method (RBM). The RBM is built from a few ground-state snapshots, i.e., lowest eigenvectors of the full system Hamiltonian computed at well-chosen points in the parameter space of interest. We put forward a greedy-strategy to assemble such small-dimensional basis, i.e., to select where to spend the numerical effort needed for the snapshots. Once the RBM is assembled, physical observables required for mapping out the phase-diagram (e.g., structure factors) can be computed for any parameter value with a modest computational complexity, considerably lower than the one associated to the underlying Hilbert space dimension. We benchmark the method in two test cases, a chain of excited Rydberg atoms and a geometrically frustrated antiferromagnetic two-dimensional lattice model, and illustrate the accuracy of the approach. In particular, we find that the ground-manifold can be approximated to sufficient accuracy with a moderate number of basis functions, which increases very mildly when the number of microscopic constituents grows - in stark contrast to the exponential growth of the Hilbert space needed to describe each of the few snapshots. A combination of the presented RBM approach with other numerical techniques circumventing even the latter big cost, e.g., Tensor Network methods, is a tantalising outlook of this work.}, + archiveprefix = {arXiv}, + keywords = {Condensed Matter - Strongly Correlated Electrons,Quantum Physics}, + file = {/home/johannes/Nextcloud/Zotero/Herbst et al_2021_Surrogate models for quantum spin systems based on reduced order modeling.pdf;/home/johannes/Zotero/storage/F4FW6AHT/2110.html} +} + +@article{hermannDeepneuralnetworkSolutionElectronic2020, + title = {Deep-Neural-Network Solution of the Electronic {{Schrödinger}} Equation}, + author = {Hermann, Jan and Schätzle, Zeno and Noé, Frank}, + date = {2020-10}, + journaltitle = {Nature Chemistry}, + shortjournal = {Nat. Chem.}, + volume = {12}, + number = {10}, + pages = {891--897}, + publisher = {{Nature Publishing Group}}, + issn = {1755-4349}, + doi = {10.1038/s41557-020-0544-y}, + url = {https://www.nature.com/articles/s41557-020-0544-y}, + urldate = {2022-03-28}, + abstract = {The electronic Schrödinger equation can only be solved analytically for the hydrogen atom, and the numerically exact full configuration-interaction method is exponentially expensive in the number of electrons. Quantum Monte Carlo methods are a possible way out: they scale well for large molecules, they can be parallelized and their accuracy has, as yet, been only limited by the flexibility of the wavefunction ansatz used. Here we propose PauliNet, a deep-learning wavefunction ansatz that achieves nearly exact solutions of the electronic Schrödinger equation for molecules with up to 30 electrons. PauliNet has a multireference Hartree–Fock solution built in as a baseline, incorporates the physics of valid wavefunctions and is trained using variational quantum Monte Carlo. PauliNet outperforms previous state-of-the-art variational ansatzes for atoms, diatomic molecules and a strongly correlated linear H10, and matches the accuracy of highly specialized quantum chemistry methods on the transition-state energy of cyclobutadiene, while being computationally efficient.}, + issue = {10}, + langid = {english}, + keywords = {Computational chemistry,Method development,Physical chemistry,Quantum chemistry,Theoretical chemistry}, + file = {/home/johannes/Nextcloud/Zotero/Hermann et al_2020_Deep-neural-network solution of the electronic Schrödinger equation.pdf;/home/johannes/Zotero/storage/V947YTSM/s41557-020-0544-y.html} +} + +@article{herrCompressingPhysicsAutoencoder2019, + title = {Compressing Physics with an Autoencoder: {{Creating}} an Atomic Species Representation to Improve Machine Learning Models in the Chemical Sciences}, + shorttitle = {Compressing Physics with an Autoencoder}, + author = {Herr, John E. and Koh, Kevin and Yao, Kun and Parkhill, John}, + date = {2019-08-28}, + journaltitle = {The Journal of Chemical Physics}, + shortjournal = {J. Chem. Phys.}, + volume = {151}, + number = {8}, + pages = {084103}, + publisher = {{American Institute of Physics}}, + issn = {0021-9606}, + doi = {10.1063/1.5108803}, + url = {https://aip.scitation.org/doi/full/10.1063/1.5108803}, + urldate = {2022-01-02}, + file = {/home/johannes/Nextcloud/Zotero/Herr et al_2019_Compressing physics with an autoencoder.pdf} +} + +@article{himanenDScribeLibraryDescriptors2020, + title = {{{DScribe}}: {{Library}} of Descriptors for Machine Learning in Materials Science}, + shorttitle = {{{DScribe}}}, + author = {Himanen, Lauri and Jäger, Marc O. J. and Morooka, Eiaki V. and Federici Canova, Filippo and Ranawat, Yashasvi S. and Gao, David Z. and Rinke, Patrick and Foster, Adam S.}, + date = {2020-02-01}, + journaltitle = {Computer Physics Communications}, + shortjournal = {Computer Physics Communications}, + volume = {247}, + pages = {106949}, + issn = {0010-4655}, + doi = {10.1016/j.cpc.2019.106949}, + url = {https://www.sciencedirect.com/science/article/pii/S0010465519303042}, + urldate = {2021-05-13}, + abstract = {DScribe is a software package for machine learning that provides popular feature transformations (“descriptorsâ€) for atomistic materials simulations. DScribe accelerates the application of machine learning for atomistic property prediction by providing user-friendly, off-the-shelf descriptor implementations. The package currently contains implementations for Coulomb matrix, Ewald sum matrix, sine matrix, Many-body Tensor Representation (MBTR), Atom-centered Symmetry Function (ACSF) and Smooth Overlap of Atomic Positions (SOAP). Usage of the package is illustrated for two different applications: formation energy prediction for solids and ionic charge prediction for atoms in organic molecules. The package is freely available under the open-source Apache License 2.0. Program summary Program Title: DScribe Program Files doi: http://dx.doi.org/10.17632/vzrs8n8pk6.1 Licensing provisions: Apache-2.0 Programming language: Python/C/C++ Supplementary material: Supplementary Information as PDF Nature of problem: The application of machine learning for materials science is hindered by the lack of consistent software implementations for feature transformations. These feature transformations, also called descriptors, are a key step in building machine learning models for property prediction in materials science. Solution method: We have developed a library for creating common descriptors used in machine learning applied to materials science. We provide an implementation the following descriptors: Coulomb matrix, Ewald sum matrix, sine matrix, Many-body Tensor Representation (MBTR), Atom-centered Symmetry Functions (ACSF) and Smooth Overlap of Atomic Positions (SOAP). The library has a python interface with computationally intensive routines written in C or C++. The source code, tutorials and documentation are provided online. A continuous integration mechanism is set up to automatically run a series of regression tests and check code coverage when the codebase is updated.}, + langid = {english}, + keywords = {ACSF,descriptors,DScribe,library,materials,Matrix descriptors,MBTR,ML,Open source,Python,rec-by-ruess,SOAP,with-code}, + file = {/home/johannes/Nextcloud/Zotero/Himanen et al_2020_DScribe_Supplementary.pdf;/home/johannes/Nextcloud/Zotero/Himanen et al_2020_DScribe.pdf;/home/johannes/Zotero/storage/IZ66FYIN/S0010465519303042.html} +} + +@book{himanenMaterialsInformaticsAugmenting2020, + title = {Materials {{Informatics}} - {{Augmenting Materials Research}} with {{Data-driven Design}} and {{Machine Learning}}}, + author = {Himanen, Lauri}, + date = {2020}, + publisher = {{Aalto University}}, + issn = {1799-4942 (electronic)}, + url = {https://aaltodoc.aalto.fi:443/handle/123456789/43027}, + urldate = {2021-05-13}, + abstract = {Materials science is the systematic study and development of materials and their properties. Materials informatics and data-driven materials science are umbrella terms for the scientific practice of systematically extracting knowledge from data produced in materials science. This practice differs from traditional scientific approaches in materials research by the volume of processed data and the more automated way information is extracted. This data-driven approach — sometimes referred to as the 4th paradigm of science — is largely driven by the use of modern hardware and software for data production and storage, the Open Science movement and the methodological developments in data mining and machine learning. This dissertation reviews how materials informatics can be effectively applied to accelerate materials science, focusing on computational, atomistic materials modelling. The topic is divided into two different areas: how the data-driven design and tools are being used to re-imagine the life-cycle of materials data and how machine learning, in particular, can be used to complement existing research methodologies in materials science. These topics are explored by investigating the historical development of materials informatics and by highlighting the modern tools and techniques. This discussion provides a guide for anyone interested in deploying these methods in their research and also covers some of the key challenges that the field of materials informatics still faces. After this overview, the original materials informatics research performed during the studies is summarized. First, the open-source software libraries developed for materials informatics are introduced. These libraries deal specifically with tasks related to the automated structural classification of complex atomistic geometries and the efficient description of materials for machine learning. Next, the studies related to materials discovery using data mining and machine learning are discussed. The first study leverages materials databases in the search for optimal coating materials for perovskite-based photovoltaics while the second study focuses on using machine learning for identifying catalytically active sites on nanoclusters.}, + isbn = {978-952-60-8951-5}, + langid = {english}, + keywords = {DScribe,ML,review}, + annotation = {Accepted: 2020-02-11T10:01:06Z}, + file = {/home/johannes/Nextcloud/Zotero/Himanen_2020_Materials Informatics - Augmenting Materials Research with Data-driven Design.pdf;/home/johannes/Zotero/storage/5N3DHF4G/43027.html} +} + +@article{hirohataReviewSpintronicsPrinciples2020, + title = {Review on Spintronics: {{Principles}} and Device Applications}, + shorttitle = {Review on Spintronics}, + author = {Hirohata, Atsufumi and Yamada, Keisuke and Nakatani, Yoshinobu and Prejbeanu, Ioan-Lucian and Diény, Bernard and Pirro, Philipp and Hillebrands, Burkard}, + date = {2020-09-01}, + journaltitle = {Journal of Magnetism and Magnetic Materials}, + shortjournal = {Journal of Magnetism and Magnetic Materials}, + volume = {509}, + pages = {166711}, + issn = {0304-8853}, + doi = {10.1016/j.jmmm.2020.166711}, + url = {https://www.sciencedirect.com/science/article/pii/S0304885320302353}, + urldate = {2022-05-09}, + abstract = {Spintronics is one of the emerging fields for the next-generation nanoelectronic devices to reduce their power consumption and to increase their memory and processing capabilities. Such devices utilise the spin degree of freedom of electrons and/or holes, which can also interact with their orbital moments. In these devices, the spin polarisation is controlled either by magnetic layers used as spin-polarisers or analysers or via spin–orbit coupling. Spin waves can also be used to carry spin current. In this review, the fundamental physics of these phenomena is described first with respect to the spin generation methods as detailed in Sections 2~\textasciitilde ~9. The recent development in their device applications then follows in Sections 10 and 11. Future perspectives are provided at the end.}, + langid = {english}, + keywords = {Dzyaloshinskii–Moriya interaction,Electric field,Electromagnetic wave,Hard disk drive,Landau-Lifshits-Gilbert equation,Magnetic damping,Magnetic random access memory,Magnetic sensor,Magnetic skyrmion,Neuromorphic,Racetrack memory,Spin Hall effects,Spin Nernst effect,Spin Seebeck effect,Spin-current generation,Spin-orbit effects,Spin-transfer torque,Spintronics}, + file = {/home/johannes/Nextcloud/Zotero/Hirohata et al_2020_Review on spintronics.pdf} +} + +@article{hohenbergInhomogeneousElectronGas1964, + title = {Inhomogeneous {{Electron Gas}}}, + author = {Hohenberg, P.}, + date = {1964}, + journaltitle = {Physical Review}, + shortjournal = {Phys. Rev.}, + volume = {136}, + pages = {B864-B871}, + doi = {10.1103/PhysRev.136.B864}, + issue = {3B}, + keywords = {DFT,HKT,original publication}, + file = {/home/johannes/Nextcloud/Zotero/Hohenberg_1964_Inhomogeneous Electron Gas.pdf;/home/johannes/Zotero/storage/BRS4FL49/PhysRev.136.html} +} + +@article{holecAtomisticModelingBasedDesign2017, + title = {Atomistic {{Modeling-Based Design}} of {{Novel Materials}}}, + author = {Holec, David and Zhou, Liangcai and Riedl, Helmut and Koller, Christian M. and Mayrhofer, Paul H. and Friák, Martin and Å ob, MojmÃr and Körmann, Fritz and Neugebauer, Jörg and Music, Denis and Hartmann, Markus A. and Fischer, Franz D.}, + date = {2017}, + journaltitle = {Advanced Engineering Materials}, + volume = {19}, + number = {4}, + pages = {1600688}, + issn = {1527-2648}, + doi = {10.1002/adem.201600688}, + url = {https://onlinelibrary.wiley.com/doi/abs/10.1002/adem.201600688}, + urldate = {2022-05-13}, + abstract = {Modern materials science increasingly advances via a knowledge-based development rather than a trial-and-error procedure. Gathering large amounts of data and getting deep understanding of non-trivial relationships between synthesis of materials, their structure and properties is experimentally a tedious work. Here, theoretical modeling plays a vital role. In this review paper we briefly introduce modeling approaches employed in materials science, their principles and fields of application. We then focus on atomistic modeling methods, mostly quantum mechanical ones but also Monte Carlo and classical molecular dynamics, to demonstrate their practical use on selected examples.}, + langid = {english}, + keywords = {materials,mechanical,MPI Eisenforschung,multiscale}, + annotation = {\_eprint: https://onlinelibrary.wiley.com/doi/pdf/10.1002/adem.201600688}, + file = {/home/johannes/Nextcloud/Zotero/Holec et al_2017_Atomistic Modeling-Based Design of Novel Materials.pdf;/home/johannes/Zotero/storage/NHGDI7UK/adem.html} +} + +@article{hollingsworthCanExactConditions2018, + title = {Can Exact Conditions Improve Machine-Learned Density Functionals?}, + author = {Hollingsworth, Jacob and Li, Li (æŽåŠ›) and Baker, Thomas E. and Burke, Kieron}, + date = {2018-06-28}, + journaltitle = {The Journal of Chemical Physics}, + shortjournal = {J. Chem. Phys.}, + volume = {148}, + number = {24}, + pages = {241743}, + publisher = {{American Institute of Physics}}, + issn = {0021-9606}, + doi = {10.1063/1.5025668}, + url = {https://aip.scitation.org/doi/full/10.1063/1.5025668}, + urldate = {2022-07-08}, + abstract = {Historical methods of functional development in density functional theory have often been guided by analytic conditions that constrain the exact functional one is trying to approximate. Recently, machine-learned functionals have been created by interpolating the results from a small number of exactly solved systems to unsolved systems that are similar in nature. For a simple one-dimensional system, using an exact condition, we find improvements in the learning curves of a machine learning approximation to the non-interacting kinetic energy functional. We also find that the significance of the improvement depends on the nature of the interpolation manifold of the machine-learned functional.}, + keywords = {DFA,DFT,KRR,ML,ML-DFA,ML-DFT,ML-ESM}, + file = {/home/johannes/Nextcloud/Zotero/Hollingsworth et al_2018_Can exact conditions improve machine-learned density functionals.pdf} +} + +@book{hollPhysicsbasedDeepLearning2021, + title = {Physics-Based {{Deep Learning}}}, + author = {Holl, Philipp and Mueller, Maximilian and Schnell, Patrick and Trost, Felix and Thuerey, Nils and Um, Kiwon}, + date = {2021}, + publisher = {{WWW}}, + url = {https://physicsbaseddeeplearning.org}, + urldate = {2022-10-02}, + abstract = {Welcome to the Physics-based Deep Learning Book (v0.2) 👋 TL;DR: This document contains a practical and comprehensive introduction of everything related to deep learning in the context of physical simulations. As much as possible, all topics come with hands-on code examples in the form of Jupyter notebooks to quickly get started. Beyond standard supervised learning from data, we’ll look at physical loss constraints, more tightly coupled learning algorithms with differentiable simulations, training algorithms tailored to physics problems, as well as reinforcement learning and uncertainty modeling. We live in exciting times: these methods have a huge potential to fundamentally change what computer simulations can achieve.}, + keywords = {autodiff,basics,book,Deep learning,learning material,ML,online book,physics-based deep learning,physics-informed ML,PINN}, + file = {/home/johannes/Zotero/storage/Q7HV8D2L/intro.html} +} + +@article{hongReducingTimeDiscovery2021, + title = {Reducing {{Time}} to {{Discovery}}: {{Materials}} and {{Molecular Modeling}}, {{Imaging}}, {{Informatics}}, and {{Integration}}}, + shorttitle = {Reducing {{Time}} to {{Discovery}}}, + author = {Hong, Seungbum and Liow, Chi Hao and Yuk, Jong Min and Byon, Hye Ryung and Yang, Yongsoo and Cho, EunAe and Yeom, Jiwon and Park, Gun and Kang, Hyeonmuk and Kim, Seunggu and Shim, Yoonsu and Na, Moony and Jeong, Chaehwa and Hwang, Gyuseong and Kim, Hongjun and Kim, Hoon and Eom, Seongmun and Cho, Seongwoo and Jun, Hosun and Lee, Yongju and Baucour, Arthur and Bang, Kihoon and Kim, Myungjoon and Yun, Seokjung and Ryu, Jeongjae and Han, Youngjoon and Jetybayeva, Albina and Choi, Pyuck-Pa and Agar, Joshua C. and Kalinin, Sergei V. and Voorhees, Peter W. and Littlewood, Peter and Lee, Hyuck Mo}, + date = {2021-03-23}, + journaltitle = {ACS nano}, + shortjournal = {ACS Nano}, + volume = {15}, + number = {3}, + eprint = {33577296}, + eprinttype = {pmid}, + pages = {3971--3995}, + issn = {1936-086X}, + doi = {10.1021/acsnano.1c00211}, + abstract = {Multiscale and multimodal imaging of material structures and properties provides solid ground on which materials theory and design can flourish. Recently, KAIST announced 10 flagship research fields, which include KAIST Materials Revolution: Materials and Molecular Modeling, Imaging, Informatics and Integration (M3I3). The M3I3 initiative aims to reduce the time for the discovery, design and development of materials based on elucidating multiscale processing-structure-property relationship and materials hierarchy, which are to be quantified and understood through a combination of machine learning and scientific insights. In this review, we begin by introducing recent progress on related initiatives around the globe, such as the Materials Genome Initiative (U.S.), Materials Informatics (U.S.), the Materials Project (U.S.), the Open Quantum Materials Database (U.S.), Materials Research by Information Integration Initiative (Japan), Novel Materials Discovery (E.U.), the NOMAD repository (E.U.), Materials Scientific Data Sharing Network (China), Vom Materials Zur Innovation (Germany), and Creative Materials Discovery (Korea), and discuss the role of multiscale materials and molecular imaging combined with machine learning in realizing the vision of M3I3. Specifically, microscopies using photons, electrons, and physical probes will be revisited with a focus on the multiscale structural hierarchy, as well as structure-property relationships. Additionally, data mining from the literature combined with machine learning will be shown to be more efficient in finding the future direction of materials structures with improved properties than the classical approach. Examples of materials for applications in energy and information will be reviewed and discussed. A case study on the development of a Ni-Co-Mn cathode materials illustrates M3I3's approach to creating libraries of multiscale structure-property-processing relationships. We end with a future outlook toward recent developments in the field of M3I3.}, + langid = {english}, + keywords = {KAIST,Li-ion battery,M3I3,machine learning,materials and molecular modeling,materials imaging,materials informatics,materials integration}, + file = {/home/johannes/Nextcloud/Zotero/Hong et al_2021_Reducing Time to Discovery.pdf} +} + +@article{huberAiiDAScalableComputational2020, + title = {{{AiiDA}} 1.0, a Scalable Computational Infrastructure for Automated Reproducible Workflows and Data Provenance}, + author = {Huber, Sebastiaan P. and Zoupanos, Spyros and Uhrin, Martin and Talirz, Leopold and Kahle, Leonid and Häuselmann, Rico and Gresch, Dominik and Müller, Tiziano and Yakutovich, Aliaksandr V. and Andersen, Casper W. and Ramirez, Francisco F. and Adorf, Carl S. and Gargiulo, Fernando and Kumbhar, Snehal and Passaro, Elsa and Johnston, Conrad and Merkys, Andrius and Cepellotti, Andrea and Mounet, Nicolas and Marzari, Nicola and Kozinsky, Boris and Pizzi, Giovanni}, + date = {2020-09-08}, + journaltitle = {Scientific Data}, + shortjournal = {Sci Data}, + volume = {7}, + number = {1}, + pages = {300}, + publisher = {{Nature Publishing Group}}, + issn = {2052-4463}, + doi = {10.1038/s41597-020-00638-4}, + url = {https://www.nature.com/articles/s41597-020-00638-4}, + urldate = {2021-06-29}, + abstract = {The ever-growing availability of computing power and the sustained development of advanced computational methods have contributed much to recent scientific progress. These developments present new challenges driven by the sheer amount of calculations and data to manage. Next-generation exascale supercomputers will harden these challenges, such that automated and scalable solutions become crucial. In recent years, we have been developing AiiDA (aiida.net), a robust open-source high-throughput infrastructure addressing the challenges arising from the needs of automated workflow management and data provenance recording. Here, we introduce developments and capabilities required to reach sustained performance, with AiiDA supporting throughputs of tens of thousands processes/hour, while automatically preserving and storing the full data provenance in a relational database making it queryable and traversable, thus enabling high-performance data analytics. AiiDA’s workflow language provides advanced automation, error handling features and a flexible plugin model to allow interfacing with external simulation software. The associated plugin registry enables seamless sharing of extensions, empowering a vibrant user community dedicated to making simulations more robust, user-friendly and reproducible.}, + issue = {1}, + langid = {english}, + keywords = {AiiDA,original publication}, + annotation = {Bandiera\_abtest: a Cc\_license\_type: cc\_by Cg\_type: Nature Research Journals Primary\_atype: Research Subject\_term: Computational methods;Research management Subject\_term\_id: computational-methods;research-management}, + file = {/home/johannes/Nextcloud/Zotero/Huber et al_2020_AiiDA 1.pdf;/home/johannes/Zotero/storage/SQ25VE8T/s41597-020-00638-4.html} +} + +@unpublished{huberCommonWorkflowsComputing2021, + title = {Common Workflows for Computing Material Properties Using Different Quantum Engines}, + author = {Huber, Sebastiaan P. and Bosoni, Emanuele and Bercx, Marnik and Bröder, Jens and Degomme, Augustin and Dikan, Vladimir and Eimre, Kristjan and Flage-Larsen, Espen and Garcia, Alberto and Genovese, Luigi and Gresch, Dominik and Johnston, Conrad and Petretto, Guido and Poncé, Samuel and Rignanese, Gian-Marco and Sewell, Christopher J. and Smit, Berend and Tseplyaev, Vasily and Uhrin, Martin and Wortmann, Daniel and Yakutovich, Aliaksandr V. and Zadoks, Austin and Zarabadi-Poor, Pezhman and Zhu, Bonan and Marzari, Nicola and Pizzi, Giovanni}, + date = {2021-05-11}, + eprint = {2105.05063}, + eprinttype = {arxiv}, + primaryclass = {cond-mat}, + url = {http://arxiv.org/abs/2105.05063}, + urldate = {2021-06-23}, + abstract = {The prediction of material properties through electronic-structure simulations based on density-functional theory has become routinely common, thanks, in part, to the steady increase in the number and robustness of available simulation packages. This plurality of codes and methods aiming to solve similar problems is both a boon and a burden. While providing great opportunities for cross-verification, these packages adopt different methods, algorithms, and paradigms, making it challenging to choose, master, and efficiently use any one for a given task. Leveraging recent advances in managing reproducible scientific workflows, we demonstrate how developing common interfaces for workflows that automatically compute material properties can tackle the challenge mentioned above, greatly simplifying interoperability and cross-verification. We introduce design rules for reproducible and reusable code-agnostic workflow interfaces to compute well-defined material properties, which we implement for eleven different quantum engines and use to compute three different material properties. Each implementation encodes carefully selected simulation parameters and workflow logic, making the implementer's expertise of the quantum engine directly available to non-experts. Full provenance and reproducibility of the workflows is guaranteed through the use of the AiiDA infrastructure. All workflows are made available as open-source and come pre-installed with the Quantum Mobile virtual machine, making their use straightforward.}, + archiveprefix = {arXiv}, + keywords = {AiiDA,workflows}, + file = {/home/johannes/Nextcloud/Zotero/Huber et al_2021_Common workflows for computing material properties using different quantum.pdf;/home/johannes/Zotero/storage/9AJRBXBR/2105.html} +} + +@unpublished{huoUnifiedRepresentationMolecules2018, + title = {Unified {{Representation}} of {{Molecules}} and {{Crystals}} for {{Machine Learning}}}, + author = {Huo, Haoyan and Rupp, Matthias}, + date = {2018-01-02}, + eprint = {1704.06439}, + eprinttype = {arxiv}, + primaryclass = {cond-mat, physics:physics}, + url = {http://arxiv.org/abs/1704.06439}, + urldate = {2021-06-29}, + abstract = {Accurate simulations of atomistic systems from first principles are limited by computational cost. In high-throughput settings, machine learning can potentially reduce these costs significantly by accurately interpolating between reference calculations. For this, kernel learning approaches crucially require a single Hilbert space accommodating arbitrary atomistic systems. We introduce a many-body tensor representation that is invariant to translations, rotations and nuclear permutations of same elements, unique, differentiable, can represent molecules and crystals, and is fast to compute. Empirical evidence is presented for energy prediction errors below 1 kcal/mol for 7k organic molecules and 5 meV/atom for 11k elpasolite crystals. Applicability is demonstrated for phase diagrams of Pt-group/transition-metal binary systems.}, + archiveprefix = {arXiv}, + keywords = {descriptors,MBTR,ML,original publication}, + file = {/home/johannes/Nextcloud/Zotero/Huo_Rupp_2018_Unified Representation of Molecules and Crystals for Machine Learning.pdf;/home/johannes/Zotero/storage/EZJ986KS/1704.html} +} + +@article{hutsonArtificialIntelligenceFaces2018, + title = {Artificial Intelligence Faces Reproducibility Crisis}, + author = {Hutson, Matthew}, + date = {2018-02-16}, + journaltitle = {Science (New York, N.Y.)}, + shortjournal = {Science}, + volume = {359}, + number = {6377}, + eprint = {29449469}, + eprinttype = {pmid}, + pages = {725--726}, + issn = {1095-9203}, + doi = {10.1126/science.359.6377.725}, + langid = {english}, + keywords = {Artificial Intelligence,Reproducibility of Results} +} + +@inproceedings{idreosOverviewDataExploration2015, + title = {Overview of {{Data Exploration Techniques}}}, + booktitle = {Proceedings of the 2015 {{ACM SIGMOD International Conference}} on {{Management}} of {{Data}}}, + author = {Idreos, Stratos and Papaemmanouil, Olga and Chaudhuri, Surajit}, + date = {2015-05-27}, + series = {{{SIGMOD}} '15}, + pages = {277--281}, + publisher = {{Association for Computing Machinery}}, + location = {{New York, NY, USA}}, + doi = {10.1145/2723372.2731084}, + url = {https://doi.org/10.1145/2723372.2731084}, + urldate = {2022-10-02}, + abstract = {Data exploration is about efficiently extracting knowledge from data even if we do not know exactly what we are looking for. In this tutorial, we survey recent developments in the emerging area of database systems tailored for data exploration. We discuss new ideas on how to store and access data as well as new ideas on how to interact with a data system to enable users and applications to quickly figure out which data parts are of interest. In addition, we discuss how to exploit lessons-learned from past research, the new challenges data exploration crafts, emerging applications and future research directions.}, + isbn = {978-1-4503-2758-9}, + keywords = {data exploration,Database,unsupervised learning,visualization}, + file = {/home/johannes/Nextcloud/Zotero/Idreos et al_2015_Overview of Data Exploration Techniques.pdf} +} + +@article{imbalzanoAutomaticSelectionAtomic2018, + title = {Automatic Selection of Atomic Fingerprints and Reference Configurations for Machine-Learning Potentials}, + author = {Imbalzano, Giulio and Anelli, Andrea and Giofré, Daniele and Klees, Sinja and Behler, Jörg and Ceriotti, Michele}, + date = {2018-04-30}, + journaltitle = {The Journal of Chemical Physics}, + shortjournal = {J. Chem. Phys.}, + volume = {148}, + number = {24}, + pages = {241730}, + publisher = {{American Institute of Physics}}, + issn = {0021-9606}, + doi = {10.1063/1.5024611}, + url = {https://aip.scitation.org/doi/10.1063/1.5024611}, + urldate = {2021-05-18}, + abstract = {Machine learning of atomic-scale properties is revolutionizing molecular modeling, making it possible to evaluate inter-atomic potentials with first-principles accuracy, at a fraction of the costs. The accuracy, speed, and reliability of machine learning potentials, however, depend strongly on the way atomic configurations are represented, i.e., the choice of descriptors used as input for the machine learning method. The raw Cartesian coordinates are typically transformed in “fingerprints,†or “symmetry functions,†that are designed to encode, in addition to the structure, important properties of the potential energy surface like its invariances with respect to rotation, translation, and permutation of like atoms. Here we discuss automatic protocols to select a number of fingerprints out of a large pool of candidates, based on the correlations that are intrinsic to the training data. This procedure can greatly simplify the construction of neural network potentials that strike the best balance between accuracy and computational efficiency and has the potential to accelerate by orders of magnitude the evaluation of Gaussian approximation potentials based on the smooth overlap of atomic positions kernel. We present applications to the construction of neural network potentials for water and for an Al–Mg–Si alloy and to the prediction of the formation energies of small organic molecules using Gaussian process regression.}, + keywords = {ACSF,autoML,descriptors,GAP,ML,MLP}, + file = {/home/johannes/Nextcloud/Zotero/Imbalzano et al_2018_Automatic selection of atomic fingerprints and reference configurations for.pdf;/home/johannes/Zotero/storage/DBZXGYRI/1.html} +} + +@online{ImprovingDensityFunctional, + title = {Improving {{Density Functional Theory}} with {{Machine Learning}} - {{ProQuest}}}, + url = {https://www.proquest.com/openview/f9f5f9fd53c33fcce033b3403f865e90/1?accountid=15951&cbl=18750&diss=y&parentSessionId=aO8yAFVn7V%2F%2Byvnb4zczB1fFkg9jKvE%2F%2FXvtAdNSJQc%3D&parentSessionId=mU4GnqS%2F4JNlsQ5daD%2Bnp9f3OJ4k9apMVBrSWVCKrbE%3D&pq-origsite=gscholar}, + urldate = {2021-12-14}, + abstract = {Explore millions of resources from scholarly journals, books, newspapers, videos and more, on the ProQuest Platform.}, + langid = {english}, + keywords = {DFT,dissertation,ML,preview}, + file = {/home/johannes/Zotero/storage/WLWNEY4Q/1.html} +} + +@article{jablonkaBigDataSciencePorous2020, + title = {Big-{{Data Science}} in {{Porous Materials}}: {{Materials Genomics}} and {{Machine Learning}}}, + shorttitle = {Big-{{Data Science}} in {{Porous Materials}}}, + author = {Jablonka, Kevin Maik and Ongari, Daniele and Moosavi, Seyed Mohamad and Smit, Berend}, + date = {2020-08-26}, + journaltitle = {Chemical Reviews}, + shortjournal = {Chem. Rev.}, + volume = {120}, + number = {16}, + pages = {8066--8129}, + publisher = {{American Chemical Society}}, + issn = {0009-2665}, + doi = {10.1021/acs.chemrev.0c00004}, + url = {https://doi.org/10.1021/acs.chemrev.0c00004}, + urldate = {2021-05-30}, + abstract = {By combining metal nodes with organic linkers we can potentially synthesize millions of possible metal–organic frameworks (MOFs). The fact that we have so many materials opens many exciting avenues but also create new challenges. We simply have too many materials to be processed using conventional, brute force, methods. In this review, we show that having so many materials allows us to use big-data methods as a powerful technique to study these materials and to discover complex correlations. The first part of the review gives an introduction to the principles of big-data science. We show how to select appropriate training sets, survey approaches that are used to represent these materials in feature space, and review different learning architectures, as well as evaluation and interpretation strategies. In the second part, we review how the different approaches of machine learning have been applied to porous materials. In particular, we discuss applications in the field of gas storage and separation, the stability of these materials, their electronic properties, and their synthesis. Given the increasing interest of the scientific community in machine learning, we expect this list to rapidly expand in the coming years.}, + keywords = {descriptors,materials discovery,ML,MLP,review}, + file = {/home/johannes/Nextcloud/Zotero/Jablonka et al_2020_Big-Data Science in Porous Materials.pdf} +} + +@article{jacobsMaterialsSimulationToolkit2020, + title = {The {{Materials Simulation Toolkit}} for {{Machine}} Learning ({{MAST-ML}}): {{An}} Automated Open Source Toolkit to Accelerate Data-Driven Materials Research}, + shorttitle = {The {{Materials Simulation Toolkit}} for {{Machine}} Learning ({{MAST-ML}})}, + author = {Jacobs, Ryan and Mayeshiba, Tam and Afflerbach, Ben and Miles, Luke and Williams, Max and Turner, Matthew and Finkel, Raphael and Morgan, Dane}, + date = {2020-04-15}, + journaltitle = {Computational Materials Science}, + shortjournal = {Computational Materials Science}, + volume = {176}, + pages = {109544}, + issn = {0927-0256}, + doi = {10.1016/j.commatsci.2020.109544}, + url = {https://www.sciencedirect.com/science/article/pii/S0927025620300355}, + urldate = {2021-06-26}, + abstract = {As data science and machine learning methods are taking on an increasingly important role in the materials research community, there is a need for the development of machine learning software tools that are easy to use (even for nonexperts with no programming ability), provide flexible access to the most important algorithms, and codify best practices of machine learning model development and evaluation. Here, we introduce the Materials Simulation Toolkit for Machine Learning (MAST-ML), an open source Python-based software package designed to broaden and accelerate the use of machine learning in materials science research. MAST-ML provides predefined routines for many input setup, model fitting, and post-analysis tasks, as well as a simple structure for executing a multi-step machine learning model workflow. In this paper, we describe how MAST-ML is used to streamline and accelerate the execution of machine learning problems. We walk through how to acquire and run MAST-ML, demonstrate how to execute different components of a supervised machine learning workflow via a customized input file, and showcase a number of features and analyses conducted automatically during a MAST-ML run. Further, we demonstrate the utility of MAST-ML by showcasing examples of recent materials informatics studies which used MAST-ML to formulate and evaluate various machine learning models for an array of materials applications. Finally, we lay out a vision of how MAST-ML, together with complementary software packages and emerging cyberinfrastructure, can advance the rapidly growing field of materials informatics, with a focus on producing machine learning models easily, reproducibly, and in a manner that facilitates model evolution and improvement in the future.}, + langid = {english}, + keywords = {library,materials informatics,with-code}, + file = {/home/johannes/Nextcloud/Zotero/Jacobs et al_2020_The Materials Simulation Toolkit for Machine learning (MAST-ML).pdf;/home/johannes/Zotero/storage/5R5YHQE4/S0927025620300355.html} +} + +@article{jainCommentaryMaterialsProject2013, + title = {Commentary: {{The Materials Project}}: {{A}} Materials Genome Approach to Accelerating Materials Innovation}, + shorttitle = {Commentary}, + author = {Jain, Anubhav and Ong, Shyue Ping and Hautier, Geoffroy and Chen, Wei and Richards, William Davidson and Dacek, Stephen and Cholia, Shreyas and Gunter, Dan and Skinner, David and Ceder, Gerbrand and Persson, Kristin A.}, + date = {2013-07-01}, + journaltitle = {APL Materials}, + volume = {1}, + number = {1}, + pages = {011002}, + publisher = {{American Institute of Physics}}, + doi = {10.1063/1.4812323}, + url = {https://aip.scitation.org/doi/10.1063%2F1.4812323}, + urldate = {2021-10-15}, + abstract = {Accelerating the discovery of advanced materials is essential for human welfare and sustainable, clean energy. In this paper, we introduce the Materials Project (www.materialsproject.org), a core program of the Materials Genome Initiative that uses high-throughput computing to uncover the properties of all known inorganic materials. This open dataset can be accessed through multiple channels for both interactive exploration and data mining. The Materials Project also seeks to create open-source platforms for developing robust, sophisticated materials analyses. Future efforts will enable users to perform ‘‘rapid-prototyping’’ of new materials in silico, and provide researchers with new avenues for cost-effective, data-driven materials design.}, + keywords = {materials project}, + file = {/home/johannes/Nextcloud/Zotero/Jain et al_2013_Commentary.pdf} +} + +@article{jainFireWorksDynamicWorkflow2015, + title = {{{FireWorks}}: A Dynamic Workflow System Designed for High-Throughput Applications}, + shorttitle = {{{FireWorks}}}, + author = {Jain, Anubhav and Ong, Shyue Ping and Chen, Wei and Medasani, Bharat and Qu, Xiaohui and Kocher, Michael and Brafman, Miriam and Petretto, Guido and Rignanese, Gian-Marco and Hautier, Geoffroy and Gunter, Daniel and Persson, Kristin A.}, + date = {2015}, + journaltitle = {Concurrency and Computation: Practice and Experience}, + volume = {27}, + number = {17}, + pages = {5037--5059}, + issn = {1532-0634}, + doi = {10.1002/cpe.3505}, + url = {https://onlinelibrary.wiley.com/doi/abs/10.1002/cpe.3505}, + urldate = {2021-10-17}, + abstract = {This paper introduces FireWorks, a workflow software for running high-throughput calculation workflows at supercomputing centers. FireWorks has been used to complete over 50 million CPU-hours worth of computational chemistry and materials science calculations at the National Energy Research Supercomputing Center. It has been designed to serve the demanding high-throughput computing needs of these applications, with extensive support for (i) concurrent execution through job packing, (ii) failure detection and correction, (iii) provenance and reporting for long-running projects, (iv) automated duplicate detection, and (v) dynamic workflows (i.e., modifying the workflow graph during runtime). We have found that these features are highly relevant to enabling modern data-driven and high-throughput science applications, and we discuss our implementation strategy that rests on Python and NoSQL databases (MongoDB). Finally, we present performance data and limitations of our approach along with planned future work. Copyright © 2015 John Wiley \& Sons, Ltd.}, + langid = {english}, + keywords = {fault-tolerant computing,high-throughput computing,scientific workflows}, + annotation = {\_eprint: https://onlinelibrary.wiley.com/doi/pdf/10.1002/cpe.3505}, + file = {/home/johannes/Nextcloud/Zotero/Jain et al_2015_FireWorks.pdf;/home/johannes/Zotero/storage/FFWIWLTR/cpe.html} +} + +@article{janssenPyironIntegratedDevelopment2019, + title = {Pyiron: {{An}} Integrated Development Environment for Computational Materials Science}, + shorttitle = {Pyiron}, + author = {Janssen, Jan and Surendralal, Sudarsan and Lysogorskiy, Yury and Todorova, Mira and Hickel, Tilmann and Drautz, Ralf and Neugebauer, Jörg}, + date = {2019-06-01}, + journaltitle = {Computational Materials Science}, + shortjournal = {Computational Materials Science}, + volume = {163}, + pages = {24--36}, + issn = {0927-0256}, + doi = {10.1016/j.commatsci.2018.07.043}, + url = {https://www.sciencedirect.com/science/article/pii/S0927025618304786}, + urldate = {2021-10-17}, + abstract = {To support and accelerate the development of simulation protocols in atomistic modelling, we introduce an integrated development environment (IDE) for computational materials science called pyiron (http://pyiron.org). The pyiron IDE combines a web based source code editor, a job management system for build automation, and a hierarchical data management solution. The core components of the pyiron IDE are pyiron objects based on an abstract class, which links application structures such as atomistic structures, projects, jobs, simulation protocols and computing resources with persistent storage and an interactive user environment. The simulation protocols within the pyiron IDE are constructed using the Python programming language. To highlight key concepts of this tool as well as to demonstrate its ability to simplify the implementation and testing of simulation protocols we discuss two applications. In these examples we show how pyiron supports the whole life cycle of a typical simulation, seamlessly combines ab initio with empirical potential calculations, and how complex feedback loops can be implemented. While originally developed with focus on ab initio thermodynamics simulations, the concepts and implementation of pyiron are general thus allowing to employ it for a wide range of simulation topics.}, + langid = {english}, + keywords = {Complex simulation protocols,Integrated development environment,Modelling workflow}, + file = {/home/johannes/Zotero/storage/TNV7XY35/S0927025618304786.html} +} + +@inproceedings{jiaPushingLimitMolecular2020, + title = {Pushing the {{Limit}} of {{Molecular Dynamics}} with {{Ab Initio Accuracy}} to 100 {{Million Atoms}} with {{Machine Learning}}}, + booktitle = {{{SC20}}: {{International Conference}} for {{High Performance Computing}}, {{Networking}}, {{Storage}} and {{Analysis}}}, + author = {Jia, Weile and Wang, Han and Chen, Mohan and Lu, Denghui and Lin, Lin and Car, Roberto and Weinan, E and Zhang, Linfeng}, + date = {2020-11}, + pages = {1--14}, + doi = {10.1109/SC41405.2020.00009}, + abstract = {For 35 years, ab initio molecular dynamics (AIMD) has been the method of choice for modeling complex atomistic phenomena from first principles. However, most AIMD applications are limited by computational cost to systems with thousands of atoms at most. We report that a machine learning based simulation protocol (Deep Potential Molecular Dynamics), while retaining ab initio accuracy, can simulate more than 1 nanosecond-long trajectory of over 100 million atoms per day, using a highly optimized code (GPU DeePMD-kit) on the Summit supercomputer. Our code can efficiently scale up to the entire Summit supercomputer, attaining 91 PFLOPS in double precision (45.5\% of the peak) and 162/275 PFLOPS in mixed-single/half precision. The great accomplishment of this work is that it opens the door to simulating unprecedented size and time scales with ab initio accuracy. It also poses new challenges to the next-generation supercomputer for a better integration of machine learning and physical modeling.}, + eventtitle = {{{SC20}}: {{International Conference}} for {{High Performance Computing}}, {{Networking}}, {{Storage}} and {{Analysis}}}, + keywords = {100 million atoms,DeePMD-kit,MD,ML,MLP,record,Supercomputer}, + file = {/home/johannes/Nextcloud/Zotero/Jia et al_2020_Pushing the Limit of Molecular Dynamics with Ab Initio Accuracy to 100 Million.pdf;/home/johannes/Zotero/storage/UML425XW/9355242.html} +} + +@book{johanssonNumericalPythonScientific2019, + title = {Numerical {{Python}}: {{Scientific Computing}} and {{Data Science Applications}} with {{Numpy}}, {{SciPy}} and {{Matplotlib}}}, + shorttitle = {Numerical {{Python}}}, + author = {Johansson, Robert}, + date = {2019}, + publisher = {{Apress}}, + location = {{Berkeley, CA}}, + doi = {10.1007/978-1-4842-4246-9}, + url = {http://link.springer.com/10.1007/978-1-4842-4246-9}, + urldate = {2021-05-04}, + isbn = {978-1-4842-4245-2 978-1-4842-4246-9}, + langid = {english}, + keywords = {computational science,data science,general,numerical,practice,python,scientific computing}, + file = {/home/johannes/Books/data_science/_general/practice/Johansson_NumericalPython_2e-2019.epub} +} + +@book{johnvonneumann-institutfurcomputingComputationalNanoscienceIt2006, + title = {Computational Nanoscience: Do It Yourself! : {{Winter School}}, 14-22 {{February}} 2006, {{Forschungszentrum Julich}}, {{Germany}} : Lecture Notes}, + shorttitle = {Computational Nanoscience}, + editor = {{John von Neumann-Institut für Computing} and {Winter School} and Grotendorst, Johannes and Blügel, Stefan and Marx, Dominik and {John von Neumann-Institut für Computing}}, + date = {2006}, + publisher = {{John von Neumann Institut for Computing}}, + location = {{Julich, Germany}}, + url = {http://hdl.handle.net/2128/2943}, + editora = {Mavropoulos, Phivos and Zeller, Rudolf and Lounis, Samir and Dederichs, Peter H.}, + editoratype = {collaborator}, + isbn = {978-3-00-017350-9}, + langid = {english}, + keywords = {FZJ,KKR,NIC,NIC winter school,PGI-1/IAS-1}, + annotation = {OCLC: 77518371}, + file = {/home/johannes/Nextcloud/Zotero/John von Neumann-Institut für Computing et al_2006_Computational nanoscience.pdf} +} + +@misc{jorgensenDeepDFTNeuralMessage2020, + title = {{{DeepDFT}}: {{Neural Message Passing Network}} for {{Accurate Charge Density Prediction}}}, + shorttitle = {{{DeepDFT}}}, + author = {Jørgensen, Peter Bjørn and Bhowmik, Arghya}, + date = {2020-11-04}, + number = {arXiv:2011.03346}, + eprint = {2011.03346}, + eprinttype = {arxiv}, + primaryclass = {physics}, + publisher = {{arXiv}}, + doi = {10.48550/arXiv.2011.03346}, + url = {http://arxiv.org/abs/2011.03346}, + urldate = {2022-07-10}, + abstract = {We introduce DeepDFT, a deep learning model for predicting the electronic charge density around atoms, the fundamental variable in electronic structure simulations from which all ground state properties can be calculated. The model is formulated as neural message passing on a graph, consisting of interacting atom vertices and special query point vertices for which the charge density is predicted. The accuracy and scalability of the model are demonstrated for molecules, solids and liquids. The trained model achieves lower average prediction errors than the observed variations in charge density obtained from density functional theory simulations using different exchange correlation functionals.}, + archiveprefix = {arXiv}, + keywords = {Computer Science - Machine Learning,Physics - Computational Physics}, + file = {/home/johannes/Nextcloud/Zotero/Jørgensen_Bhowmik_2020_DeepDFT.pdf;/home/johannes/Zotero/storage/QXJKV745/2011.html} +} + +@misc{jorgensenGraphNeuralNetworks2021, + title = {Graph Neural Networks for Fast Electron Density Estimation of Molecules, Liquids, and Solids}, + author = {Jørgensen, Peter Bjørn and Bhowmik, Arghya}, + date = {2021-12-01}, + number = {arXiv:2112.00652}, + eprint = {2112.00652}, + eprinttype = {arxiv}, + primaryclass = {physics, stat}, + publisher = {{arXiv}}, + doi = {10.48550/arXiv.2112.00652}, + url = {http://arxiv.org/abs/2112.00652}, + urldate = {2022-07-10}, + abstract = {Electron density \$\textbackslash rho(\textbackslash vec\{r\})\$ is the fundamental variable in the calculation of ground state energy with density functional theory (DFT). Beyond total energy, features in \$\textbackslash rho(\textbackslash vec\{r\})\$ distribution and modifications in \$\textbackslash rho(\textbackslash vec\{r\})\$ are often used to capture critical physicochemical phenomena in functional materials and molecules at the electronic scale. Methods providing access to \$\textbackslash rho(\textbackslash vec\{r\})\$ of complex disordered systems with little computational cost can be a game changer in the expedited exploration of materials phase space towards the inverse design of new materials with better functionalities. We present a machine learning framework for the prediction of \$\textbackslash rho(\textbackslash vec\{r\})\$. The model is based on equivariant graph neural networks and the electron density is predicted at special query point vertices that are part of the message passing graph, but only receive messages. The model is tested across multiple data sets of molecules (QM9), liquid ethylene carbonate electrolyte (EC) and LixNiyMnzCo(1-y-z)O2 lithium ion battery cathodes (NMC). For QM9 molecules, the accuracy of the proposed model exceeds typical variability in \$\textbackslash rho(\textbackslash vec\{r\})\$ obtained from DFT done with different exchange-correlation functional and show beyond the state of the art accuracy. The accuracy is even better for the mixed oxide (NMC) and electrolyte (EC) datasets. The linear scaling model's capacity to probe thousands of points simultaneously permits calculation of \$\textbackslash rho(\textbackslash vec\{r\})\$ for large complex systems many orders of magnitude faster than DFT allowing screening of disordered functional materials.}, + archiveprefix = {arXiv}, + keywords = {Physics - Computational Physics,Statistics - Machine Learning}, + file = {/home/johannes/Nextcloud/Zotero/Jørgensen_Bhowmik_2021_Graph neural networks for fast electron density estimation of molecules,.pdf;/home/johannes/Zotero/storage/MBXG22TT/2112.html} +} + +@online{julichJulichKKRCodes, + title = {The {{Jülich KKR Codes}}}, + shorttitle = {Jülich {{KKR}} Codes}, + author = {Jülich, Forschungszentrum}, + url = {https://jukkr.fz-juelich.de/}, + urldate = {2021-10-17}, + abstract = {The Jülich family of codes implementing the Korringa-Kohn-Rostoker Green Function method for electronic structure calculations}, + langid = {american}, + file = {/home/johannes/Zotero/storage/ZX3TRV7K/jukkr.fz-juelich.de.html} +} + +@article{kalitaHowWellDoes2022, + title = {How {{Well Does Kohn}}–{{Sham Regularizer Work}} for {{Weakly Correlated Systems}}?}, + author = {Kalita, Bhupalee and Pederson, Ryan and Chen, Jielun and Li, Li and Burke, Kieron}, + date = {2022-03-24}, + journaltitle = {The Journal of Physical Chemistry Letters}, + shortjournal = {J. Phys. Chem. Lett.}, + volume = {13}, + number = {11}, + pages = {2540--2547}, + publisher = {{American Chemical Society}}, + doi = {10.1021/acs.jpclett.2c00371}, + url = {https://doi.org/10.1021/acs.jpclett.2c00371}, + urldate = {2022-07-10}, + abstract = {Kohn–Sham regularizer (KSR) is a differentiable machine learning approach to finding the exchange-correlation functional in Kohn–Sham density functional theory that works for strongly correlated systems. Here we test KSR for a weak correlation. We propose spin-adapted KSR (sKSR) with trainable local, semilocal, and nonlocal approximations found by minimizing density and total energy loss. We assess the atoms-to-molecules generalizability by training on one-dimensional (1D) H, He, Li, Be, and Be2+ and testing on 1D hydrogen chains, LiH, BeH2, and helium hydride complexes. The generalization error from our semilocal approximation is comparable to other differentiable approaches, but our nonlocal functional outperforms any existing machine learning functionals, predicting ground-state energies of test systems with a mean absolute error of 2.7 mH.}, + keywords = {DFT,Kohn-Sham regularizer,ML,ML-DFA,ML-DFT,ML-ESM,spin-dependent,spin-polarized}, + file = {/home/johannes/Nextcloud/Zotero/Kalita et al_2022_How Well Does Kohn–Sham Regularizer Work for Weakly Correlated Systems.pdf;/home/johannes/Zotero/storage/TCWGCAZA/acs.jpclett.html} +} + +@article{kalitaLearningApproximateDensity2021, + title = {Learning to {{Approximate Density Functionals}}}, + author = {Kalita, Bhupalee and Li, Li and McCarty, Ryan J. and Burke, Kieron}, + date = {2021-02-16}, + journaltitle = {Accounts of Chemical Research}, + shortjournal = {Acc. Chem. Res.}, + volume = {54}, + number = {4}, + pages = {818--826}, + publisher = {{American Chemical Society}}, + issn = {0001-4842}, + doi = {10.1021/acs.accounts.0c00742}, + url = {https://doi.org/10.1021/acs.accounts.0c00742}, + urldate = {2021-12-14}, + abstract = {ConspectusDensity functional theory (DFT) calculations are used in over 40,000 scientific papers each year, in chemistry, materials science, and far beyond. DFT is extremely useful because it is computationally much less expensive than ab initio electronic structure methods and allows systems of considerably larger size to be treated. However, the accuracy of any Kohn–Sham DFT calculation is limited by the approximation chosen for the exchange-correlation (XC) energy. For more than half a century, humans have developed the art of such approximations, using general principles, empirical data, or a combination of both, typically yielding useful results, but with errors well above the chemical accuracy limit (1 kcal/mol). Over the last 15 years, machine learning (ML) has made major breakthroughs in many applications and is now being applied to electronic structure calculations. This recent rise of ML begs the question: Can ML propose or improve density functional approximations? Success could greatly enhance the accuracy and usefulness of DFT calculations without increasing the cost.In this work, we detail efforts in this direction, beginning with an elementary proof of principle from 2012, namely, finding the kinetic energy of several Fermions in a box using kernel ridge regression. This is an example of orbital-free DFT, for which a successful general-purpose scheme could make even DFT calculations run much faster. We trace the development of that work to state-of-the-art molecular dynamics simulations of resorcinol with chemical accuracy. By training on ab initio examples, one bypasses the need to find the XC functional explicitly. We also discuss how the exchange-correlation energy itself can be modeled with such methods, especially for strongly correlated materials. Finally, we show how deep neural networks with differentiable programming can be used to construct accurate density functionals from very few data points by using the Kohn–Sham equations themselves as a regularizer. All these cases show that ML can create approximations of greater accuracy than humans, and is capable of finding approximations that can deal with difficult cases such as strong correlation. However, such ML-designed functionals have not been implemented in standard codes because of one last great challenge: generalization. We discuss how effortlessly human-designed functionals can be applied to a wide range of situations, and how difficult that is for ML.}, + keywords = {DFT,Kohn-Sham regularizer,ML,ML-DFA,ML-DFT,ML-ESM,OF-DFT,review}, + file = {/home/johannes/Nextcloud/Zotero/Kalita et al_2021_Learning to Approximate Density Functionals.pdf;/home/johannes/Zotero/storage/MUYRWPH9/acs.accounts.html} +} + +@unpublished{kalitaUsingMachineLearning2021, + title = {Using {{Machine Learning}} to {{Find New Density Functionals}}}, + author = {Kalita, Bhupalee and Burke, Kieron}, + date = {2021-12-03}, + eprint = {2112.05554}, + eprinttype = {arxiv}, + primaryclass = {physics}, + url = {http://arxiv.org/abs/2112.05554}, + urldate = {2022-03-28}, + abstract = {Machine learning has now become an integral part of research and innovation. The field of machine learning density functional theory has continuously expanded over the years while making several noticeable advances. We briefly discuss the status of this field and point out some current and future challenges. We also talk about how state-of-the-art science and technology tools can help overcome these challenges. This draft is a part of the "Roadmap on Machine Learning in Electronic Structure" to be published in Electronic Structure (EST).}, + archiveprefix = {arXiv}, + keywords = {DFT,Kohn-Sham regularizer,ML,ML-DFA,ML-DFT,ML-ESM}, + file = {/home/johannes/Nextcloud/Zotero/Kalita_Burke_2021_Using Machine Learning to Find New Density Functionals.pdf;/home/johannes/Zotero/storage/6FMA3TRD/2112.html} +} + +@inproceedings{kanterDeepFeatureSynthesis2015, + title = {Deep Feature Synthesis: {{Towards}} Automating Data Science Endeavors}, + shorttitle = {Deep Feature Synthesis}, + booktitle = {2015 {{IEEE International Conference}} on {{Data Science}} and {{Advanced Analytics}} ({{DSAA}})}, + author = {Kanter, James Max and Veeramachaneni, Kalyan}, + date = {2015-10}, + pages = {1--10}, + doi = {10.1109/DSAA.2015.7344858}, + abstract = {In this paper, we develop the Data Science Machine, which is able to derive predictive models from raw data automatically. To achieve this automation, we first propose and develop the Deep Feature Synthesis algorithm for automatically generating features for relational datasets. The algorithm follows relationships in the data to a base field, and then sequentially applies mathematical functions along that path to create the final feature. Second, we implement a generalizable machine learning pipeline and tune it using a novel Gaussian Copula process based approach. We entered the Data Science Machine in 3 data science competitions that featured 906 other data science teams. Our approach beats 615 teams in these data science competitions. In 2 of the 3 competitions we beat a majority of competitors, and in the third, we achieved 94\% of the best competitor's score. In the best case, with an ongoing competition, we beat 85.6\% of the teams and achieved 95.7\% of the top submissions score.}, + eventtitle = {2015 {{IEEE International Conference}} on {{Data Science}} and {{Advanced Analytics}} ({{DSAA}})}, + keywords = {Data models,Feature extraction,ML,thesis}, + file = {/home/johannes/Nextcloud/Zotero/Kanter_Veeramachaneni_2015_Deep feature synthesis.pdf;/home/johannes/Zotero/storage/FTYFE5ZI/7344858.html} +} + +@article{kasimLearningExchangecorrelationFunctional2021, + title = {Learning the Exchange-Correlation Functional from Nature with Fully Differentiable Density Functional Theory}, + author = {Kasim, Muhammad F. and Vinko, Sam M.}, + date = {2021-09-15}, + journaltitle = {Physical Review Letters}, + shortjournal = {Phys. Rev. Lett.}, + volume = {127}, + number = {12}, + eprint = {2102.04229}, + eprinttype = {arxiv}, + pages = {126403}, + issn = {0031-9007, 1079-7114}, + doi = {10.1103/PhysRevLett.127.126403}, + url = {http://arxiv.org/abs/2102.04229}, + urldate = {2022-01-02}, + abstract = {Improving the predictive capability of molecular properties in ab initio simulations is essential for advanced material discovery. Despite recent progress making use of machine learning, utilizing deep neural networks to improve quantum chemistry modelling remains severely limited by the scarcity and heterogeneity of appropriate experimental data. Here we show how training a neural network to replace the exchange-correlation functional within a fully-differentiable three-dimensional Kohn-Sham density functional theory (DFT) framework can greatly improve simulation accuracy. Using only eight experimental data points on diatomic molecules, our trained exchange-correlation networks enable improved prediction accuracy of atomization energies across a collection of 104 molecules containing new bonds and atoms that are not present in the training dataset.}, + archiveprefix = {arXiv}, + keywords = {Computer Science - Machine Learning,Physics - Chemical Physics,Physics - Computational Physics}, + file = {/home/johannes/Nextcloud/Zotero/Kasim_Vinko_2021_Learning the exchange-correlation functional from nature with fully.pdf;/home/johannes/Zotero/storage/TFQLR3CJ/2102.html} +} + +@article{kaundinyaPredictionElectronDensity2022, + title = {Prediction of the {{Electron Density}} of {{States}} for {{Crystalline Compounds}} with {{Atomistic Line Graph Neural Networks}} ({{ALIGNN}})}, + author = {Kaundinya, Prathik R. and Choudhary, Kamal and Kalidindi, Surya R.}, + date = {2022-04-01}, + journaltitle = {JOM}, + shortjournal = {JOM}, + volume = {74}, + number = {4}, + pages = {1395--1405}, + issn = {1543-1851}, + doi = {10.1007/s11837-022-05199-y}, + url = {https://doi.org/10.1007/s11837-022-05199-y}, + urldate = {2022-06-30}, + abstract = {Machine learning (ML)-based models have greatly enhanced the traditional materials discovery and design pipeline. Specifically, in recent years, surrogate ML models for material property prediction have demonstrated success in predicting discrete scalar-valued target properties to within reasonable accuracy of their DFT-computed values. However, accurate prediction of spectral targets, such as the electron density of states (DOS), poses a much more challenging problem due to the complexity of the target, and the limited amount of available training data. In this study, we present an extension of the recently developed atomistic line graph neural network to accurately predict DOS of a large set of material unit cell structures, trained to the publicly available JARVIS-DFT dataset. Furthermore, we evaluate two methods of representation of the target quantity: a direct discretized spectrum, and a compressed low-dimensional representation obtained using an autoencoder. Through this work, we demonstrate the utility of graph-based featurization and modeling methods in the prediction of complex targets that depend on both chemistry and directional characteristics of material structures.}, + langid = {english}, + keywords = {_tablet,ALIGNN,autoencoder,DFT,dimensionality reduction,dimensionality reduction of target,GNN,JARVIS,JARVIS-DFT,ML,ML-DFT,ML-ESM,MPNN,prediction from structure,prediction of LDOS}, + file = {/home/johannes/Nextcloud/Zotero/Kaundinya et al_2022_Prediction of the Electron Density of States for Crystalline Compounds with.pdf} +} + +@article{keimerPhysicsQuantumMaterials2017, + title = {The Physics of Quantum Materials}, + author = {Keimer, B. and Moore, J. E.}, + date = {2017-11}, + journaltitle = {Nature Physics}, + shortjournal = {Nature Phys}, + volume = {13}, + number = {11}, + pages = {1045--1055}, + publisher = {{Nature Publishing Group}}, + issn = {1745-2481}, + doi = {10.1038/nphys4302}, + url = {https://www.nature.com/articles/nphys4302}, + urldate = {2021-08-24}, + abstract = {The physical description of all materials is rooted in quantum mechanics, which describes how atoms bond and electrons interact at a fundamental level. Although these quantum effects can in many cases be approximated by a classical description at the macroscopic level, in recent years there has been growing interest in material systems where quantum effects remain manifest over a wider range of energy and length scales. Such quantum materials include superconductors, graphene, topological insulators, Weyl semimetals, quantum spin liquids, and spin ices. Many of them derive their properties from reduced dimensionality, in particular from confinement of electrons to two-dimensional sheets. Moreover, they tend to be materials in which electrons cannot be considered as independent particles but interact strongly and give rise to collective excitations known as quasiparticles. In all cases, however, quantum-mechanical effects fundamentally alter properties of the material. This Review surveys the electronic properties of quantum materials through the prism of the electron wavefunction, and examines how its entanglement and topology give rise to a rich variety of quantum states and phases; these are less classically describable than conventional ordered states also driven by quantum mechanics, such as ferromagnetism.}, + issue = {11}, + langid = {english}, + annotation = {Bandiera\_abtest: a Cg\_type: Nature Research Journals Primary\_atype: Reviews Subject\_term: Quantum physics;Theoretical physics Subject\_term\_id: quantum-physics;theoretical-physics}, + file = {/home/johannes/Nextcloud/Zotero/Keimer_Moore_2017_The physics of quantum materials.pdf} +} + +@article{khorshidiAmpModularApproach2016, + title = {Amp: {{A}} Modular Approach to Machine Learning in Atomistic Simulations}, + shorttitle = {Amp}, + author = {Khorshidi, Alireza and Peterson, Andrew A.}, + date = {2016-10-01}, + journaltitle = {Computer Physics Communications}, + shortjournal = {Computer Physics Communications}, + volume = {207}, + pages = {310--324}, + issn = {0010-4655}, + doi = {10.1016/j.cpc.2016.05.010}, + url = {https://www.sciencedirect.com/science/article/pii/S0010465516301266}, + urldate = {2021-08-21}, + abstract = {Electronic structure calculations, such as those employing Kohn–Sham density functional theory or ab initio wavefunction theories, have allowed for atomistic-level understandings of a wide variety of phenomena and properties of matter at small scales. However, the computational cost of electronic structure methods drastically increases with length and time scales, which makes these methods difficult for long time-scale molecular dynamics simulations or large-sized systems. Machine-learning techniques can provide accurate potentials that can match the quality of electronic structure calculations, provided sufficient training data. These potentials can then be used to rapidly simulate large and long time-scale phenomena at similar quality to the parent electronic structure approach. Machine-learning potentials usually take a bias-free mathematical form and can be readily developed for a wide variety of systems. Electronic structure calculations have favorable properties–namely that they are noiseless and targeted training data can be produced on-demand–that make them particularly well-suited for machine learning. This paper discusses our modular approach to atomistic machine learning through the development of the open-source Atomistic Machine-learning Package (Amp), which allows for representations of both the total and atom-centered potential energy surface, in both periodic and non-periodic systems. Potentials developed through the atom-centered approach are simultaneously applicable for systems with various sizes. Interpolation can be enhanced by introducing custom descriptors of the local environment. We demonstrate this in the current work for Gaussian-type, bispectrum, and Zernike-type descriptors. Amp ~has an intuitive and modular structure with an interface through the python scripting language yet has parallelizable fortran components for demanding tasks; it is designed to integrate closely with the widely used Atomic Simulation Environment (ASE), which makes it compatible with a wide variety of commercial and open-source electronic structure codes. We finally demonstrate that the neural network model inside Amp ~can accurately interpolate electronic structure energies as well as forces of thousands of multi-species atomic systems. Program summary Program title: Amp Catalogue identifier: AFAK\_v1\_0 Program summary URL:http://cpc.cs.qub.ac.uk/summaries/AFAK\_v1\_0.html Program obtainable from: CPC Program Library, Queen’s University, Belfast, N. Ireland Licensing provisions: yes No. of lines in distributed program, including test data, etc.: 21239 No. of bytes in distributed program, including test data, etc.: 1412975 Distribution format: tar.gz Programming language: Python, Fortran. Computer: PC, Mac. Operating system: Linux, Mac, Windows. Has the code been vectorized or parallelized?: Yes RAM: Variable, depending on the number and size of atomic systems. Classification: 16.1, 2.1. External routines: ASE, NumPy, SciPy, f2py, matplotlib Nature of problem: Atomic interactions within many-body systems typically have complicated functional forms, difficult to represent in simple pre-decided closed-forms. Solution method: Machine learning provides flexible functional forms that can be improved as new situations are encountered. Typically, interatomic potentials yield from machine learning simultaneously apply to different system sizes. Unusual features: Amp is as modular as possible, providing a framework for the user to create atomic environment descriptor and regression model at will. Moreover, it has Atomic Simulation Environment (ASE) interface, facilitating interactive collaboration with other electronic structure calculators within ASE. Running time: Variable, depending on the number and size of atomic systems.}, + langid = {english}, + keywords = {ACSF,ase,bispectrum,BPNN,BPSF,descriptors,DFT,library,MD,ML,MLP,models,neural network potentials,PES,SOAP,with-code,Zernike descriptors}, + file = {/home/johannes/Nextcloud/Zotero/Khorshidi_Peterson_2016_Amp.pdf;/home/johannes/Zotero/storage/PFLW4RH4/S0010465516301266.html} +} + +@misc{kidgerNeuralDifferentialEquations2022, + title = {On {{Neural Differential Equations}}}, + author = {Kidger, Patrick}, + date = {2022-02-04}, + number = {arXiv:2202.02435}, + eprint = {2202.02435}, + eprinttype = {arxiv}, + primaryclass = {cs, math, stat}, + publisher = {{arXiv}}, + doi = {10.48550/arXiv.2202.02435}, + url = {http://arxiv.org/abs/2202.02435}, + urldate = {2022-09-07}, + abstract = {The conjoining of dynamical systems and deep learning has become a topic of great interest. In particular, neural differential equations (NDEs) demonstrate that neural networks and differential equation are two sides of the same coin. Traditional parameterised differential equations are a special case. Many popular neural network architectures, such as residual networks and recurrent networks, are discretisations. NDEs are suitable for tackling generative problems, dynamical systems, and time series (particularly in physics, finance, ...) and are thus of interest to both modern machine learning and traditional mathematical modelling. NDEs offer high-capacity function approximation, strong priors on model space, the ability to handle irregular data, memory efficiency, and a wealth of available theory on both sides. This doctoral thesis provides an in-depth survey of the field. Topics include: neural ordinary differential equations (e.g. for hybrid neural/mechanistic modelling of physical systems); neural controlled differential equations (e.g. for learning functions of irregular time series); and neural stochastic differential equations (e.g. to produce generative models capable of representing complex stochastic dynamics, or sampling from complex high-dimensional distributions). Further topics include: numerical methods for NDEs (e.g. reversible differential equations solvers, backpropagation through differential equations, Brownian reconstruction); symbolic regression for dynamical systems (e.g. via regularised evolution); and deep implicit models (e.g. deep equilibrium models, differentiable optimisation). We anticipate this thesis will be of interest to anyone interested in the marriage of deep learning with dynamical systems, and hope it will provide a useful reference for the current state of the art.}, + archiveprefix = {arXiv}, + keywords = {backpropagation,Deep learning,differential equations,NDE,PINN,thesis}, + file = {/home/johannes/Nextcloud/Zotero/Kidger_2022_On Neural Differential Equations.pdf;/home/johannes/Zotero/storage/EHARV7VZ/2202.html} +} + +@article{kippChiralHallEffect2021, + title = {The Chiral {{Hall}} Effect in Canted Ferromagnets and Antiferromagnets}, + author = {Kipp, Jonathan and Samanta, Kartik and Lux, Fabian R. and Merte, Maximilian and Go, Dongwook and Hanke, Jan-Philipp and Redies, Matthias and Freimuth, Frank and Blügel, Stefan and Ležaić, Marjana and Mokrousov, Yuriy}, + date = {2021-05-14}, + journaltitle = {Communications Physics}, + shortjournal = {Commun Phys}, + volume = {4}, + number = {1}, + pages = {1--12}, + publisher = {{Nature Publishing Group}}, + issn = {2399-3650}, + doi = {10.1038/s42005-021-00587-3}, + url = {https://www.nature.com/articles/s42005-021-00587-3}, + urldate = {2022-06-01}, + abstract = {The anomalous Hall effect has been indispensable in our understanding of numerous magnetic phenomena. This concerns both ferromagnetic materials, as well as diverse classes of antiferromagnets, where in addition to the anomalous and recently discovered crystal Hall effect, the topological Hall effect in noncoplanar antiferromagnets has been a subject of intensive research in the past decades. Here, we uncover a distinct flavor of the Hall effect emerging in generic canted spin systems. We demonstrate that upon canting, the anomalous Hall effect acquires a contribution which is sensitive to the sense of imprinted vector chirality among spins. We explore the origins and basic properties of corresponding chiral Hall effect, and closely tie it to the symmetry properties of the system. Our findings suggest that the chiral Hall effect and corresponding chiral magneto-optical effects emerge as useful tools in characterizing an interplay of structure and chirality in complex magnets, as well as in tracking their chiral dynamics and fluctuations.}, + issue = {1}, + langid = {english}, + keywords = {Magnetic properties and materials,Spintronics}, + file = {/home/johannes/Nextcloud/Zotero/Kipp et al_2021_The chiral Hall effect in canted ferromagnets and antiferromagnets.pdf} +} + +@article{kirkpatrickPushingFrontiersDensity2021, + title = {Pushing the Frontiers of Density Functionals by Solving the Fractional Electron Problem}, + author = {Kirkpatrick, James and McMorrow, Brendan and Turban, David H. P. and Gaunt, Alexander L. and Spencer, James S. and Matthews, Alexander G. D. G. and Obika, Annette and Thiry, Louis and Fortunato, Meire and Pfau, David and Castellanos, Lara Román and Petersen, Stig and Nelson, Alexander W. R. and Kohli, Pushmeet and Mori-Sánchez, Paula and Hassabis, Demis and Cohen, Aron J.}, + date = {2021-12-10}, + journaltitle = {Science}, + volume = {374}, + number = {6573}, + pages = {1385--1389}, + publisher = {{American Association for the Advancement of Science}}, + doi = {10.1126/science.abj6511}, + url = {https://www.science.org/doi/10.1126/science.abj6511}, + urldate = {2022-05-13}, + keywords = {DeepMind,density functional,DFT,DM21,ML,ML-DFA,ML-DFT,ML-ESM,molecules,original publication,with-code}, + file = {/home/johannes/Nextcloud/Zotero/Kirkpatrick et al_2021_Pushing the frontiers of density functionals by solving the fractional electron.pdf} +} + +@unpublished{klicperaGemNetUniversalDirectional2021, + title = {{{GemNet}}: {{Universal Directional Graph Neural Networks}} for {{Molecules}}}, + shorttitle = {{{GemNet}}}, + author = {Klicpera, Johannes and Becker, Florian and Günnemann, Stephan}, + date = {2021-11-29}, + eprint = {2106.08903}, + eprinttype = {arxiv}, + primaryclass = {physics, stat}, + url = {http://arxiv.org/abs/2106.08903}, + urldate = {2022-01-02}, + abstract = {Effectively predicting molecular interactions has the potential to accelerate molecular dynamics by multiple orders of magnitude and thus revolutionize chemical simulations. Graph neural networks (GNNs) have recently shown great successes for this task, overtaking classical methods based on fixed molecular kernels. However, they still appear very limited from a theoretical perspective, since regular GNNs cannot distinguish certain types of graphs. In this work we close this gap between theory and practice. We show that GNNs with directed edge embeddings and two-hop message passing are indeed universal approximators for predictions that are invariant to translation, and equivariant to permutation and rotation. We then leverage these insights and multiple structural improvements to propose the geometric message passing neural network (GemNet). We demonstrate the benefits of the proposed changes in multiple ablation studies. GemNet outperforms previous models on the COLL, MD17, and OC20 datasets by 34\%, 41\%, and 20\%, respectively, and performs especially well on the most challenging molecules. Our implementation is available online.}, + archiveprefix = {arXiv}, + keywords = {Computer Science - Machine Learning,Physics - Chemical Physics,Physics - Computational Physics,Statistics - Machine Learning}, + file = {/home/johannes/Nextcloud/Zotero/Klicpera et al_2021_GemNet.pdf;/home/johannes/Zotero/storage/NPJ8A2J3/2106.html} +} + +@article{klintenbergComputationalSearchStrong2014, + title = {Computational {{Search}} for {{Strong Topological Insulators}}: {{An Exercise}} in {{Data Mining}} and {{Electronic Structure}}}, + shorttitle = {Computational {{Search}} for {{Strong Topological Insulators}}}, + author = {Klintenberg, M. and Haraldsen, J. and Balatsky, A.}, + date = {2014-06-19}, + journaltitle = {Applied Physics Research}, + volume = {6}, + number = {4}, + pages = {p31}, + issn = {1916-9639}, + doi = {10.5539/apr.v6n4p31}, + url = {http://www.ccsenet.org/journal/index.php/apr/article/view/37961}, + urldate = {2021-05-21}, + abstract = {We report a data-mining investigation for the search of topological insulators by examining individual electronic structures for over 60,000 materials. Using a data-mining algorithm, we survey changes in band inversion with and without spin-orbit coupling by screening the calculated electronic band structure for a small gap and a change concavity at high-symmetry points. Overall, we were able to identify a number of topological candidates with varying structures and composition. Our overall goal is expand the realm of predictive theory into the determination of new and exotic complex materials through the data mining of electronic structure.}, + issue = {4}, + langid = {english}, + keywords = {classification,DFT,FP-LMTO,LDA,materials screening,topological insulator}, + file = {/home/johannes/Nextcloud/Zotero/Klintenberg et al_2014_Computational Search for Strong Topological Insulators3.pdf;/home/johannes/Nextcloud/Zotero/Klintenberg et al_2014_Computational Search for Strong Topological Insulators4.pdf;/home/johannes/Zotero/storage/9MDA4BT7/37961.html} +} + +@unpublished{klusSymmetricAntisymmetricKernels2021, + title = {Symmetric and Antisymmetric Kernels for Machine Learning Problems in Quantum Physics and Chemistry}, + author = {Klus, Stefan and Gelß, Patrick and Nüske, Feliks and Noé, Frank}, + date = {2021-03-31}, + eprint = {2103.17233}, + eprinttype = {arxiv}, + primaryclass = {math-ph, physics:physics, physics:quant-ph, stat}, + url = {http://arxiv.org/abs/2103.17233}, + urldate = {2021-05-13}, + abstract = {We derive symmetric and antisymmetric kernels by symmetrizing and antisymmetrizing conventional kernels and analyze their properties. In particular, we compute the feature space dimensions of the resulting polynomial kernels, prove that the reproducing kernel Hilbert spaces induced by symmetric and antisymmetric Gaussian kernels are dense in the space of symmetric and antisymmetric functions, and propose a Slater determinant representation of the antisymmetric Gaussian kernel, which allows for an efficient evaluation even if the state space is high-dimensional. Furthermore, we show that by exploiting symmetries or antisymmetries the size of the training data set can be significantly reduced. The results are illustrated with guiding examples and simple quantum physics and chemistry applications.}, + archiveprefix = {arXiv}, + keywords = {kernel methods,ML,models}, + file = {/home/johannes/Nextcloud/Zotero/Klus et al_2021_Symmetric and antisymmetric kernels for machine learning problems in quantum.pdf;/home/johannes/Zotero/storage/WM8YDGB2/2103.html} +} + +@unpublished{knosgaardRepresentingIndividualElectronic2021, + title = {Representing Individual Electronic States in Crystals for Machine Learning Quasiparticle Band Structures}, + author = {Knøsgaard, Nikolaj Rørbæk and Thygesen, Kristian Sommer}, + date = {2021-07-13}, + eprint = {2107.06029}, + eprinttype = {arxiv}, + primaryclass = {physics}, + url = {http://arxiv.org/abs/2107.06029}, + urldate = {2021-08-05}, + abstract = {We address the problem of representing quantum states of electrons in a solid for the purpose of machine leaning state-specific electronic properties. Specifically, we construct a fingerprint based on energy decomposed operator matrix elements (ENDOME) and radially decomposed projected density of states (RAD-PDOS), which are both obtainable from a standard density functional theory (DFT) calculation. Using such fingerprints we train a gradient boosting model on a set of 46k G\$\_0\$W\$\_0\$ quasiparticle energies. The resulting model predicts the self-energy correction of states in materials not seen by the model with a mean absolute error of 0.14 eV. By including the material's calculated dielectric constant in the fingerprint the error can be further reduced by 30\%, which we find is due to an enhanced ability to learn the correlation/screening part of the self-energy. Our work paves the way for accurate estimates of quasiparticle band structures at the cost of a standard DFT calculation.}, + archiveprefix = {arXiv}, + keywords = {descriptors,DFT,electronic state,electronic state descriptors,ENDOME,materials,ML}, + file = {/home/johannes/Nextcloud/Zotero/Knøsgaard_Thygesen_2021_Representing individual electronic states in crystals for machine learning.pdf;/home/johannes/Zotero/storage/HWNEX2AE/2107.html} +} + +@article{kocerContinuousOptimallyComplete2020, + title = {Continuous and Optimally Complete Description of Chemical Environments Using {{Spherical Bessel}} Descriptors}, + author = {Kocer, Emir and Mason, Jeremy K. and Erturk, Hakan}, + date = {2020-01-01}, + journaltitle = {AIP Advances}, + shortjournal = {AIP Advances}, + volume = {10}, + number = {1}, + pages = {015021}, + publisher = {{American Institute of Physics}}, + doi = {10.1063/1.5111045}, + url = {https://aip.scitation.org/doi/abs/10.1063/1.5111045}, + urldate = {2021-05-13}, + keywords = {descriptors,ML,MLP,SB descriptors,SOAP,Zernike descriptors}, + file = {/home/johannes/Nextcloud/Zotero/Kocer et al_2020_Continuous and optimally complete description of chemical environments using.pdf;/home/johannes/Zotero/storage/NZCCBRZE/1.html} +} + +@article{koFourthgenerationHighdimensionalNeural2021, + title = {A Fourth-Generation High-Dimensional Neural Network Potential with Accurate Electrostatics Including Non-Local Charge Transfer}, + author = {Ko, Tsz Wai and Finkler, Jonas A. and Goedecker, Stefan and Behler, Jörg}, + date = {2021-01-15}, + journaltitle = {Nature Communications}, + shortjournal = {Nat Commun}, + volume = {12}, + number = {1}, + pages = {398}, + publisher = {{Nature Publishing Group}}, + issn = {2041-1723}, + doi = {10.1038/s41467-020-20427-2}, + url = {https://www.nature.com/articles/s41467-020-20427-2}, + urldate = {2021-07-22}, + abstract = {Machine learning potentials have become an important tool for atomistic simulations in many fields, from chemistry via molecular biology to materials science. Most of the established methods, however, rely on local properties and are thus unable to take global changes in the electronic structure into account, which result from long-range charge transfer or different charge states. In this work we overcome this limitation by introducing a fourth-generation high-dimensional neural network potential that combines a charge equilibration scheme employing environment-dependent atomic electronegativities with accurate atomic energies. The method, which is able to correctly describe global charge distributions in arbitrary systems, yields much improved energies and substantially extends the applicability of modern machine learning potentials. This is demonstrated for a series of systems representing typical scenarios in chemistry and materials science that are incorrectly described by current methods, while the fourth-generation neural network potential is in excellent agreement with electronic structure calculations.}, + issue = {1}, + langid = {english}, + keywords = {CENT,HDNNP,rec-by-bluegel}, + annotation = {Bandiera\_abtest: a Cc\_license\_type: cc\_by Cg\_type: Nature Research Journals Primary\_atype: Research Subject\_term: Computational methods;Density functional theory;Method development;Molecular dynamics Subject\_term\_id: computational-methods;density-functional-theory;method-development;molecular-dynamics}, + file = {/home/johannes/Nextcloud/Zotero/Ko et al_2021_A fourth-generation high-dimensional neural network potential with accurate.pdf;/home/johannes/Zotero/storage/2Z8H4HHW/s41467-020-20427-2.html} +} + +@article{koGeneralPurposeMachineLearning2021, + title = {General-{{Purpose Machine Learning Potentials Capturing Nonlocal Charge Transfer}}}, + author = {Ko, Tsz Wai and Finkler, Jonas A. and Goedecker, Stefan and Behler, Jörg}, + date = {2021-02-16}, + journaltitle = {Accounts of Chemical Research}, + shortjournal = {Acc. Chem. Res.}, + volume = {54}, + number = {4}, + pages = {808--817}, + publisher = {{American Chemical Society}}, + issn = {0001-4842}, + doi = {10.1021/acs.accounts.0c00689}, + url = {https://doi.org/10.1021/acs.accounts.0c00689}, + urldate = {2021-05-18}, + abstract = {ConspectusThe development of first-principles-quality machine learning potentials (MLP) has seen tremendous progress, now enabling computer simulations of complex systems for which sufficiently accurate interatomic potentials have not been available. These advances and the increasing use of MLPs for more and more diverse systems gave rise to new questions regarding their applicability and limitations, which has constantly driven new developments. The resulting MLPs can be classified into several generations depending on the types of systems they are able to describe. First-generation MLPs, as introduced 25 years ago, have been applicable to low-dimensional systems such as small molecules. MLPs became a practical tool for complex systems in chemistry and materials science with the introduction of high-dimensional neural network potentials (HDNNP) in 2007, which represented the first MLP of the second generation. Second-generation MLPs are based on the concept of locality and express the total energy as a sum of environment-dependent atomic energies, which allows applications to very large systems containing thousands of atoms with linearly scaling computational costs. Since second-generation MLPs do not consider interactions beyond the local chemical environments, a natural extension has been the inclusion of long-range interactions without truncation, mainly electrostatics, employing environment-dependent charges establishing the third MLP generation. A variety of second- and, to some extent, also third-generation MLPs are currently the standard methods in ML-based atomistic simulations.In spite of countless successful applications, in recent years it has been recognized that the accuracy of MLPs relying on local atomic energies and charges is still insufficient for systems with long-ranged dependencies in the electronic structure. These can, for instance, result from nonlocal charge transfer or ionization and are omnipresent in many important types of systems and chemical processes such as the protonation and deprotonation of organic and biomolecules, redox reactions, and defects and doping in materials. In all of these situations, small local modifications can change the system globally, resulting in different equilibrium structures, charge distributions, and reactivity. These phenomena cannot be captured by second- and third-generation MLPs. Consequently, the inclusion of nonlocal phenomena has been identified as a next key step in the development of a new fourth generation of MLPs. While a first fourth-generation MLP, the charge equilibration neural network technique (CENT), was introduced in 2015, only very recently have a range of new general-purpose methods applicable to a broad range of physical scenarios emerged. In this Account, we show how fourth-generation HDNNPs can be obtained by combining the concepts of CENT and second-generation HDNNPs. These new MLPs allow for a highly accurate description of systems where nonlocal charge transfer is important.}, + keywords = {HDNNP,long-range interaction,ML,MLP,models}, + file = {/home/johannes/Nextcloud/Zotero/Ko et al_2021_General-Purpose Machine Learning Potentials Capturing Nonlocal Charge Transfer.pdf} +} + +@article{kohnSelfConsistentEquationsIncluding1965, + title = {Self-{{Consistent Equations Including Exchange}} and {{Correlation Effects}}}, + author = {Kohn, W.}, + date = {1965}, + journaltitle = {Physical Review}, + shortjournal = {Phys. Rev.}, + volume = {140}, + pages = {A1133-A1138}, + doi = {10.1103/PhysRev.140.A1133}, + issue = {4A}, + keywords = {DFT,KS-DFT,original publication}, + file = {/home/johannes/Nextcloud/Zotero/Kohn_1965_Self-Consistent Equations Including Exchange and Correlation Effects.pdf;/home/johannes/Zotero/storage/4CF9DCKS/PhysRev.140.html} +} + +@article{korshunovaOpenChemDeepLearning2021, + title = {{{OpenChem}}: {{A Deep Learning Toolkit}} for {{Computational Chemistry}} and {{Drug Design}}}, + shorttitle = {{{OpenChem}}}, + author = {Korshunova, Maria and Ginsburg, Boris and Tropsha, Alexander and Isayev, Olexandr}, + date = {2021-01-25}, + journaltitle = {Journal of Chemical Information and Modeling}, + shortjournal = {J. Chem. Inf. Model.}, + volume = {61}, + number = {1}, + pages = {7--13}, + publisher = {{American Chemical Society}}, + issn = {1549-9596}, + doi = {10.1021/acs.jcim.0c00971}, + url = {https://doi.org/10.1021/acs.jcim.0c00971}, + urldate = {2021-07-21}, + abstract = {Deep learning models have demonstrated outstanding results in many data-rich areas of research, such as computer vision and natural language processing. Currently, there is a rise of deep learning in computational chemistry and materials informatics, where deep learning could be effectively applied in modeling the relationship between chemical structures and their properties. With the immense growth of chemical and materials data, deep learning models can begin to outperform conventional machine learning techniques such as random forest, support vector machines, and nearest neighbor. Herein, we introduce OpenChem, a PyTorch-based deep learning toolkit for computational chemistry and drug design. OpenChem offers easy and fast model development, modular software design, and several data preprocessing modules. It is freely available via the GitHub repository.}, + keywords = {chemistry,GCN,GNN,library,ML,models,pytorch}, + file = {/home/johannes/Nextcloud/Zotero/Korshunova et al_2021_OpenChem.pdf;/home/johannes/Zotero/storage/U5ZHRH93/acs.jcim.html} +} + +@article{kosmaStrongSpinorbitTorque2020, + title = {Strong Spin-Orbit Torque Effect on Magnetic Defects Due to Topological Surface State Electrons in {{Bi2Te3}}}, + shorttitle = {Strong Spin-Orbit Torque Effect on Magnetic Defects Due to Topological Surface State Electrons In}, + author = {Kosma, Adamantia}, + date = {2020}, + journaltitle = {Physical Review B}, + shortjournal = {Phys. Rev. B}, + volume = {102}, + number = {14}, + doi = {10.1103/PhysRevB.102.144424}, + keywords = {Boltzmann theory,DFT,IFF,KKR,PGI-1/IAS-1,rec-by-ruess,Spin-orbit effects,surface physics,topological insulator,topological spin textures,transport properties}, + file = {/home/johannes/Nextcloud/Zotero/Kosma_2020_Strong spin-orbit torque effect on magnetic defects due to topological surface.pdf;/home/johannes/Zotero/storage/5JLDY6FT/PhysRevB.102.html} +} + +@misc{krennPredictingFutureAI2022, + title = {Predicting the {{Future}} of {{AI}} with {{AI}}: {{High-quality}} Link Prediction in an Exponentially Growing Knowledge Network}, + shorttitle = {Predicting the {{Future}} of {{AI}} with {{AI}}}, + author = {Krenn, Mario and Buffoni, Lorenzo and Coutinho, Bruno and Eppel, Sagi and Foster, Jacob Gates and Gritsevskiy, Andrew and Lee, Harlin and Lu, Yichao and Moutinho, Joao P. and Sanjabi, Nima and Sonthalia, Rishi and Tran, Ngoc Mai and Valente, Francisco and Xie, Yangxinyu and Yu, Rose and Kopp, Michael}, + date = {2022-09-23}, + number = {arXiv:2210.00881}, + eprint = {2210.00881}, + eprinttype = {arxiv}, + primaryclass = {cs}, + publisher = {{arXiv}}, + doi = {10.48550/arXiv.2210.00881}, + url = {http://arxiv.org/abs/2210.00881}, + urldate = {2022-10-05}, + abstract = {A tool that could suggest new personalized research directions and ideas by taking insights from the scientific literature could significantly accelerate the progress of science. A field that might benefit from such an approach is artificial intelligence (AI) research, where the number of scientific publications has been growing exponentially over the last years, making it challenging for human researchers to keep track of the progress. Here, we use AI techniques to predict the future research directions of AI itself. We develop a new graph-based benchmark based on real-world data -- the Science4Cast benchmark, which aims to predict the future state of an evolving semantic network of AI. For that, we use more than 100,000 research papers and build up a knowledge network with more than 64,000 concept nodes. We then present ten diverse methods to tackle this task, ranging from pure statistical to pure learning methods. Surprisingly, the most powerful methods use a carefully curated set of network features, rather than an end-to-end AI approach. It indicates a great potential that can be unleashed for purely ML approaches without human knowledge. Ultimately, better predictions of new future research directions will be a crucial component of more advanced research suggestion tools.}, + archiveprefix = {arXiv}, + keywords = {General ML,literature analysis}, + file = {/home/johannes/Nextcloud/Zotero/Krenn et al_2022_Predicting the Future of AI with AI.pdf;/home/johannes/Zotero/storage/MZBX2N4K/2210.html} +} + +@article{krennSelfreferencingEmbeddedStrings2020, + title = {Self-Referencing Embedded Strings ({{SELFIES}}): {{A}} 100\% Robust Molecular String Representation}, + shorttitle = {Self-Referencing Embedded Strings ({{SELFIES}})}, + author = {Krenn, Mario and Häse, Florian and Nigam, AkshatKumar and Friederich, Pascal and Aspuru-Guzik, Alan}, + date = {2020-11}, + journaltitle = {Machine Learning: Science and Technology}, + shortjournal = {Mach. Learn.: Sci. Technol.}, + volume = {1}, + number = {4}, + pages = {045024}, + publisher = {{IOP Publishing}}, + issn = {2632-2153}, + doi = {10.1088/2632-2153/aba947}, + url = {https://doi.org/10.1088/2632-2153/aba947}, + urldate = {2021-08-05}, + abstract = {The discovery of novel materials and functional molecules can help to solve some of society’s most urgent challenges, ranging from efficient energy harvesting and storage to uncovering novel pharmaceutical drug candidates. Traditionally matter engineering–generally denoted as inverse design–was based massively on human intuition and high-throughput virtual screening. The last few years have seen the emergence of significant interest in computer-inspired designs based on evolutionary or deep learning methods. The major challenge here is that the standard strings molecular representation SMILES shows substantial weaknesses in that task because large fractions of strings do not correspond to valid molecules. Here, we solve this problem at a fundamental level and introduce SELFIES (SELF-referencIng Embedded Strings), a string-based representation of molecules which is 100\% robust. Every SELFIES string corresponds to a valid molecule, and SELFIES can represent every molecule. SELFIES can be directly applied in arbitrary machine learning models without the adaptation of the models; each of the generated molecule candidates is valid. In our experiments, the model’s internal memory stores two orders of magnitude more diverse molecules than a similar test with SMILES. Furthermore, as all molecules are valid, it allows for explanation and interpretation of the internal working of the generative models.}, + langid = {english}, + keywords = {chemistry,descriptors,GAN,library,ML,molecules,SELFIES,SMILES,VAE}, + file = {/home/johannes/Nextcloud/Zotero/Krenn et al_2020_Self-referencing embedded strings (SELFIES).pdf} +} + +@article{kulikRoadmapMachineLearning2022, + title = {Roadmap on {{Machine Learning}} in {{Electronic Structure}}}, + author = {Kulik, Heather and Hammerschmidt, Thomas and Schmidt, Jonathan and Botti, Silvana and Marques, Miguel A. L. and Boley, Mario and Scheffler, Matthias and Todorović, Milica and Rinke, Patrick and Oses, Corey and Smolyanyuk, Andriy and Curtarolo, Stefano and Tkatchenko, Alexandre and Bartok, Albert and Manzhos, Sergei and Ihara, Manabu and Carrington, Tucker and Behler, Jörg and Isayev, Olexandr and Veit, Max and Grisafi, Andrea and Nigam, Jigyasa and Ceriotti, Michele and Schütt, Kristoff T and Westermayr, Julia and Gastegger, Michael and Maurer, Reinhard and Kalita, Bhupalee and Burke, Kieron and Nagai, Ryo and Akashi, Ryosuke and Sugino, Osamu and Hermann, Jan and Noé, Frank and Pilati, Sebastiano and Draxl, Claudia and Kuban, Martin and Rigamonti, Santiago and Scheidgen, Markus and Esters, Marco and Hicks, David and Toher, Cormac and Balachandran, Prasanna and Tamblyn, Isaac and Whitelam, Stephen and Bellinger, Colin and Ghiringhelli, Luca M.}, + date = {2022}, + journaltitle = {Electronic Structure}, + shortjournal = {Electron. Struct.}, + issn = {2516-1075}, + doi = {10.1088/2516-1075/ac572f}, + url = {http://iopscience.iop.org/article/10.1088/2516-1075/ac572f}, + urldate = {2022-03-28}, + abstract = {In recent years, we have been witnessing a paradigm shift in computational materials science. In fact, traditional methods, mostly developed in the second half of the XXth century, are being complemented, extended, and sometimes even completely replaced by faster, simpler, and often more accurate approaches. The new approaches, that we collectively label by machine learning, have their origins in the fields of informatics and artificial intelligence, but are making rapid inroads in all other branches of science. With this in mind, this Roadmap article, consisting of multiple contributions from experts across the field, discusses the use of machine learning in materials science, and share perspectives on current and future challenges in problems as diverse as the prediction of materials properties, the construction of force-fields, the development of exchange correlation functionals for density-functional theory, the solution of the many-body problem, and more. In spite of the already numerous and exciting success stories, we are just at the beginning of a long path that will reshape materials science for the many challenges of the XXIth century.}, + langid = {english}, + keywords = {_tablet,descriptors,DFT,electronic structure theory,MD,ML,ML-DFT,ML-ESM,models,review,roadmap,surrogate model}, + file = {/home/johannes/Nextcloud/Zotero/false;/home/johannes/Nextcloud/Zotero/Kulik et al_2022_Roadmap on Machine Learning in Electronic Structure.pdf} +} + +@article{kumarTopologicalQuantumMaterials2021, + title = {Topological {{Quantum Materials}} from the {{Viewpoint}} of {{Chemistry}}}, + author = {Kumar, Nitesh and Guin, Satya N. and Manna, Kaustuv and Shekhar, Chandra and Felser, Claudia}, + date = {2021-03-10}, + journaltitle = {Chemical Reviews}, + shortjournal = {Chem. Rev.}, + volume = {121}, + number = {5}, + pages = {2780--2815}, + publisher = {{American Chemical Society}}, + issn = {0009-2665}, + doi = {10.1021/acs.chemrev.0c00732}, + url = {https://doi.org/10.1021/acs.chemrev.0c00732}, + urldate = {2021-05-13}, + abstract = {Topology, a mathematical concept, has recently become a popular and truly transdisciplinary topic encompassing condensed matter physics, solid state chemistry, and materials science. Since there is a direct connection between real space, namely atoms, valence electrons, bonds, and orbitals, and reciprocal space, namely bands and Fermi surfaces, via symmetry and topology, classifying topological materials within a single-particle picture is possible. Currently, most materials are classified as trivial insulators, semimetals, and metals or as topological insulators, Dirac and Weyl nodal-line semimetals, and topological metals. The key ingredients for topology are certain symmetries, the inert pair effect of the outer electrons leading to inversion of the conduction and valence bands, and spin–orbit coupling. This review presents the topological concepts related to solids from the viewpoint of a solid-state chemist, summarizes techniques for growing single crystals, and describes basic physical property measurement techniques to characterize topological materials beyond their structure and provide examples of such materials. Finally, a brief outlook on the impact of topology in other areas of chemistry is provided at the end of the article.}, + keywords = {_tablet,chemistry,topological insulator}, + file = {/home/johannes/Nextcloud/Zotero/Kumar et al_2021_Topological Quantum Materials from the Viewpoint of Chemistry.pdf} +} + +@misc{lamGraphCastLearningSkillful2022, + title = {{{GraphCast}}: {{Learning}} Skillful Medium-Range Global Weather Forecasting}, + shorttitle = {{{GraphCast}}}, + author = {Lam, Remi and Sanchez-Gonzalez, Alvaro and Willson, Matthew and Wirnsberger, Peter and Fortunato, Meire and Pritzel, Alexander and Ravuri, Suman and Ewalds, Timo and Alet, Ferran and Eaton-Rosen, Zach and Hu, Weihua and Merose, Alexander and Hoyer, Stephan and Holland, George and Stott, Jacklynn and Vinyals, Oriol and Mohamed, Shakir and Battaglia, Peter}, + date = {2022-12-24}, + number = {arXiv:2212.12794}, + eprint = {2212.12794}, + eprinttype = {arxiv}, + primaryclass = {physics}, + publisher = {{arXiv}}, + doi = {10.48550/arXiv.2212.12794}, + url = {http://arxiv.org/abs/2212.12794}, + urldate = {2022-12-31}, + abstract = {We introduce a machine-learning (ML)-based weather simulator--called "GraphCast"--which outperforms the most accurate deterministic operational medium-range weather forecasting system in the world, as well as all previous ML baselines. GraphCast is an autoregressive model, based on graph neural networks and a novel high-resolution multi-scale mesh representation, which we trained on historical weather data from the European Centre for Medium-Range Weather Forecasts (ECMWF)'s ERA5 reanalysis archive. It can make 10-day forecasts, at 6-hour time intervals, of five surface variables and six atmospheric variables, each at 37 vertical pressure levels, on a 0.25-degree latitude-longitude grid, which corresponds to roughly 25 x 25 kilometer resolution at the equator. Our results show GraphCast is more accurate than ECMWF's deterministic operational forecasting system, HRES, on 90.0\% of the 2760 variable and lead time combinations we evaluated. GraphCast also outperforms the most accurate previous ML-based weather forecasting model on 99.2\% of the 252 targets it reported. GraphCast can generate a 10-day forecast (35 gigabytes of data) in under 60 seconds on Cloud TPU v4 hardware. Unlike traditional forecasting methods, ML-based forecasting scales well with data: by training on bigger, higher quality, and more recent data, the skill of the forecasts can improve. Together these results represent a key step forward in complementing and improving weather modeling with ML, open new opportunities for fast, accurate forecasting, and help realize the promise of ML-based simulation in the physical sciences.}, + archiveprefix = {arXiv}, + keywords = {/unread,DeepMind,GNN,PDE,SciML,simulation,weather forecasting}, + file = {/home/johannes/Nextcloud/Zotero/Lam et al_2022_GraphCast.pdf;/home/johannes/Zotero/storage/8UD54ESE/2212.html} +} + +@unpublished{langerRepresentationsMoleculesMaterials2021, + title = {Representations of Molecules and Materials for Interpolation of Quantum-Mechanical Simulations via Machine Learning}, + author = {Langer, Marcel F. and Goeßmann, Alex and Rupp, Matthias}, + date = {2021-02-09}, + eprint = {2003.12081}, + eprinttype = {arxiv}, + primaryclass = {cond-mat, physics:physics}, + url = {http://arxiv.org/abs/2003.12081}, + urldate = {2021-05-13}, + abstract = {Computational study of molecules and materials from first principles is a cornerstone of physics, chemistry, and materials science, but limited by the cost of accurate and precise simulations. In settings involving many simulations, machine learning can reduce these costs, often by orders of magnitude, by interpolating between reference simulations. This requires representations that describe any molecule or material and support interpolation. We comprehensively review and discuss current representations and relations between them, using a unified mathematical framework based on many-body functions, group averaging, and tensor products. For selected state-of-the-art representations, we compare energy predictions for organic molecules, binary alloys, and Al-Ga-In sesquioxides in numerical experiments controlled for data distribution, regression method, and hyper-parameter optimization.}, + archiveprefix = {arXiv}, + keywords = {_tablet,ACE,BoB,BS,CM,descriptors,GPR,KRR,library,materials,MBTR,ML,models,MTP,review,SOAP,with-code}, + file = {/home/johannes/Nextcloud/Zotero/Langer et al_2021_Representations of molecules and materials for interpolation of.pdf;/home/johannes/Zotero/storage/5BG77UWY/2003.html} +} + +@article{langerRepresentationsMoleculesMaterials2022, + title = {Representations of Molecules and Materials for Interpolation of Quantum-Mechanical Simulations via Machine Learning}, + author = {Langer, Marcel F. and Goeßmann, Alex and Rupp, Matthias}, + date = {2022-03-16}, + journaltitle = {npj Computational Materials}, + shortjournal = {npj Comput Mater}, + volume = {8}, + number = {1}, + pages = {1--14}, + publisher = {{Nature Publishing Group}}, + issn = {2057-3960}, + doi = {10.1038/s41524-022-00721-x}, + url = {https://www.nature.com/articles/s41524-022-00721-x}, + urldate = {2022-08-17}, + abstract = {Computational study of molecules and materials from first principles is a cornerstone of physics, chemistry, and materials science, but limited by the cost of accurate and precise simulations. In settings involving many simulations, machine learning can reduce these costs, often by orders of magnitude, by interpolating between reference simulations. This requires representations that describe any molecule or material and support interpolation. We comprehensively review and discuss current representations and relations between them. For selected state-of-the-art representations, we compare energy predictions for organic molecules, binary alloys, and Al–Ga–In sesquioxides in numerical experiments controlled for data distribution, regression method, and hyper-parameter optimization.}, + issue = {1}, + langid = {english}, + keywords = {_tablet,ACE,benchmarking,BoB,BS,CM,descriptor comparison,descriptors,GPR,KRR,library,materials,MBTR,ML,models,MTP,review,SOAP,with-code}, + file = {/home/johannes/Nextcloud/Zotero/Langer et al_2022_Representations of molecules and materials for interpolation of.pdf;/home/johannes/Zotero/storage/9RVUDSSX/s41524-022-00721-x.html} +} + +@article{larsenAtomicSimulationEnvironment2017, + title = {The Atomic Simulation Environment—a {{Python}} Library for Working with Atoms}, + author = {Larsen, Ask Hjorth and Mortensen, Jens Jørgen and Blomqvist, Jakob and Castelli, Ivano E. and Christensen, Rune and Du\textbackslash lak, Marcin and Friis, Jesper and Groves, Michael N. and Hammer, Bjørk and Hargus, Cory and Hermes, Eric D. and Jennings, Paul C. and Jensen, Peter Bjerre and Kermode, James and Kitchin, John R. and Kolsbjerg, Esben Leonhard and Kubal, Joseph and Kaasbjerg, Kristen and Lysgaard, Steen and Maronsson, Jón Bergmann and Maxson, Tristan and Olsen, Thomas and Pastewka, Lars and Peterson, Andrew and Rostgaard, Carsten and Schiøtz, Jakob and Schütt, Ole and Strange, Mikkel and Thygesen, Kristian S. and Vegge, Tejs and Vilhelmsen, Lasse and Walter, Michael and Zeng, Zhenhua and Jacobsen, Karsten W.}, + date = {2017-06}, + shortjournal = {J. Phys.: Condens. Matter}, + volume = {29}, + number = {27}, + pages = {273002}, + publisher = {{IOP Publishing}}, + issn = {0953-8984}, + doi = {10.1088/1361-648X/aa680e}, + url = {https://doi.org/10.1088/1361-648x/aa680e}, + urldate = {2021-10-17}, + abstract = {The atomic simulation environment (ASE) is a software package written in the Python programming language with the aim of setting up, steering, and analyzing atomistic simulations. In ASE, tasks are fully scripted in Python. The powerful syntax of Python combined with the NumPy array library make it possible to perform very complex simulation tasks. For example, a sequence of calculations may be performed with the use of a simple ‘for-loop’ construction. Calculations of energy, forces, stresses and other quantities are performed through interfaces to many external electronic structure codes or force fields using a uniform interface. On top of this calculator interface, ASE provides modules for performing many standard simulation tasks such as structure optimization, molecular dynamics, handling of constraints and performing nudged elastic band calculations.}, + langid = {english}, + file = {/home/johannes/Nextcloud/Zotero/Larsen et al_2017_The atomic simulation environment—a Python library for working with atoms.pdf} +} + +@online{LearningLJPotential, + title = {Learning a {{LJ}} Potential — {{PiNN}} Documentation}, + url = {https://teoroo-pinn.readthedocs.io/en/latest/notebooks/Learn_LJ_potential.html}, + urldate = {2021-05-13}, + keywords = {ANN,MD,ML,MLP,notebook}, + file = {/home/johannes/Zotero/storage/VDHVAB3I/Learn_LJ_potential.html} +} + +@article{lehtolaAssessmentInitialGuesses2019, + title = {Assessment of {{Initial Guesses}} for {{Self-Consistent Field Calculations}}. {{Superposition}} of {{Atomic Potentials}}: {{Simple}} yet {{Efficient}}}, + shorttitle = {Assessment of {{Initial Guesses}} for {{Self-Consistent Field Calculations}}. {{Superposition}} of {{Atomic Potentials}}}, + author = {Lehtola, Susi}, + date = {2019-03-12}, + journaltitle = {Journal of Chemical Theory and Computation}, + shortjournal = {J. Chem. Theory Comput.}, + volume = {15}, + number = {3}, + pages = {1593--1604}, + publisher = {{American Chemical Society}}, + issn = {1549-9618}, + doi = {10.1021/acs.jctc.8b01089}, + url = {https://doi.org/10.1021/acs.jctc.8b01089}, + urldate = {2022-05-17}, + abstract = {Electronic structure calculations, such as in the Hartree–Fock or Kohn–Sham density functional approach, require an initial guess for the molecular orbitals. The quality of the initial guess has a significant impact on the speed of convergence of the self-consistent field (SCF) procedure. Popular choices for the initial guess include the one-electron guess from the core Hamiltonian, the extended Hückel method, and the superposition of atomic densities (SAD). Here, we discuss alternative guesses obtained from the superposition of atomic potentials (SAP), which is easily implementable even in real-space calculations. We also discuss a variant of SAD which produces guess orbitals by purification of the density matrix that could also be used in real-space calculations, as well as a parameter-free variant of the extended Hückel method, which resembles the SAP method and is easy to implement on top of existing SAD infrastructure. The performance of the core Hamiltonian, the SAD, and the SAP guesses as well as the extended Hückel variant is assessed in nonrelativistic calculations on a data set of 259 molecules ranging from the first to the fourth periods by projecting the guess orbitals onto precomputed, converged SCF solutions in single- to triple-ζ basis sets. It is shown that the proposed SAP guess is the best guess on average. The extended Hückel guess offers a good alternative, with less scatter in accuracy.}, + keywords = {initial guess,SCF}, + file = {/home/johannes/Nextcloud/Zotero/Lehtola_2019_Assessment of Initial Guesses for Self-Consistent Field Calculations.pdf} +} + +@article{lehtolaRecentDevelopmentsLibxc2018, + title = {Recent Developments in Libxc — {{A}} Comprehensive Library of Functionals for Density Functional Theory}, + author = {Lehtola, Susi and Steigemann, Conrad and Oliveira, Micael J. T. and Marques, Miguel A. L.}, + date = {2018-01-01}, + journaltitle = {SoftwareX}, + shortjournal = {SoftwareX}, + volume = {7}, + pages = {1--5}, + issn = {2352-7110}, + doi = {10.1016/j.softx.2017.11.002}, + url = {https://www.sciencedirect.com/science/article/pii/S2352711017300602}, + urldate = {2021-12-14}, + abstract = {libxc is a library of exchange–correlation functionals for density-functional theory. We are concerned with semi-local functionals (or the semi-local part of hybrid functionals), namely local-density approximations, generalized-gradient approximations, and meta-generalized-gradient approximations. Currently we include around 400 functionals for the exchange, correlation, and the kinetic energy, spanning more than 50 years of research. Moreover, libxc is by now used by more than 20 codes, not only from the atomic, molecular, and solid-state physics, but also from the quantum chemistry communities.}, + langid = {english}, + keywords = {DFT,library,with-code}, + file = {/home/johannes/Nextcloud/Zotero/Lehtola et al_2018_Recent developments in libxc — A comprehensive library of functionals for.pdf} +} + +@article{leiDesignAnalysisMachine2019, + title = {Design and Analysis of Machine Learning Exchange-Correlation Functionals via Rotationally Invariant Convolutional Descriptors}, + author = {Lei, Xiangyun and Medford, Andrew J.}, + date = {2019-06-12}, + journaltitle = {Physical Review Materials}, + shortjournal = {Phys. Rev. Materials}, + volume = {3}, + number = {6}, + pages = {063801}, + publisher = {{American Physical Society}}, + doi = {10.1103/PhysRevMaterials.3.063801}, + url = {https://link.aps.org/doi/10.1103/PhysRevMaterials.3.063801}, + urldate = {2022-07-05}, + abstract = {In this work we explore the potential of a data-driven approach to the design of exchange-correlation (xc) functionals. The approach, inspired by convolutional filters in computer vision and surrogate functions from optimization, utilizes convolutions of the electron density to form a feature space to represent local electronic environments and neural networks to map the features to the exchange-correlation energy density. These features are orbital free, and provide a systematic route to including information at various length scales. This work shows that convolutional descriptors are theoretically capable of an exact representation of the electron density, and proposes Maxwell-Cartesian spherical harmonic kernels as a class of rotationally invariant descriptors for the construction of machine learned functionals. The approach is demonstrated using data from the B3LYP functional on a number of small molecules containing C, H, O, and N along with a neural network regression model. The machine learned functionals are compared to standard physical approximations and the accuracy is assessed for the absolute energy of each molecular system as well as formation energies. The results indicate that it is possible to reproduce the exchange-correlation portion of B3LYP formation energies to within chemical accuracy using orbital-free descriptors with a spatial extent of 0.2 Ã…. The findings provide empirical insight into the spatial range of electron exchange, and suggest that the combination of convolutional descriptors and machine learning regression models is a promising framework for xc functional design, although challenges remain in obtaining training data and generating models consistent with pseudopotentials.}, + keywords = {B3LYP,CNN,DFT,grid-based descriptors,LDA,MCSH,ML,ML-DFA,ML-DFT,ML-ESM,prediction from density,VWN}, + file = {/home/johannes/Nextcloud/Zotero/Lei_Medford_2019_Design and analysis of machine learning exchange-correlation functionals via.pdf;/home/johannes/Zotero/storage/RNGY77UQ/Lei and Medford - 2019 - Design and analysis of machine learning exchange-c.pdf;/home/johannes/Zotero/storage/D2P5RDDM/PhysRevMaterials.3.html} +} + +@article{lejaeghereErrorEstimatesSolidState2014, + title = {Error {{Estimates}} for {{Solid-State Density-Functional Theory Predictions}}: {{An Overview}} by {{Means}} of the {{Ground-State Elemental Crystals}}}, + shorttitle = {Error {{Estimates}} for {{Solid-State Density-Functional Theory Predictions}}}, + author = {Lejaeghere, K. and Van Speybroeck, V. and Van Oost, G. and Cottenier, S.}, + date = {2014-01-01}, + journaltitle = {Critical Reviews in Solid State and Materials Sciences}, + volume = {39}, + number = {1}, + pages = {1--24}, + publisher = {{Taylor \& Francis}}, + issn = {1040-8436}, + doi = {10.1080/10408436.2013.772503}, + url = {https://doi.org/10.1080/10408436.2013.772503}, + urldate = {2021-10-15}, + abstract = {Predictions of observable properties by density-functional theory calculations (DFT) are used increasingly often by experimental condensed-matter physicists and materials engineers as data. These predictions are used to analyze recent measurements, or to plan future experiments in a rational way. Increasingly more experimental scientists in these fields therefore face the natural question: what is the expected error for such a first-principles prediction? Information and experience about this question is implicitly available in the computational community, scattered over two decades of literature. The present review aims to summarize and quantify this implicit knowledge. This eventually leads to a practical protocol that allows any scientist—experimental or theoretical—to determine justifiable error estimates for many basic property predictions, without having to perform additional DFT calculations. A central role is played by a large and diverse test set of crystalline solids, containing all ground-state elemental crystals (except most lanthanides). For several properties of each crystal, the difference between DFT results and experimental values is assessed. We discuss trends in these deviations and review explanations suggested in the literature. A prerequisite for such an error analysis is that different implementations of the same first-principles formalism provide the same predictions. Therefore, the reproducibility of predictions across several mainstream methods and codes is discussed too. A quality factor Δ expresses the spread in predictions from two distinct DFT implementations by a single number. To compare the PAW method to the highly accurate APW+lo approach, a code assessment of VASP and GPAW (PAW) with respect to WIEN2k (APW+lo) yields Δ-values of 1.9 and 3.3 meV/atom, respectively. In both cases the PAW potentials recommended by the respective codes have been used. These differences are an order of magnitude smaller than the typical difference with experiment, and therefore predictions by APW+lo and PAW are for practical purposes identical.}, + keywords = {benchmarking,code comparison,density-functional theory,error estimate}, + annotation = {\_eprint: https://doi.org/10.1080/10408436.2013.772503}, + file = {/home/johannes/Nextcloud/Zotero/Lejaeghere et al_2014_Error Estimates for Solid-State Density-Functional Theory Predictions.pdf;/home/johannes/Zotero/storage/92BC3LBZ/10408436.2013.html} +} + +@unpublished{lewisLearningElectronDensities2021, + title = {Learning Electron Densities in the Condensed-Phase}, + author = {Lewis, Alan M. and Grisafi, Andrea and Ceriotti, Michele and Rossi, Mariana}, + date = {2021-06-09}, + eprint = {2106.05364}, + eprinttype = {arxiv}, + primaryclass = {cond-mat, physics:physics}, + url = {http://arxiv.org/abs/2106.05364}, + urldate = {2021-06-29}, + abstract = {We introduce a local machine-learning method for predicting the electron densities of periodic systems. The framework is based on a numerical, atom-centred auxiliary basis, which enables an accurate expansion of the all-electron density in a form suitable for learning isolated and periodic systems alike. We show that using this formulation the electron densities of metals, semiconductors and molecular crystals can all be accurately predicted using a symmetry-adapted Gaussian process regression model, properly adjusted for the non-orthogonal nature of the basis. These predicted densities enable the efficient calculation of electronic properties which present errors on the order of tens of meV/atom when compared to ab initio density-functional calculations. We demonstrate the key power of this approach by using a model trained on ice unit cells containing only 4 water molecules to predict the electron densities of cells containing up to 512 molecules, and see no increase in the magnitude of the errors of derived electronic properties when increasing the system size. Indeed, we find that these extrapolated derived energies are more accurate than those predicted using a direct machine-learning model.}, + archiveprefix = {arXiv}, + keywords = {DFT,GPR,library,ML,ML-DFT,ML-ESM,models,molecules,molecules & solids,prediction of electron density,prediction of ground-state properties,Resolution of the identity,SA-GPR,SALTED,solids,with-code}, + file = {/home/johannes/Nextcloud/Zotero/false;/home/johannes/Nextcloud/Zotero/Lewis et al_2021_Learning electron densities in the condensed-phase.pdf;/home/johannes/Zotero/storage/IC2NJGYT/2106.html} +} + +@article{lewisLearningElectronDensities2021a, + title = {Learning {{Electron Densities}} in the {{Condensed Phase}}}, + author = {Lewis, Alan M. and Grisafi, Andrea and Ceriotti, Michele and Rossi, Mariana}, + date = {2021-11-09}, + journaltitle = {Journal of Chemical Theory and Computation}, + shortjournal = {J. Chem. Theory Comput.}, + volume = {17}, + number = {11}, + pages = {7203--7214}, + publisher = {{American Chemical Society}}, + issn = {1549-9618}, + doi = {10.1021/acs.jctc.1c00576}, + url = {https://doi.org/10.1021/acs.jctc.1c00576}, + urldate = {2022-08-22}, + abstract = {We introduce a local machine-learning method for predicting the electron densities of periodic systems. The framework is based on a numerical, atom-centered auxiliary basis, which enables an accurate expansion of the all-electron density in a form suitable for learning isolated and periodic systems alike. We show that, using this formulation, the electron densities of metals, semiconductors, and molecular crystals can all be accurately predicted using symmetry-adapted Gaussian process regression models, properly adjusted for the nonorthogonal nature of the basis. These predicted densities enable the efficient calculation of electronic properties, which present errors on the order of tens of meV/atom when compared to ab initio density-functional calculations. We demonstrate the key power of this approach by using a model trained on ice unit cells containing only 4 water molecules to predict the electron densities of cells containing up to 512 molecules and see no increase in the magnitude of the errors of derived electronic properties when increasing the system size. Indeed, we find that these extrapolated derived energies are more accurate than those predicted using a direct machine-learning model. Finally, on heterogeneous data sets SALTED can predict electron densities with errors below 4\%.}, + keywords = {_tablet,DFT,GPR,library,ML,ML-DFT,ML-ESM,models,molecules,molecules & solids,prediction of electron density,prediction of ground-state properties,Resolution of the identity,SA-GPR,SALTED,solids,with-code}, + file = {/home/johannes/Nextcloud/Zotero/Lewis et al_2021_Learning Electron Densities in the Condensed Phase.pdf;/home/johannes/Zotero/storage/S9FT2FEZ/acs.jctc.html} +} + +@unpublished{liDeepNeuralNetwork2021, + title = {Deep {{Neural Network Representation}} of {{Density Functional Theory Hamiltonian}}}, + author = {Li, He and Wang, Zun and Zou, Nianlong and Ye, Meng and Duan, Wenhui and Xu, Yong}, + date = {2021-04-08}, + eprint = {2104.03786}, + eprinttype = {arxiv}, + primaryclass = {cond-mat, physics:physics, physics:quant-ph}, + url = {http://arxiv.org/abs/2104.03786}, + urldate = {2022-01-02}, + abstract = {The marriage of density functional theory (DFT) and deep learning methods has the potential to revolutionize modern research of material science. Here we study the crucial problem of representing DFT Hamiltonian for crystalline materials of arbitrary configurations via deep neural network. A general framework is proposed to deal with the infinite dimensionality and covariance transformation of DFT Hamiltonian matrix in virtue of locality and use message passing neural network together with graph representation for deep learning. Our example study on graphene-based systems demonstrates that high accuracy (\$\textbackslash sim\$meV) and good transferability can be obtained for DFT Hamiltonian, ensuring accurate predictions of materials properties without DFT. The Deep Hamiltonian method provides a solution to the accuracy-efficiency dilemma of DFT and opens new opportunities to explore large-scale materials and physics.}, + archiveprefix = {arXiv}, + keywords = {Condensed Matter - Disordered Systems and Neural Networks,Condensed Matter - Materials Science,Condensed Matter - Mesoscale and Nanoscale Physics,Physics - Computational Physics,Quantum Physics}, + file = {/home/johannes/Nextcloud/Zotero/Li et al_2021_Deep Neural Network Representation of Density Functional Theory Hamiltonian.pdf;/home/johannes/Zotero/storage/B7RUP7VH/2104.html} +} + +@article{liDLHubSimplifyingPublication2021, + title = {{{DLHub}}: {{Simplifying}} Publication, Discovery, and Use of Machine Learning Models in Science}, + shorttitle = {{{DLHub}}}, + author = {Li, Zhuozhao and Chard, Ryan and Ward, Logan and Chard, Kyle and Skluzacek, Tyler J. and Babuji, Yadu and Woodard, Anna and Tuecke, Steven and Blaiszik, Ben and Franklin, Michael J. and Foster, Ian}, + date = {2021-01-01}, + journaltitle = {Journal of Parallel and Distributed Computing}, + shortjournal = {Journal of Parallel and Distributed Computing}, + volume = {147}, + pages = {64--76}, + issn = {0743-7315}, + doi = {10.1016/j.jpdc.2020.08.006}, + url = {https://www.sciencedirect.com/science/article/pii/S0743731520303464}, + urldate = {2022-01-03}, + abstract = {Machine Learning (ML) has become a critical tool enabling new methods of analysis and driving deeper understanding of phenomena across scientific disciplines. There is a growing need for “learning systems†to support various phases in the ML lifecycle. While others have focused on supporting model development, training, and inference, few have focused on the unique challenges inherent in science, such as the need to publish and share models and to serve them on a range of available computing resources. In this paper, we present the Data and Learning Hub for science (DLHub), a learning system designed to support these use cases. Specifically, DLHub enables publication of models, with descriptive metadata, persistent identifiers, and flexible access control. It packages arbitrary models into portable servable containers, and enables low-latency, distributed serving of these models on heterogeneous compute resources. We show that DLHub supports low-latency model inference comparable to other model serving systems including TensorFlow Serving, SageMaker, and Clipper, and improved performance, by up to 95\%, with batching and memoization enabled. We also show that DLHub can scale to concurrently serve models on 500 containers. Finally, we describe five case studies that highlight the use of DLHub for scientific applications.}, + langid = {english}, + keywords = {DLHub,Learning systems,Machine learning,Model serving}, + file = {/home/johannes/Zotero/storage/9B88LYEZ/S0743731520303464.html} +} + +@article{liKohnShamEquationsRegularizer2021, + title = {Kohn-{{Sham Equations}} as {{Regularizer}}: {{Building Prior Knowledge}} into {{Machine-Learned Physics}}}, + shorttitle = {Kohn-{{Sham Equations}} as {{Regularizer}}}, + author = {Li, Li and Hoyer, Stephan and Pederson, Ryan and Sun, Ruoxi and Cubuk, Ekin D. and Riley, Patrick and Burke, Kieron}, + date = {2021-01-20}, + journaltitle = {Physical Review Letters}, + shortjournal = {Phys. Rev. Lett.}, + volume = {126}, + number = {3}, + pages = {036401}, + publisher = {{American Physical Society}}, + doi = {10.1103/PhysRevLett.126.036401}, + url = {https://link.aps.org/doi/10.1103/PhysRevLett.126.036401}, + urldate = {2022-07-07}, + abstract = {Including prior knowledge is important for effective machine learning models in physics and is usually achieved by explicitly adding loss terms or constraints on model architectures. Prior knowledge embedded in the physics computation itself rarely draws attention. We show that solving the Kohn-Sham equations when training neural networks for the exchange-correlation functional provides an implicit regularization that greatly improves generalization. Two separations suffice for learning the entire one-dimensional H2 dissociation curve within chemical accuracy, including the strongly correlated region. Our models also generalize to unseen types of molecules and overcome self-interaction error.}, + keywords = {autodiff,CNN,DFT,JAX,JAX-DFT,Kohn-Sham regularizer,ML,ML-DFA,ML-DFT,ML-ESM,molecules,original publication,prediction from density,regularization,RNN}, + file = {/home/johannes/Nextcloud/Zotero/Li et al_2021_Kohn-Sham Equations as Regularizer.pdf;/home/johannes/Zotero/storage/CAFV9KV8/Li et al_2021_Kohn-Sham Equations as Regularizer.pdf;/home/johannes/Zotero/storage/QQA9HJV3/Li et al. - 2021 - Kohn-Sham Equations as Regularizer Building Prior.gif;/home/johannes/Zotero/storage/2MCFRSEU/PhysRevLett.126.html} +} + +@article{liMetallizationSuperconductivityDense2014, + title = {The Metallization and Superconductivity of Dense Hydrogen Sulfide}, + author = {Li, Yinwei and Hao, Jian and Liu, Hanyu and Li, Yanling and Ma, Yanming}, + date = {2014-05-07}, + journaltitle = {The Journal of Chemical Physics}, + shortjournal = {J. Chem. Phys.}, + volume = {140}, + number = {17}, + pages = {174712}, + publisher = {{American Institute of Physics}}, + issn = {0021-9606}, + doi = {10.1063/1.4874158}, + url = {https://aip.scitation.org/doi/10.1063/1.4874158}, + urldate = {2021-10-21}, + abstract = {Hydrogen sulfide (H2S) is a prototype molecular system and a sister molecule of water (H2O). The phase diagram of solid H2S at high pressures remains largely unexplored arising from the challenges in dealing with the pressure-induced weakening of S–H bond and larger atomic core difference between H and S. Metallization is yet achieved for H2O, but it was observed for H2S above 96 GPa. However, the metallic structure of H2S remains elusive, greatly impeding the understanding of its metallicity and the potential superconductivity. We have performed an extensive structural study on solid H2S at pressure ranges of 10–200 GPa through an unbiased structure prediction method based on particle swarm optimization algorithm. Besides the findings of candidate structures for nonmetallic phases IV and V, we are able to establish stable metallic structures violating an earlier proposal of elemental decomposition into sulfur and hydrogen [R. Rousseau, M. Boero, M. Bernasconi, M. Parrinello, and K. Terakura, Phys. Rev. Lett. 85, 1254 (2000)]. Our study unravels a superconductive potential of metallic H2S with an estimated maximal transition temperature of ∼80 K at 160 GPa, higher than those predicted for most archetypal hydrogen-containing compounds (e.g., SiH4, GeH4, etc.).}, + keywords = {applications of DFT,DFT,master-thesis,superconductor}, + file = {/home/johannes/Nextcloud/Zotero/Li et al_2014_The metallization and superconductivity of dense hydrogen sulfide.pdf} +} + +@article{lindmaaTheoreticalPredictionProperties2017, + title = {Theoretical Prediction of Properties of Atomistic Systems : {{Density}} Functional Theory and Machine Learning}, + shorttitle = {Theoretical Prediction of Properties of Atomistic Systems}, + author = {Lindmaa, Alexander}, + date = {2017}, + publisher = {{Linköping University Electronic Press}}, + url = {http://urn.kb.se/resolve?urn=urn:nbn:se:liu:diva-139767}, + urldate = {2021-06-26}, + abstract = {The prediction of ground state properties of atomistic systems is of vital importance in technological advances as well as in the physical sciences. Fundamentally, these predictions are based on a ...}, + langid = {english}, + keywords = {Coulomb matrix,descriptors,DFT,Ewald sum matrix,kinetic energy density,KRR,ML,models,PCA,prediction of ground-state properties,prediction of solid formation energy}, + file = {/home/johannes/Nextcloud/Zotero/Lindmaa_2017_Theoretical prediction of properties of atomistic systems.pdf;/home/johannes/Zotero/storage/TVX96NQ7/record.html} +} + +@misc{liptonTroublingTrendsMachine2018, + title = {Troubling {{Trends}} in {{Machine Learning Scholarship}}}, + author = {Lipton, Zachary C. and Steinhardt, Jacob}, + date = {2018-07-26}, + number = {arXiv:1807.03341}, + eprint = {1807.03341}, + eprinttype = {arxiv}, + primaryclass = {cs, stat}, + publisher = {{arXiv}}, + doi = {10.48550/arXiv.1807.03341}, + url = {http://arxiv.org/abs/1807.03341}, + urldate = {2022-06-27}, + abstract = {Collectively, machine learning (ML) researchers are engaged in the creation and dissemination of knowledge about data-driven algorithms. In a given paper, researchers might aspire to any subset of the following goals, among others: to theoretically characterize what is learnable, to obtain understanding through empirically rigorous experiments, or to build a working system that has high predictive accuracy. While determining which knowledge warrants inquiry may be subjective, once the topic is fixed, papers are most valuable to the community when they act in service of the reader, creating foundational knowledge and communicating as clearly as possible. Recent progress in machine learning comes despite frequent departures from these ideals. In this paper, we focus on the following four patterns that appear to us to be trending in ML scholarship: (i) failure to distinguish between explanation and speculation; (ii) failure to identify the sources of empirical gains, e.g., emphasizing unnecessary modifications to neural architectures when gains actually stem from hyper-parameter tuning; (iii) mathiness: the use of mathematics that obfuscates or impresses rather than clarifies, e.g., by confusing technical and non-technical concepts; and (iv) misuse of language, e.g., by choosing terms of art with colloquial connotations or by overloading established technical terms. While the causes behind these patterns are uncertain, possibilities include the rapid expansion of the community, the consequent thinness of the reviewer pool, and the often-misaligned incentives between scholarship and short-term measures of success (e.g., bibliometrics, attention, and entrepreneurial opportunity). While each pattern offers a corresponding remedy (don't do it), we also discuss some speculative suggestions for how the community might combat these trends.}, + archiveprefix = {arXiv}, + keywords = {criticism,ML,research ethics,state of a field}, + file = {/home/johannes/Nextcloud/Zotero/Lipton_Steinhardt_2018_Troubling Trends in Machine Learning Scholarship.pdf;/home/johannes/Zotero/storage/HK89ZR8C/1807.html} +} + +@article{liuDensityEstimationUsing2021, + title = {Density Estimation Using Deep Generative Neural Networks}, + author = {Liu, Qiao and Xu, Jiaze and Jiang, Rui and Wong, Wing Hung}, + date = {2021-04-13}, + journaltitle = {Proceedings of the National Academy of Sciences}, + volume = {118}, + number = {15}, + pages = {e2101344118}, + publisher = {{Proceedings of the National Academy of Sciences}}, + doi = {10.1073/pnas.2101344118}, + url = {https://www.pnas.org/doi/10.1073/pnas.2101344118}, + urldate = {2022-07-08}, + keywords = {density,General ML}, + file = {/home/johannes/Nextcloud/Zotero/Liu et al_2021_Density estimation using deep generative neural networks.pdf} +} + +@article{liuImprovingPerformanceLongRangeCorrected2017, + title = {Improving the {{Performance}} of {{Long-Range-Corrected Exchange-Correlation Functional}} with an {{Embedded Neural Network}}}, + author = {Liu, Qin and Wang, JingChun and Du, PengLi and Hu, LiHong and Zheng, Xiao and Chen, GuanHua}, + date = {2017-09-28}, + journaltitle = {The Journal of Physical Chemistry A}, + shortjournal = {J. Phys. Chem. A}, + volume = {121}, + number = {38}, + pages = {7273--7281}, + publisher = {{American Chemical Society}}, + issn = {1089-5639}, + doi = {10.1021/acs.jpca.7b07045}, + url = {https://doi.org/10.1021/acs.jpca.7b07045}, + urldate = {2022-07-05}, + abstract = {A machine-learning-based exchange-correlation functional is proposed for general-purpose density functional theory calculations. It is built upon the long-range-corrected Becke–Lee–Yang–Parr (LC–BLYP) functional, along with an embedded neural network which determines the value of the range-separation parameter μ for every individual system. The structure and the weights of the neural network are optimized with a reference data set containing 368 highly accurate thermochemical and kinetic energies. The newly developed functional (LC–BLYP–NN) achieves a balanced performance for a variety of energetic properties investigated. It largely improves the accuracy of atomization energies and heats of formation on which the original LC–BLYP with a fixed μ performs rather poorly. Meanwhile, it yields a similar or slightly compromised accuracy for ionization potentials, electron affinities, and reaction barriers, for which the original LC–BLYP works reasonably well. This work clearly highlights the potential usefulness of machine-learning techniques for improving density functional calculations.}, + keywords = {autoencoder,BLYP,compositional descriptors,DFT,ML,ML-DFA,ML-DFT,ML-ESM,molecules,NN,prediction of Exc,used small dataset}, + file = {/home/johannes/Nextcloud/Zotero/Liu et al_2017_Improving the Performance of Long-Range-Corrected Exchange-Correlation.pdf;/home/johannes/Zotero/storage/76EWRKPT/acs.jpca.html} +} + +@article{liUnderstandingMachinelearnedDensity2016, + title = {Understanding Machine-Learned Density Functionals}, + author = {Li, Li and Snyder, John C. and Pelaschier, Isabelle M. and Huang, Jessica and Niranjan, Uma-Naresh and Duncan, Paul and Rupp, Matthias and Müller, Klaus-Robert and Burke, Kieron}, + date = {2016}, + journaltitle = {International Journal of Quantum Chemistry}, + volume = {116}, + number = {11}, + pages = {819--833}, + issn = {1097-461X}, + doi = {10.1002/qua.25040}, + url = {https://onlinelibrary.wiley.com/doi/abs/10.1002/qua.25040}, + urldate = {2022-07-08}, + abstract = {Machine learning (ML) is an increasingly popular statistical tool for analyzing either measured or calculated data sets. Here, we explore its application to a well-defined physics problem, investigating issues of how the underlying physics is handled by ML, and how self-consistent solutions can be found by limiting the domain in which ML is applied. The particular problem is how to find accurate approximate density functionals for the kinetic energy (KE) of noninteracting electrons. Kernel ridge regression is used to approximate the KE of non-interacting fermions in a one dimensional box as a functional of their density. The properties of different kernels and methods of cross-validation are explored, reproducing the physics faithfully in some cases, but not others. We also address how self-consistency can be achieved with information on only a limited electronic density domain. Accurate constrained optimal densities are found via a modified Euler-Lagrange constrained minimization of the machine-learned total energy, despite the poor quality of its functional derivative. A projected gradient descent algorithm is derived using local principal component analysis. Additionally, a sparse grid representation of the density can be used without degrading the performance of the methods. The implications for machine-learned density functional approximations are discussed. © 2015 Wiley Periodicals, Inc.}, + langid = {english}, + keywords = {_tablet,DFA,DFT,KRR,ML,ML-DFA,ML-DFT,ML-ESM,tutorial}, + annotation = {\_eprint: https://onlinelibrary.wiley.com/doi/pdf/10.1002/qua.25040}, + file = {/home/johannes/Nextcloud/Zotero/Li et al_2016_Understanding machine-learned density functionals.pdf;/home/johannes/Zotero/storage/ZPNDJ7AU/qua.html} +} + +@article{liuSingleNNModifiedBehler2020, + title = {{{SingleNN}}: {{Modified Behler}}–{{Parrinello Neural Network}} with {{Shared Weights}} for {{Atomistic Simulations}} with {{Transferability}}}, + shorttitle = {{{SingleNN}}}, + author = {Liu, Mingjie and Kitchin, John R.}, + date = {2020-08-13}, + journaltitle = {The Journal of Physical Chemistry C}, + shortjournal = {J. Phys. Chem. C}, + volume = {124}, + number = {32}, + pages = {17811--17818}, + publisher = {{American Chemical Society}}, + issn = {1932-7447}, + doi = {10.1021/acs.jpcc.0c04225}, + url = {https://doi.org/10.1021/acs.jpcc.0c04225}, + urldate = {2021-06-24}, + abstract = {In this article, we introduce the SingleNN, which is a modified version of the Behler–Parrinello Neural Network (BPNN) where the neural networks for the prediction of atomic energy for different elements are combined into a single network with shared weights. Using a data set containing Cu, Ge, Li, Mo, Ni, and Si, we demonstrate that SingleNN could achieve an accuracy that is on a par with BPNN for energy and force predictions. Furthermore, we demonstrate that SingleNN could learn a common transformation for the fingerprints of atoms to a latent space in which the atomic energies of the atoms are nearly linear. Using the common transformation, we could fit the data with new elements by changing only weights in the output layer in the neural network. In this way, with a moderate compromise in accuracy, we can speed up the training process significantly and potentially reduce the amount of training data needed.}, + keywords = {ACSF,ANN,BPNN,BPSF,descriptors,ML,MLP,models,SingleNN,surrogate model}, + file = {/home/johannes/Nextcloud/Zotero/Liu_Kitchin_2020_SingleNN.pdf} +} + +@thesis{lohOvercomingDataScarcity2021, + type = {Thesis}, + title = {Overcoming {{Data Scarcity}} in {{Deep Learning}} of {{Scientific Problems}}}, + author = {Loh, Charlotte Chang Le}, + date = {2021-09}, + institution = {{Massachusetts Institute of Technology}}, + url = {https://dspace.mit.edu/handle/1721.1/140165}, + urldate = {2022-05-18}, + abstract = {Data-driven approaches such as machine learning have been increasingly applied to the natural sciences, e.g. for property prediction and optimization or material discovery. An essential criteria to ensure the success of such methods is the need for extensive amounts of labeled data, making it unfeasible for data-scarce problems where labeled data generation is computationally expensive, or labour and time intensive. Here, I introduce surrogate and invariance- boosted contrastive learning (SIB-CL), a deep learning framework which overcomes data-scarcity by incorporating three “inexpensive" and easily obtainable auxiliary information. Specifically, these are: 1) abundant unlabeled data, 2) prior knowledge of known symmetries or invariances of the problem and 3) a surrogate dataset obtained at near-zero cost either from simplification or approximation. I demonstrate the effectiveness and generality of SIB-CL on various scientific problems, for example, the prediction of the density-of-states of 2D photonic crystals and solving the time-independent Schrödinger equation of 3D random potentials. SIB-CL is shown to provide orders of magnitude savings on the amount of labeled data needed when compared to conventional deep learning techniques, offering opportunities to apply data-driven methods even to data-scarce problems.}, + langid = {english}, + keywords = {contrastive learning,Deep learning,invariance,ML,NN,photonic crystals,rec-by-ruess,Schrödinger equation,small data,SSL,surrogate data,target: bandstructure,target: DOS,target: potential,TISE,transfer learning}, + annotation = {Accepted: 2022-02-07T15:28:01Z}, + file = {/home/johannes/Nextcloud/Zotero/Loh_2021_Overcoming Data Scarcity in Deep Learning of Scientific Problems.pdf;/home/johannes/Zotero/storage/BLKX75W6/140165.html} +} + +@misc{lopanitsynaModelingHighentropyTransitionmetal2022, + title = {Modeling High-Entropy Transition-Metal Alloys with Alchemical Compression}, + author = {Lopanitsyna, Nataliya and Fraux, Guillaume and Springer, Maximilian A. and De, Sandip and Ceriotti, Michele}, + date = {2022-12-26}, + number = {arXiv:2212.13254}, + eprint = {2212.13254}, + eprinttype = {arxiv}, + primaryclass = {cond-mat, physics:physics}, + publisher = {{arXiv}}, + doi = {10.48550/arXiv.2212.13254}, + url = {http://arxiv.org/abs/2212.13254}, + urldate = {2022-12-29}, + abstract = {Alloys composed of several elements in roughly equimolar composition, often referred to as high-entropy alloys, have long been of interest for their thermodynamics and peculiar mechanical properties, and more recently for their potential application in catalysis. They are a considerable challenge to traditional atomistic modeling, and also to data-driven potentials that for the most part have memory footprint, computational effort and data requirements which scale poorly with the number of elements included. We apply a recently proposed scheme to compress chemical information in a lower-dimensional space, which reduces dramatically the cost of the model with negligible loss of accuracy, to build a potential that can describe 25 d-block transition metals. The model shows semi-quantitative accuracy for prototypical alloys, and is remarkably stable when extrapolating to structures outside its training set. We use this framework to study element segregation in a computational experiment that simulates an equimolar alloy of all 25 elements, mimicking the seminal experiments by Cantor et al., and use our observations on the short-range order relations between the elements to define a data-driven set of Hume-Rothery rules that can serve as guidance for alloy design. We conclude with a study of three prototypical alloys, CoCrFeMnNi, CoCrFeMoNi and IrPdPtRhRu, determining their stability and the short-range order behavior of their constituents.}, + archiveprefix = {arXiv}, + keywords = {ACE,alchemical,chemical species scaling problem,descriptors,dimensionality reduction,high-entropy alloys,MTP,PyTorch,SOAP,transition metals,with-code}, + file = {/home/johannes/Nextcloud/Zotero/Lopanitsyna et al_2022_Modeling high-entropy transition-metal alloys with alchemical compression.pdf;/home/johannes/Zotero/storage/QNGQ9AQD/2212.html} +} + +@article{lysogorskiyPerformantImplementationAtomic2021, + title = {Performant Implementation of the Atomic Cluster Expansion ({{PACE}}) and Application to Copper and Silicon}, + author = {Lysogorskiy, Yury and van der Oord, Cas and Bochkarev, Anton and Menon, Sarath and Rinaldi, Matteo and Hammerschmidt, Thomas and Mrovec, Matous and Thompson, Aidan and Csányi, Gábor and Ortner, Christoph and Drautz, Ralf}, + date = {2021-06-28}, + journaltitle = {npj Computational Materials}, + shortjournal = {npj Comput Mater}, + volume = {7}, + number = {1}, + pages = {1--12}, + publisher = {{Nature Publishing Group}}, + issn = {2057-3960}, + doi = {10.1038/s41524-021-00559-9}, + url = {https://www.nature.com/articles/s41524-021-00559-9}, + urldate = {2022-05-11}, + abstract = {The atomic cluster expansion is a general polynomial expansion of the atomic energy in multi-atom basis functions. Here we implement the atomic cluster expansion in the performant C++ code PACE that is suitable for use in large-scale atomistic simulations. We briefly review the atomic cluster expansion and give detailed expressions for energies and forces as well as efficient algorithms for their evaluation. We demonstrate that the atomic cluster expansion as implemented in PACE shifts a previously established Pareto front for machine learning interatomic potentials toward faster and more accurate calculations. Moreover, general purpose parameterizations are presented for copper and silicon and evaluated in detail. We show that the Cu and Si potentials significantly improve on the best available potentials for highly accurate large-scale atomistic simulations.}, + issue = {1}, + langid = {english}, + keywords = {_tablet,ACE,C++,descriptors,library}, + file = {/home/johannes/Nextcloud/Zotero/Lysogorskiy et al_2021_Performant implementation of the atomic cluster expansion (PACE) and.pdf;/home/johannes/Zotero/storage/QVQD97QT/s41524-021-00559-9.html} +} + +@article{magalhaesDensityFunctionalTheory2017, + title = {Density {{Functional Theory Calculation}} of the {{Absorption Properties}} of {{Brown Carbon Chromophores Generated}} by {{Catechol Heterogeneous Ozonolysis}}}, + author = {Magalhães, Ana Catarina O. and Esteves da Silva, Joaquim C. G. and Pinto da Silva, LuÃs}, + date = {2017-08-17}, + journaltitle = {ACS Earth and Space Chemistry}, + shortjournal = {ACS Earth Space Chem.}, + volume = {1}, + number = {6}, + pages = {353--360}, + publisher = {{American Chemical Society}}, + doi = {10.1021/acsearthspacechem.7b00061}, + url = {https://doi.org/10.1021/acsearthspacechem.7b00061}, + urldate = {2021-10-21}, + abstract = {The effect of light-absorbing atmospheric particles on climate change has been incorporated into climate models, but the absence of brown carbon (BrC) in these models has been leading to significant differences between model predictions and measured data on radiative forcing. Also, little is known regarding the relationship between optical properties and chemical compositions of BrC. Thus, we have characterized the absorption properties of catechol and known heterogeneous ozonolysis products, with a theoretical approach based on density functional theory (DFT). While catechol presents a weak absorption maximum in the ultraviolet C (UVC) region, other polyaromatic derivatives present an absorption up to 6 times higher, with biphenyl-2,2′,3,3′-tetraol, biphenyl-3,3′,4,4′,5,5′-hexaol, and terphenyl-2′,3,3′,3″,4,4″-hexaol presenting the strongest absorption. Moreover, these derivatives now absorb in the ultraviolet B (UVB) and ultraviolet A (UVA) regions, which are types of actinic radiation in the ultraviolet (UV) region not filtered by atmosphere (contrary to UVC), with terphenyl molecules presenting the highest absorption maximum. Furthermore, the absorption efficiency of these compounds is potentiated in the condensed phase, such as cloud droplets, rain, fog, and water films, as a result of a higher degree of electron delocalization. This study provides reliable information regarding the absorption properties of BrC generated by catechol, which is essential for the development of accurate models of climate forcing.}, + keywords = {applications of DFT,atmospheric chemistry,DFT,master-thesis}, + file = {/home/johannes/Nextcloud/Zotero/Magalhães et al_2017_Density Functional Theory Calculation of the Absorption Properties of Brown.pdf;/home/johannes/Zotero/storage/ZRG83Z75/acsearthspacechem.html} +} + +@book{MagnetismElectronicStructure, + title = {Magnetism and the {{Electronic Structure}} of {{Crystals}}}, + url = {https://link.springer.com/book/10.1007/978-3-642-84411-9}, + urldate = {2022-06-18}, + langid = {english}, + keywords = {condensed matter,defects,DFT,magnetism}, + file = {/home/johannes/Nextcloud/Zotero/Magnetism and the Electronic Structure of Crystals.pdf;/home/johannes/Zotero/storage/QVJRNHRA/978-3-642-84411-9.html} +} + +@book{majlisQuantumTheoryMagnetism2007, + title = {The {{Quantum Theory}} of {{Magnetism}}}, + author = {Majlis, Norberto}, + date = {2007-09}, + edition = {2}, + publisher = {{WORLD SCIENTIFIC}}, + doi = {10.1142/6094}, + url = {http://www.worldscientific.com/worldscibooks/10.1142/6094}, + urldate = {2022-06-18}, + isbn = {978-981-256-792-5 978-981-277-974-8}, + langid = {english}, + keywords = {condensed matter,graduate,magnetism,textbook}, + file = {/home/johannes/Nextcloud/Zotero/Majlis_2007_The Quantum Theory of Magnetism.pdf} +} + +@article{margrafPureNonlocalMachinelearned2021, + title = {Pure Non-Local Machine-Learned Density Functional Theory for Electron Correlation}, + author = {Margraf, Johannes T. and Reuter, Karsten}, + date = {2021-01-12}, + journaltitle = {Nature Communications}, + shortjournal = {Nat Commun}, + volume = {12}, + number = {1}, + pages = {344}, + publisher = {{Nature Publishing Group}}, + issn = {2041-1723}, + doi = {10.1038/s41467-020-20471-y}, + url = {https://www.nature.com/articles/s41467-020-20471-y}, + urldate = {2021-10-15}, + abstract = {Density-functional theory (DFT) is a rigorous and (in principle) exact framework for the description of the ground state properties of atoms, molecules and solids based on their electron density. While computationally efficient density-functional approximations (DFAs) have become essential tools in computational chemistry, their (semi-)local treatment of electron correlation has a number of well-known pathologies, e.g. related to electron self-interaction. Here, we present a type of machine-learning (ML) based DFA (termed Kernel Density Functional Approximation, KDFA) that is pure, non-local and transferable, and can be efficiently trained with fully quantitative reference methods. The functionals retain the mean-field computational cost of common DFAs and are shown to be applicable to non-covalent, ionic and covalent interactions, as well as across different system sizes. We demonstrate their remarkable possibilities by computing the free energy surface for the protonated water dimer at hitherto unfeasible gold-standard coupled cluster quality on a single commodity workstation.}, + issue = {1}, + langid = {english}, + annotation = {Bandiera\_abtest: a Cc\_license\_type: cc\_by Cg\_type: Nature Research Journals Primary\_atype: Research Subject\_term: Computational chemistry;Density functional theory;Method development;Molecular dynamics Subject\_term\_id: computational-chemistry;density-functional-theory;method-development;molecular-dynamics}, + file = {/home/johannes/Nextcloud/Zotero/Margraf_Reuter_2021_Pure non-local machine-learned density functional theory for electron.pdf;/home/johannes/Zotero/storage/RCFG2NBC/s41467-020-20471-y.html} +} + +@online{MARVELDistinguishedLecture, + title = {{{MARVEL Distinguished Lecture}} — {{Georg Kresse}} - {{Events}} - Nccr-Marvel.Ch :: {{NCCR MARVEL}}}, + url = {https://nccr-marvel.ch/events/marvel-distinguished-lecture-GeorgKresse}, + urldate = {2021-05-13}, + keywords = {Bayesian regression,ML,ML-FF,MLP,models}, + file = {/home/johannes/Zotero/storage/IYCD348P/marvel-distinguished-lecture-GeorgKresse.html} +} + +@article{marzariElectronicstructureMethodsMaterials2021, + title = {Electronic-Structure Methods for Materials Design}, + author = {Marzari, Nicola and Ferretti, Andrea and Wolverton, Chris}, + date = {2021-06}, + journaltitle = {Nature Materials}, + shortjournal = {Nat. Mater.}, + volume = {20}, + number = {6}, + pages = {736--749}, + publisher = {{Nature Publishing Group}}, + issn = {1476-4660}, + doi = {10.1038/s41563-021-01013-3}, + url = {https://www.nature.com/articles/s41563-021-01013-3}, + urldate = {2022-05-13}, + abstract = {The accuracy and efficiency of electronic-structure methods to understand, predict and design the properties of materials has driven a new paradigm in research. Simulations can greatly accelerate the identification, characterization and optimization of materials, with this acceleration driven by continuous progress in theory, algorithms and hardware, and by adaptation of concepts and tools from computer science. Nevertheless, the capability to identify and characterize materials relies on the predictive accuracy of the underlying physical descriptions, and on the ability to capture the complexity of realistic systems. We provide here an overview of electronic-structure methods, of their application to the prediction of materials properties, and of the different strategies employed towards the broader goals of materials design and discovery.}, + issue = {6}, + langid = {english}, + keywords = {Electronic structure}, + file = {/home/johannes/Nextcloud/Zotero/Marzari et al_2021_Electronic-structure methods for materials design.pdf;/home/johannes/Zotero/storage/AKF7QEMC/s41563-021-01013-3.html} +} + +@inproceedings{mavropoulosKorringaKohnRostokerKKRGreen2006, + title = {The {{Korringa-Kohn-Rostoker}} ({{KKR}}) {{Green}} Function Method {{I}}. {{Electronic}} Structure of Periodic Systems}, + booktitle = {Computational {{Nanoscience}}: {{Do It Yourself}}! - {{Lecture Notes}}}, + author = {Mavropoulos, Phivos and Papanikolaou, Nikos}, + date = {2006}, + series = {{{NIC}} Series}, + volume = {31}, + pages = {131--158}, + publisher = {{NIC-Secretariat, Research Centre Jülich}}, + location = {{Jülich}}, + url = {https://juser.fz-juelich.de/record/50027/files/FZJ-2014-02214.pdf}, + urldate = {2021-06-28}, + abstract = {The Korringa-Kohn-Rostoker (KKR) method for the calculation of the electronic structure ofmaterials is founded on the concepts of the Green function and of multiple-scattering. In thismanuscript, after a short introduction to Green functions,we present a description of single-site scattering (including anisotropic potentials) and multiple-scattering theory and the KKRequations. The KKR representation of the Green function andthe algebraic Dyson equation areintroduced. We then discuss the screened KKR formalism, andits advantages in the numericaleffort for the calculation of layered systems. Finally we give a summary of the self-consistencyalgorithm for the calculation of the electronic structure.}, + eventtitle = {Computational {{Nanoscience}}: {{Do It Yourself}}!}, + isbn = {3-00-017350-1}, + keywords = {_tablet,FZJ,KKR,PGI-1/IAS-1}, + annotation = {Johannes Grotendorst, Stefan Blügel, Dominik Marx (Editors)}, + file = {/home/johannes/Nextcloud/Zotero/Mavropoulos_Papanikolaou_2006_The Korringa-Kohn-Rostoker (KKR) Green function method I.pdf} +} + +@article{mazinInverseOccamRazor2022, + title = {Inverse {{Occam}}’s Razor}, + author = {Mazin, Igor}, + date = {2022-04}, + journaltitle = {Nature Physics}, + shortjournal = {Nat. Phys.}, + volume = {18}, + number = {4}, + pages = {367--368}, + publisher = {{Nature Publishing Group}}, + issn = {1745-2481}, + doi = {10.1038/s41567-022-01575-2}, + url = {https://www.nature.com/articles/s41567-022-01575-2}, + urldate = {2022-10-21}, + abstract = {Scientists have long preferred the simplest possible explanation of their data. More recently, a worrying trend to favour unnecessarily complex interpretations has taken hold.}, + issue = {4}, + langid = {english}, + keywords = {philosophy of science,physics,rec-by-ghosh,skeptics}, + file = {/home/johannes/Nextcloud/Zotero/Mazin_2022_Inverse Occam’s razor.pdf} +} + +@article{medvedevDensityFunctionalTheory2017, + title = {Density Functional Theory Is Straying from the Path toward the Exact Functional}, + author = {Medvedev, Michael G. and Bushmarinov, Ivan S. and Sun, Jianwei and Perdew, John P. and Lyssenko, Konstantin A.}, + date = {2017-01-06}, + journaltitle = {Science}, + volume = {355}, + number = {6320}, + pages = {49--52}, + publisher = {{American Association for the Advancement of Science}}, + doi = {10.1126/science.aah5975}, + url = {https://www.science.org/doi/10.1126/science.aah5975}, + urldate = {2021-11-17}, + file = {/home/johannes/Nextcloud/Zotero/Medvedev et al_2017_Density functional theory is straying from the path toward the exact functional.pdf} +} + +@article{mehtaHighbiasLowvarianceIntroduction2019, + title = {A High-Bias, Low-Variance Introduction to {{Machine Learning}} for Physicists}, + author = {Mehta, Pankaj and Bukov, Marin and Wang, Ching-Hao and Day, Alexandre G. R. and Richardson, Clint and Fisher, Charles K. and Schwab, David J.}, + date = {2019-05-30}, + journaltitle = {Physics Reports}, + shortjournal = {Physics Reports}, + series = {A High-Bias, Low-Variance Introduction to {{Machine Learning}} for Physicists}, + volume = {810}, + pages = {1--124}, + issn = {0370-1573}, + doi = {10.1016/j.physrep.2019.03.001}, + url = {https://www.sciencedirect.com/science/article/pii/S0370157319300766}, + urldate = {2021-05-13}, + abstract = {Machine Learning (ML) is one of the most exciting and dynamic areas of modern research and application. The purpose of this review is to provide an introduction to the core concepts and tools of machine learning in a manner easily understood and intuitive to physicists. The review begins by covering fundamental concepts in ML and modern statistics such as the bias–variance tradeoff, overfitting, regularization, generalization, and gradient descent before moving on to more advanced topics in both supervised and unsupervised learning. Topics covered in the review include ensemble models, deep learning and neural networks, clustering and data visualization, energy-based models (including MaxEnt models and Restricted Boltzmann Machines), and variational methods. Throughout, we emphasize the many natural connections between ML and statistical physics. A notable aspect of the review is the use of Python Jupyter notebooks to introduce modern ML/statistical packages to readers using physics-inspired datasets (the Ising Model and Monte-Carlo simulations of supersymmetric decays of proton–proton collisions). We conclude with an extended outlook discussing possible uses of machine learning for furthering our understanding of the physical world as well as open problems in ML where physicists may be able to contribute.}, + langid = {english}, + keywords = {general,ML,notebooks,physics,rec-by-ruess,review,with-code}, + file = {/home/johannes/Nextcloud/Zotero/Mehta et al_2019_A high-bias, low-variance introduction to Machine Learning for physicists.pdf} +} + +@article{meredigCanMachineLearning2018, + title = {Can Machine Learning Identify the next High-Temperature Superconductor? {{Examining}} Extrapolation Performance for Materials Discovery}, + shorttitle = {Can Machine Learning Identify the next High-Temperature Superconductor?}, + author = {Meredig, Bryce and Antono, Erin and Church, Carena and Hutchinson, Maxwell and Ling, Julia and Paradiso, Sean and Blaiszik, Ben and Foster, Ian and Gibbons, Brenna and Hattrick-Simpers, Jason and Mehta, Apurva and Ward, Logan}, + date = {2018-10-08}, + journaltitle = {Molecular Systems Design \& Engineering}, + shortjournal = {Mol. Syst. Des. Eng.}, + volume = {3}, + number = {5}, + pages = {819--825}, + publisher = {{The Royal Society of Chemistry}}, + issn = {2058-9689}, + doi = {10.1039/C8ME00012C}, + url = {https://pubs.rsc.org/en/content/articlelanding/2018/me/c8me00012c}, + urldate = {2022-01-03}, + abstract = {Traditional machine learning (ML) metrics overestimate model performance for materials discovery. We introduce (1) leave-one-cluster-out cross-validation (LOCO CV) and (2) a simple nearest-neighbor benchmark to show that model performance in discovery applications strongly depends on the problem, data sampling, and extrapolation. Our results suggest that ML-guided iterative experimentation may outperform standard high-throughput screening for discovering breakthrough materials like high-Tc superconductors with ML.}, + langid = {english}, + file = {/home/johannes/Nextcloud/Zotero/Meredig et al_2018_Can machine learning identify the next high-temperature superconductor.pdf;/home/johannes/Zotero/storage/9WFQM4EG/c8me00012c.html} +} + +@article{merkerMachineLearningMagnetism2022, + title = {Machine Learning Magnetism Classifiers from Atomic Coordinates}, + author = {Merker, Helena A. and Heiberger, Harry and Nguyen, Linh and Liu, Tongtong and Chen, Zhantao and Andrejevic, Nina and Drucker, Nathan C. and Okabe, Ryotaro and Kim, Song Eun and Wang, Yao and Smidt, Tess and Li, Mingda}, + date = {2022-10-21}, + journaltitle = {iScience}, + shortjournal = {iScience}, + volume = {25}, + number = {10}, + pages = {105192}, + issn = {2589-0042}, + doi = {10.1016/j.isci.2022.105192}, + url = {https://www.sciencedirect.com/science/article/pii/S258900422201464X}, + urldate = {2022-10-17}, + abstract = {The determination of magnetic structure poses a long-standing challenge in condensed matter physics and materials science. Experimental techniques such as neutron diffraction are resource-limited and require complex structure refinement protocols, while computational approaches such as first-principles density functional theory (DFT) need additional semi-empirical correction, and reliable prediction is still largely limited to collinear magnetism. Here, we present a machine learning model that aims to classify the magnetic structure by inputting atomic coordinates containing transition metal and rare earth elements. By building a Euclidean equivariant neural network that preserves the crystallographic symmetry, the magnetic structure (ferromagnetic, antiferromagnetic, and non-magnetic) and magnetic propagation vector (zero or non-zero) can be predicted with an average accuracy of 77.8\% and 73.6\%. In particular, a 91\% accuracy is reached when predicting no magnetic ordering even if the structure contains magnetic element(s). Our work represents one step forward to solving the grand challenge of full magnetic structure determination.}, + langid = {english}, + keywords = {classification,e3nn,equivariant,magnetic moment,magnetic order,magnetism,ML}, + file = {/home/johannes/Nextcloud/Zotero/Merker et al_2022_Machine learning magnetism classifiers from atomic coordinates.pdf;/home/johannes/Zotero/storage/7UQX89UL/S258900422201464X.html} +} + +@article{merkysPosterioriMetadataAutomated2017, + title = {A Posteriori Metadata from Automated Provenance Tracking: Integration of {{AiiDA}} and {{TCOD}}}, + shorttitle = {A Posteriori Metadata from Automated Provenance Tracking}, + author = {Merkys, Andrius and Mounet, Nicolas and Cepellotti, Andrea and Marzari, Nicola and Gražulis, Saulius and Pizzi, Giovanni}, + date = {2017-11-14}, + journaltitle = {Journal of Cheminformatics}, + shortjournal = {Journal of Cheminformatics}, + volume = {9}, + number = {1}, + pages = {56}, + issn = {1758-2946}, + doi = {10.1186/s13321-017-0242-y}, + url = {https://doi.org/10.1186/s13321-017-0242-y}, + urldate = {2021-09-18}, + abstract = {In order to make results of computational scientific research findable, accessible, interoperable and re-usable, it is necessary to decorate them with standardised metadata. However, there are a number of technical and practical challenges that make this process difficult to achieve in practice. Here the implementation of a protocol is presented to tag crystal structures with their computed properties, without the need of human intervention to curate the data. This protocol leverages the capabilities of AiiDA, an open-source platform to manage and automate scientific computational workflows, and the TCOD, an open-access database storing computed materials properties using a well-defined and exhaustive ontology. Based on these, the complete procedure to deposit computed data in the TCOD database is automated. All relevant metadata are extracted from the full provenance information that AiiDA tracks and stores automatically while managing the calculations. Such a protocol also enables reproducibility of scientific data in the field of computational materials science. As a proof of concept, the AiiDA–TCOD interface is used to deposit 170 theoretical structures together with their computed properties and their full provenance graphs, consisting in over 4600 AiiDA nodes.}, + keywords = {AiiDA,DFT,metadata}, + file = {/home/johannes/Nextcloud/Zotero/Merkys et al_2017_A posteriori metadata from automated provenance tracking.pdf;/home/johannes/Zotero/storage/9ZIMVPJ8/s13321-017-0242-y.html} +} + +@inproceedings{missierW3CPROVFamily2013, + title = {The {{W3C PROV}} Family of Specifications for Modelling Provenance Metadata}, + booktitle = {Proceedings of the 16th {{International Conference}} on {{Extending Database Technology}}}, + author = {Missier, Paolo and Belhajjame, Khalid and Cheney, James}, + date = {2013-03-18}, + series = {{{EDBT}} '13}, + pages = {773--776}, + publisher = {{Association for Computing Machinery}}, + location = {{New York, NY, USA}}, + doi = {10.1145/2452376.2452478}, + url = {https://doi.org/10.1145/2452376.2452478}, + urldate = {2021-10-17}, + abstract = {Provenance, a form of structured metadata designed to record the origin or source of information, can be instrumental in deciding whether information is to be trusted, how it can be integrated with other diverse information sources, and how to establish attribution of information to authors throughout its history. The PROV set of specifications, produced by the World Wide Web Consortium (W3C), is designed to promote the publication of provenance information on the Web, and offers a basis for interoperability across diverse provenance management systems. The PROV provenance model is deliberately generic and domain-agnostic, but extension mechanisms are available and can be exploited for modelling specific domains. This tutorial provides an account of these specifications. Starting from intuitive and informal examples that present idiomatic provenance patterns, it progressively introduces the relational model of provenance along with the constraints model for validation of provenance documents, and concludes with example applications that show the extension points in use.}, + isbn = {978-1-4503-1597-5}, + file = {/home/johannes/Nextcloud/Zotero/Missier et al_2013_The W3C PROV family of specifications for modelling provenance metadata.pdf} +} + +@book{molnarGlobalSurrogateInterpretable, + title = {5.6 {{Global Surrogate}} | {{Interpretable Machine Learning}}}, + author = {Molnar, Christoph}, + url = {https://christophm.github.io/interpretable-ml-book/global.html}, + urldate = {2021-05-13}, + abstract = {Machine learning algorithms usually operate as black boxes and it is unclear how they derived a certain decision. This book is a guide for practitioners to make machine learning decisions interpretable.}, + keywords = {ML,surrogate model}, + file = {/home/johannes/Zotero/storage/8VNJWN2F/global.html} +} + +@article{morawietzDensityFunctionalTheoryBasedNeural2013, + title = {A {{Density-Functional Theory-Based Neural Network Potential}} for {{Water Clusters Including}} van Der {{Waals Corrections}}}, + author = {Morawietz, Tobias and Behler, Jörg}, + date = {2013-08-15}, + journaltitle = {The Journal of Physical Chemistry A}, + shortjournal = {J. Phys. Chem. A}, + volume = {117}, + number = {32}, + pages = {7356--7366}, + publisher = {{American Chemical Society}}, + issn = {1089-5639}, + doi = {10.1021/jp401225b}, + url = {https://doi.org/10.1021/jp401225b}, + urldate = {2021-05-18}, + abstract = {The fundamental importance of water for many chemical processes has motivated the development of countless efficient but approximate water potentials for large-scale molecular dynamics simulations, from simple empirical force fields to very sophisticated flexible water models. Accurate and generally applicable water potentials should fulfill a number of requirements. They should have a quality close to quantum chemical methods, they should explicitly depend on all degrees of freedom including all relevant many-body interactions, and they should be able to describe molecular dissociation and recombination. In this work, we present a high-dimensional neural network (NN) potential for water clusters based on density-functional theory (DFT) calculations, which is constructed using clusters containing up to 10 monomers and is in principle able to meet all these requirements. We investigate the reliability of specific parametrizations employing two frequently used generalized gradient approximation (GGA) exchange-correlation functionals, PBE and RPBE, as reference methods. We find that the binding energy errors of the NN potentials with respect to DFT are significantly lower than the typical uncertainties of DFT calculations arising from the choice of the exchange-correlation functional. Further, we examine the role of van der Waals interactions, which are not properly described by GGA functionals. Specifically, we incorporate the D3 scheme suggested by Grimme (J. Chem. Phys. 2010, 132, 154104) in our potentials and demonstrate that it can be applied to GGA-based NN potentials in the same way as to DFT calculations without modification. Our results show that the description of small water clusters provided by the RPBE functional is significantly improved if van der Waals interactions are included, while in case of the PBE functional, which is well-known to yield stronger binding than RPBE, van der Waals corrections lead to overestimated binding energies.}, + keywords = {chemistry,DFT,ML,MLP,models,NNP,vdW}, + file = {/home/johannes/Nextcloud/Zotero/Morawietz_Behler_2013_A Density-Functional Theory-Based Neural Network Potential for Water Clusters.pdf} +} + +@article{morenoDeepLearningHohenbergKohn2020, + title = {Deep {{Learning}} the {{Hohenberg-Kohn Maps}} of {{Density Functional Theory}}}, + author = {Moreno, Javier Robledo and Carleo, Giuseppe and Georges, Antoine}, + date = {2020-08-12}, + journaltitle = {Physical Review Letters}, + shortjournal = {Phys. Rev. Lett.}, + volume = {125}, + number = {7}, + pages = {076402}, + publisher = {{American Physical Society}}, + doi = {10.1103/PhysRevLett.125.076402}, + url = {https://link.aps.org/doi/10.1103/PhysRevLett.125.076402}, + urldate = {2021-12-14}, + abstract = {A striking consequence of the Hohenberg-Kohn theorem of density functional theory is the existence of a bijection between the local density and the ground-state many-body wave function. Here we study the problem of constructing approximations to the Hohenberg-Kohn map using a statistical learning approach. Using supervised deep learning with synthetic data, we show that this map can be accurately constructed for a chain of one-dimensional interacting spinless fermions in different phases of this model including the charge ordered Mott insulator and metallic phases and the critical point separating them. However, we also find that the learning is less effective across quantum phase transitions, suggesting an intrinsic difficulty in efficiently learning nonsmooth functional relations. We further study the problem of directly reconstructing complex observables from simple local density measurements, proposing a scheme amenable to statistical learning from experimental data.}, + keywords = {DFT,dunno,ML,ML-DFT,ML-ESM,Mott insulator,rec-by-bluegel}, + file = {/home/johannes/Nextcloud/Zotero/Moreno et al_2020_Deep Learning the Hohenberg-Kohn Maps of Density Functional Theory.pdf;/home/johannes/Zotero/storage/LWQ2IF97/Moreno et al_2020_Deep Learning the Hohenberg-Kohn Maps of Density Functional Theory2.pdf;/home/johannes/Zotero/storage/BCHNQKQ9/PhysRevLett.125.html} +} + +@article{morenoMachineLearningBand2021, + title = {Machine Learning Band Gaps from the Electron Density}, + author = {Moreno, Javier Robledo}, + date = {2021}, + journaltitle = {Physical Review Materials}, + shortjournal = {Phys. Rev. Materials}, + volume = {5}, + number = {8}, + doi = {10.1103/PhysRevMaterials.5.083802}, + keywords = {BPNN,DFT,ML,models,prediction from density,prediction of bandgap,rec-by-kim}, + file = {/home/johannes/Nextcloud/Zotero/Moreno_2021_Machine learning band gaps from the electron density.pdf;/home/johannes/Zotero/storage/B9EJXFVY/PhysRevMaterials.5.html} +} + +@article{morganOpportunitiesChallengesMachine2020, + title = {Opportunities and {{Challenges}} for {{Machine Learning}} in {{Materials Science}}}, + author = {Morgan, Dane and Jacobs, Ryan}, + date = {2020-07-01}, + journaltitle = {Annual Review of Materials Research}, + shortjournal = {Annu. Rev. Mater. Res.}, + volume = {50}, + number = {1}, + pages = {71--103}, + issn = {1531-7331, 1545-4118}, + doi = {10.1146/annurev-matsci-070218-010015}, + url = {https://www.annualreviews.org/doi/10.1146/annurev-matsci-070218-010015}, + urldate = {2021-06-29}, + abstract = {Advances in machine learning have impacted myriad areas of materials science, such as the discovery of novel materials and the improvement of molecular simulations, with likely many more important developments to come. Given the rapid changes in this field, it is challenging to understand both the breadth of opportunities and the best practices for their use. In this review, we address aspects of both problems by providing an overview of the areas in which machine learning has recently had significant impact in materials science, and then we provide a more detailed discussion on determining the accuracy and domain of applicability of some common types of machine learning models. Finally, we discuss some opportunities and challenges for the materials community to fully utilize the capabilities of machine learning.}, + langid = {english}, + file = {/home/johannes/Nextcloud/Zotero/Morgan_Jacobs_2020_Opportunities and Challenges for Machine Learning in Materials Science.pdf} +} + +@article{morgensternStrongWeak3D2021, + title = {Strong and {{Weak 3D Topological Insulators Probed}} by {{Surface Science Methods}}}, + author = {Morgenstern, Markus and Pauly, Christian and Kellner, Jens and Liebmann, Marcus and Pratzer, Marco and Bihlmayer, Gustav and Eschbach, Markus and Plucinski, Lukacz and Otto, Sebastian and Rasche, Bertold and Ruck, Michael and Richter, Manuel and Just, Sven and Lüpke, Felix and Voigtländer, Bert}, + date = {2021}, + journaltitle = {physica status solidi (b)}, + volume = {258}, + number = {1}, + pages = {2000060}, + issn = {1521-3951}, + doi = {10.1002/pssb.202000060}, + url = {https://onlinelibrary.wiley.com/doi/abs/10.1002/pssb.202000060}, + urldate = {2022-05-30}, + abstract = {The contributions of surface science methods to discover and improve 3D topological insulator materials are reviewed herein, illustrated with examples from the authors’ own work. In particular, it is demonstrated that spin-polarized angular-resolved photoelectron spectroscopy is instrumental to evidence the spin-helical surface Dirac cone, to tune its Dirac point energy toward the Fermi level, and to discover novel types of topological insulators such as dual ones or switchable ones in phase change materials. Moreover, procedures are introduced to spatially map potential fluctuations by scanning tunneling spectroscopy and to identify topological edge states in weak topological insulators.}, + langid = {english}, + keywords = {angular-resolved photoelectron spectroscopy,scanning tunneling spectroscopy,spin-polarized topological insulators}, + annotation = {\_eprint: https://onlinelibrary.wiley.com/doi/pdf/10.1002/pssb.202000060}, + file = {/home/johannes/Nextcloud/Zotero/Morgenstern et al_2021_Strong and Weak 3D Topological Insulators Probed by Surface Science Methods.pdf;/home/johannes/Zotero/storage/4RCNJ2RK/pssb.html} +} + +@misc{morrowHowValidateMachinelearned2022, + title = {How to Validate Machine-Learned Interatomic Potentials}, + author = {Morrow, Joe D. and Gardner, John L. A. and Deringer, Volker L.}, + date = {2022-11-28}, + number = {arXiv:2211.12484}, + eprint = {2211.12484}, + eprinttype = {arxiv}, + primaryclass = {cond-mat, physics:physics}, + publisher = {{arXiv}}, + doi = {10.48550/arXiv.2211.12484}, + url = {http://arxiv.org/abs/2211.12484}, + urldate = {2023-01-02}, + abstract = {Machine learning (ML) approaches enable large-scale atomistic simulations with near-quantum-mechanical accuracy. With the growing availability of these methods there arises a need for careful validation, particularly for physically agnostic models - that is, for potentials which extract the nature of atomic interactions from reference data. Here, we review the basic principles behind ML potentials and their validation for atomic-scale materials modeling. We discuss best practice in defining error metrics based on numerical performance as well as physically guided validation. We give specific recommendations that we hope will be useful for the wider community, including those researchers who intend to use ML potentials for materials "off the shelf".}, + archiveprefix = {arXiv}, + keywords = {benchmarking,best practices,how-to,ML,MLP,tutorial}, + file = {/home/johannes/Nextcloud/Zotero/Morrow et al_2022_How to validate machine-learned interatomic potentials.pdf;/home/johannes/Zotero/storage/TW3TCHB3/2211.html} +} + +@thesis{mozumderDesignMagneticInteractions2022, + type = {mathesis}, + title = {Design of Magnetic Interactions in Doped Topological Insulators}, + author = {Mozumder, Rubel}, + date = {2022-04-12}, + institution = {{Heinrich Heine University Düsseldorf}}, + abstract = {Magnetic impurities and their long-range interaction (ferromagnetic) order play pivotal roles in the topological phase transition from QSHI to QAHI. This transition transforms helical edge states belonging to the QSHI for 2D TIs (surface states for 3D TI) to the chiral edge states in QAHI for 2D TIs (surface states for 3D TI). Due to such chiral states, the QAHIs forbid back-scattering in electron conducting channels, which in turn provide passionless current and increase energy efficiency for conduct- ing channels. The chiral states are consist of single spin electrons which provide spin currents from conventional charge currents. Regarding the properties of QAHIs, the QAHIs opens a new venue for low-energy elec- tronics, spintronics and quantum computation [9]. Independently, the V- [10] and Cr-doped [11] as well as their co-doping [12] (Sb, Bi)2 Te3 shows stable QAHE but with very low temperatures (≤ 0.5K). In this high throughput ab-initio work, we will investigate other possible co-doping, dimer calculations, from the d-block elements in 3D TI Bi2 Te3 . For this purpose, we have extended AiiDA-KKR plugins by developing combine- impurity workflow called combine imps wc using GF formulation of DFT code (KKR-GF method) and the new workflow is capable to run multi- impurity calculations. Here, the dimer calculations are in the main fo- cus, and from the calculation results we will analyze Heisenberg isotropic collinear interaction (Jij ), Dzyaloshinskii–Moriya interaction (DMI, Dij ), and their ratio for each possible impurity couple. Finally, using the ob- tained Jij data we have implemented some linear regression machine learn- ing tools to understand better the dependency of Jij on some well-known factors e.g. inter-impurity distance, electronegativity. Our results from the notion of this work will give a list of some potential impurities and after their potential impurity combinations for stable QAHE. It will also render an impression of implementation of machine learning approach for designing better magnetic interactions in TIs.}, + langid = {english}, + pagetotal = {85}, + keywords = {_tablet,master-thesis,PGI-1/IAS-1,thesis}, + file = {/home/johannes/Nextcloud/Zotero/Mozumder_2022_Design of magnetic interactions in doped topological insulators.pdf} +} + +@article{mullerSpiritMultifunctionalFramework2019, + title = {\emph{Spirit}: {{Multifunctional}} Framework for Atomistic Spin Simulations}, + shorttitle = {\emph{Spirit}}, + author = {Müller, Gideon P.}, + date = {2019}, + journaltitle = {Physical Review B}, + shortjournal = {Phys. Rev. B}, + volume = {99}, + number = {22}, + doi = {10.1103/PhysRevB.99.224414}, + keywords = {browser-based visualization,interactive visualization,library,PGI-1/IAS-1,spin dynamics,Spirit,visualization,web app,with-code}, + file = {/home/johannes/Nextcloud/Zotero/Müller_2019_iSpirit-i.pdf;/home/johannes/Zotero/storage/NXE55BTB/PhysRevB.99.html} +} + +@book{MultipleScatteringTheory, + title = {Multiple {{Scattering Theory}}}, + doi = {10.1088/2053-2563/aae7d8}, + url = {https://iopscience.iop.org/book/978-0-7503-1490-9}, + urldate = {2021-12-02}, + isbn = {978-0-7503-1490-9}, + langid = {english}, + file = {/home/johannes/Nextcloud/Zotero/Multiple Scattering Theory.pdf;/home/johannes/Zotero/storage/UYLUXULV/978-0-7503-1490-9.html} +} + +@unpublished{musaelianLearningLocalEquivariant2022, + title = {Learning {{Local Equivariant Representations}} for {{Large-Scale Atomistic Dynamics}}}, + author = {Musaelian, Albert and Batzner, Simon and Johansson, Anders and Sun, Lixin and Owen, Cameron J. and Kornbluth, Mordechai and Kozinsky, Boris}, + date = {2022-04-11}, + eprint = {2204.05249}, + eprinttype = {arxiv}, + primaryclass = {cond-mat, physics:physics}, + url = {http://arxiv.org/abs/2204.05249}, + urldate = {2022-04-14}, + abstract = {A simultaneously accurate and computationally efficient parametrization of the energy and atomic forces of molecules and materials is a long-standing goal in the natural sciences. In pursuit of this goal, neural message passing has lead to a paradigm shift by describing many-body correlations of atoms through iteratively passing messages along an atomistic graph. This propagation of information, however, makes parallel computation difficult and limits the length scales that can be studied. Strictly local descriptor-based methods, on the other hand, can scale to large systems but do not currently match the high accuracy observed with message passing approaches. This work introduces Allegro, a strictly local equivariant deep learning interatomic potential that simultaneously exhibits excellent accuracy and scalability of parallel computation. Allegro learns many-body functions of atomic coordinates using a series of tensor products of learned equivariant representations, but without relying on message passing. Allegro obtains improvements over state-of-the-art methods on the QM9 and revised MD-17 data sets. A single tensor product layer is shown to outperform existing deep message passing neural networks and transformers on the QM9 benchmark. Furthermore, Allegro displays remarkable generalization to out-of-distribution data. Molecular dynamics simulations based on Allegro recover structural and kinetic properties of an amorphous phosphate electrolyte in excellent agreement with first principles calculations. Finally, we demonstrate the parallel scaling of Allegro with a dynamics simulation of 100 million atoms.}, + archiveprefix = {arXiv}, + keywords = {Allegro,GNN,MD,ML,MLP,MPNN,NequIP}, + file = {/home/johannes/Nextcloud/Zotero/Musaelian et al_2022_Learning Local Equivariant Representations for Large-Scale Atomistic Dynamics.pdf;/home/johannes/Zotero/storage/3GTGKKHF/2204.html} +} + +@article{musilEfficientImplementationAtomdensity2021, + title = {Efficient Implementation of Atom-Density Representations}, + author = {Musil, Félix and Veit, Max and Goscinski, Alexander and Fraux, Guillaume and Willatt, Michael J. and Stricker, Markus and Junge, Till and Ceriotti, Michele}, + date = {2021-03-16}, + journaltitle = {The Journal of Chemical Physics}, + shortjournal = {J. Chem. Phys.}, + volume = {154}, + number = {11}, + pages = {114109}, + publisher = {{American Institute of Physics}}, + issn = {0021-9606}, + doi = {10.1063/5.0044689}, + url = {https://aip.scitation.org/doi/10.1063/5.0044689}, + urldate = {2021-05-13}, + abstract = {Physically motivated and mathematically robust atom-centered representations of molecular structures are key to the success of modern atomistic machine learning. They lie at the foundation of a wide range of methods to predict the properties of both materials and molecules and to explore and visualize their chemical structures and compositions. Recently, it has become clear that many of the most effective representations share a fundamental formal connection. They can all be expressed as a discretization of n-body correlation functions of the local atom density, suggesting the opportunity of standardizing and, more importantly, optimizing their evaluation. We present an implementation, named librascal, whose modular design lends itself both to developing refinements to the density-based formalism and to rapid prototyping for new developments of rotationally equivariant atomistic representations. As an example, we discuss smooth overlap of atomic position (SOAP) features, perhaps the most widely used member of this family of representations, to show how the expansion of the local density can be optimized for any choice of radial basis sets. We discuss the representation in the context of a kernel ridge regression model, commonly used with SOAP features, and analyze how the computational effort scales for each of the individual steps of the calculation. By applying data reduction techniques in feature space, we show how to reduce the total computational cost by a factor of up to 4 without affecting the model’s symmetry properties and without significantly impacting its accuracy.}, + keywords = {descriptors,library,librascal,ML,SOAP,with-code}, + file = {/home/johannes/Nextcloud/Zotero/Musil et al_2021_Efficient implementation of atom-density representations.pdf;/home/johannes/Zotero/storage/A3DVYDNI/5.html} +} + +@article{musilFastAccurateUncertainty2019, + title = {Fast and {{Accurate Uncertainty Estimation}} in {{Chemical Machine Learning}}}, + author = {Musil, Félix and Willatt, Michael J. and Langovoy, Mikhail A. and Ceriotti, Michele}, + date = {2019-02-12}, + journaltitle = {Journal of Chemical Theory and Computation}, + shortjournal = {J. Chem. Theory Comput.}, + volume = {15}, + number = {2}, + pages = {906--915}, + publisher = {{American Chemical Society}}, + issn = {1549-9618}, + doi = {10.1021/acs.jctc.8b00959}, + url = {https://doi.org/10.1021/acs.jctc.8b00959}, + urldate = {2021-05-30}, + abstract = {We present a scheme to obtain an inexpensive and reliable estimate of the uncertainty associated with the predictions of a machine-learning model of atomic and molecular properties. The scheme is based on resampling, with multiple models being generated based on subsampling of the same training data. The accuracy of the uncertainty prediction can be benchmarked by maximum likelihood estimation, which can also be used to correct for correlations between resampled models and to improve the performance of the uncertainty estimation by a cross-validation procedure. In the case of sparse Gaussian Process Regression models, this resampled estimator can be evaluated at negligible cost. We demonstrate the reliability of these estimates for the prediction of molecular and materials energetics and for the estimation of nuclear chemical shieldings in molecular crystals. Extension to estimate the uncertainty in energy differences, forces, or other correlated predictions is straightforward. This method can be easily applied to other machine-learning schemes and will be beneficial to make data-driven predictions more reliable and to facilitate training-set optimization and active-learning strategies.}, + keywords = {descriptors,GPR,library,ML,models,SA-GPR,SOAP,uncertainty quantification,with-code}, + file = {/home/johannes/Nextcloud/Zotero/Musil et al_2019_Fast and Accurate Uncertainty Estimation in Chemical Machine Learning.pdf;/home/johannes/Zotero/storage/PGUZKGX5/acs.jctc.html} +} + +@article{musilMachineLearningAtomic2019, + title = {Machine {{Learning}} at the {{Atomic Scale}}}, + author = {Musil, Félix and Ceriotti, Michele}, + date = {2019-12-18}, + journaltitle = {CHIMIA International Journal for Chemistry}, + shortjournal = {chimia (aarau)}, + volume = {73}, + number = {12}, + pages = {972--982}, + issn = {0009-4293}, + doi = {10.2533/chimia.2019.972}, + url = {https://www.ingentaconnect.com/content/10.2533/chimia.2019.972}, + urldate = {2021-05-30}, + abstract = {Statistical learning algorithms are finding more and more applications in science and technology. Atomic-scale modeling is no exception, with machine learning becoming commonplace as a tool to predict energy, forces and properties of molecules and condensed-phase systems. This short review summarizes recent progress in the field, focusing in particular on the problem of representing an atomic configuration in a mathematically robust and computationally efficient way. We also discuss some of the regression algorithms that have been used to construct surrogate models of atomic-scale properties. We then show examples of how the optimization of the machine-learning models can both incorporate and reveal insights onto the physical phenomena that underlie structure–property relations.}, + langid = {english}, + keywords = {descriptors,descriptors analysis,ML,SOAP}, + file = {/home/johannes/Nextcloud/Zotero/Musil_Ceriotti_2019_Machine Learning at the Atomic Scale.pdf} +} + +@article{musilPhysicsInspiredStructuralRepresentations2021, + title = {Physics-{{Inspired Structural Representations}} for {{Molecules}} and {{Materials}}}, + author = {Musil, Felix and Grisafi, Andrea and Bartók, Albert P. and Ortner, Christoph and Csányi, Gábor and Ceriotti, Michele}, + date = {2021-07-26}, + journaltitle = {Chemical Reviews}, + shortjournal = {Chem. Rev.}, + publisher = {{American Chemical Society}}, + issn = {0009-2665}, + doi = {10.1021/acs.chemrev.1c00021}, + url = {https://doi.org/10.1021/acs.chemrev.1c00021}, + urldate = {2021-08-16}, + abstract = {The first step in the construction of a regression model or a data-driven analysis, aiming to predict or elucidate the relationship between the atomic-scale structure of matter and its properties, involves transforming the Cartesian coordinates of the atoms into a suitable representation. The development of atomic-scale representations has played, and continues to play, a central role in the success of machine-learning methods for chemistry and materials science. This review summarizes the current understanding of the nature and characteristics of the most commonly used structural and chemical descriptions of atomistic structures, highlighting the deep underlying connections between different frameworks and the ideas that lead to computationally efficient and universally applicable models. It emphasizes the link between properties, structures, their physical chemistry, and their mathematical description, provides examples of recent applications to a diverse set of chemical and materials science problems, and outlines the open questions and the most promising research directions in the field.}, + keywords = {_tablet,ACSF,descriptor comparison,descriptor dimred,descriptors,MBTR,ML,review,SOAP}, + file = {/home/johannes/Nextcloud/Zotero/Musil et al_2021_Physics-Inspired Structural Representations for Molecules and Materials.pdf} +} + +@unpublished{musilPhysicsinspiredStructuralRepresentations2021, + title = {Physics-Inspired Structural Representations for Molecules and Materials}, + author = {Musil, Felix and Grisafi, Andrea and Bartók, Albert P. and Ortner, Christoph and Csányi, Gábor and Ceriotti, Michele}, + date = {2021-05-04}, + eprint = {2101.04673}, + eprinttype = {arxiv}, + primaryclass = {physics}, + url = {http://arxiv.org/abs/2101.04673}, + urldate = {2021-05-30}, + abstract = {The first step in the construction of a regression model or a data-driven analysis framework for matter at the atomic scale involves transforming the Cartesian coordinates that describe the positions of the atoms in the form of a representation that obeys the same symmetries as the properties of interest, and in general reflects the physical nature of the problem. The link between properties, structures, their physical chemistry and their mathematical description is strongest when it comes to applications aimed at determining a precise correspondence between atomic configurations and the quantities that one might compute by a quantum mechanical electronic-structure calculation or measure experimentally. The development of atomic-scale representations have played, and continue to play, a central role in the success of machine-learning methods that rely on this correspondence, such as interatomic potentials, as well as generic property models, structural classifiers and low-dimensional maps of structures and datasets. This review summarizes the current understanding of the nature and characteristics of the most commonly used structural and chemical descriptions of molecules and materials, highlighting the deep underlying connections between different frameworks, and the ideas that lead to computationally efficient and universally applicable models. It gives examples of recent applications to a diverse set of chemical and materials science problems, and outlines the open questions and the most promising research directions in the field.}, + archiveprefix = {arXiv}, + keywords = {_tablet,ACSF,descriptor dimred,descriptors,descriptors analysis,MBTR,ML,review,SOAP}, + file = {/home/johannes/Nextcloud/Zotero/Musil et al_2021_Physics-inspired structural representations for molecules and materials.pdf;/home/johannes/Zotero/storage/EXTUHGNH/2101.html} +} + +@article{nagaosaTopologicalPropertiesDynamics2013, + title = {Topological Properties and Dynamics of Magnetic Skyrmions}, + author = {Nagaosa, Naoto and Tokura, Yoshinori}, + date = {2013-12}, + journaltitle = {Nature Nanotechnology}, + shortjournal = {Nature Nanotech}, + volume = {8}, + number = {12}, + pages = {899--911}, + publisher = {{Nature Publishing Group}}, + issn = {1748-3395}, + doi = {10.1038/nnano.2013.243}, + url = {https://www.nature.com/articles/nnano.2013.243}, + urldate = {2021-08-24}, + abstract = {This Review covers the recent developments in the observation and modelling of magnetic skyrmions, including their topological properties, current-induced dynamics and potential in future information storage devices.}, + issue = {12}, + langid = {english}, + annotation = {Bandiera\_abtest: a Cg\_type: Nature Research Journals Primary\_atype: Reviews Subject\_term: Magnetic properties and materials Subject\_term\_id: magnetic-properties-and-materials}, + file = {/home/johannes/Nextcloud/Zotero/Nagaosa_Tokura_2013_Topological properties and dynamics of magnetic skyrmions.pdf} +} + +@article{narayanAssessingSinglecellTranscriptomic2021, + title = {Assessing Single-Cell Transcriptomic Variability through Density-Preserving Data Visualization}, + author = {Narayan, Ashwin and Berger, Bonnie and Cho, Hyunghoon}, + date = {2021-01-18}, + journaltitle = {Nature Biotechnology}, + shortjournal = {Nat Biotechnol}, + pages = {1--10}, + publisher = {{Nature Publishing Group}}, + issn = {1546-1696}, + doi = {10.1038/s41587-020-00801-7}, + url = {https://www.nature.com/articles/s41587-020-00801-7}, + urldate = {2021-06-03}, + abstract = {Nonlinear data visualization methods, such as t-distributed stochastic neighbor embedding (t-SNE) and uniform manifold approximation and projection (UMAP), summarize the complex transcriptomic landscape of single cells in two dimensions or three dimensions, but they neglect the local density of data points in the original space, often resulting in misleading visualizations where densely populated subsets of cells are given more visual space than warranted by their transcriptional diversity in the dataset. Here we present den-SNE and densMAP, which are density-preserving visualization tools based on t-SNE and UMAP, respectively, and demonstrate their ability to accurately incorporate information about transcriptomic variability into the visual interpretation of single-cell RNA sequencing data. Applied to recently published datasets, our methods reveal significant changes in transcriptomic variability in a range of biological processes, including heterogeneity in transcriptomic variability of immune cells in blood and tumor, human immune cell specialization and the developmental trajectory of Caenorhabditis elegans. Our methods are readily applicable to visualizing high-dimensional data in other scientific domains.}, + langid = {english}, + keywords = {den-SNE,density-preserving,densMAP,dimensionality reduction,library,t-SNE,UMAP,unsupervised learning,visualization,with-code}, + file = {/home/johannes/Nextcloud/Zotero/Narayan et al_2021_Assessing single-cell transcriptomic variability through density-preserving.pdf;/home/johannes/Zotero/storage/72VGP9LG/s41587-020-00801-7.html} +} + +@article{narayanDensityPreservingDataVisualization2020, + title = {Density-{{Preserving Data Visualization Unveils Dynamic Patterns}} of {{Single-Cell Transcriptomic Variability}}}, + author = {Narayan, Ashwin and Berger, Bonnie and Cho, Hyunghoon}, + date = {2020-05-14}, + journaltitle = {bioRxiv}, + pages = {2020.05.12.077776}, + publisher = {{Cold Spring Harbor Laboratory}}, + doi = {10.1101/2020.05.12.077776}, + url = {https://www.biorxiv.org/content/10.1101/2020.05.12.077776v1}, + urldate = {2021-05-15}, + abstract = {{$<$}p{$>$}Nonlinear data-visualization methods, such as t-SNE and UMAP, have become staple tools for summarizing the complex transcriptomic landscape of single cells in 2D or 3D. However, existing approaches neglect the local density of data points in the original space, often resulting in misleading visualizations where densely populated subpopulations of cells are given more visual space even if they account for only a small fraction of transcriptional diversity within the dataset. We present den-SNE and densMAP, our density-preserving visualization tools based on t-SNE and UMAP, respectively, and demonstrate their ability to facilitate more accurate visual interpretation of single-cell RNA-seq data. On recently published datasets, our methods newly reveal significant changes in transcriptomic variability within a range of biological processes, including cancer, immune cell specialization in human, and the developmental trajectory of \emph{C. elegans}. Our methods are readily applicable to visualizing high-dimensional data in other scientific domains.{$<$}/p{$>$}}, + langid = {english}, + keywords = {den-SNE,density-preserving,densMAP,dimensionality reduction,library,t-SNE,UMAP,unsupervised learning,visualization,with-code}, + file = {/home/johannes/Nextcloud/Zotero/Narayan et al_2020_Density-Preserving Data Visualization Unveils Dynamic Patterns of Single-Cell.pdf;/home/johannes/Zotero/storage/6QBY65KW/2020.05.12.html} +} + +@article{nemnesFeatureSelectionProcedures2021, + title = {Feature Selection Procedures for Combined Density Functional Theory—Artificial Neural Network Schemes}, + author = {Nemnes, George Alexandru and Filipoiu, Nicolae and Sipica, Valentin}, + date = {2021-04}, + journaltitle = {Physica Scripta}, + shortjournal = {Phys. Scr.}, + volume = {96}, + number = {6}, + pages = {065807}, + publisher = {{IOP Publishing}}, + issn = {1402-4896}, + doi = {10.1088/1402-4896/abf3f7}, + url = {https://doi.org/10.1088/1402-4896/abf3f7}, + urldate = {2021-12-14}, + abstract = {We propose a workflow which includes the essential step of feature selection in order to optimize combined density functional theory—machine learning schemes (DFT-ML). Here, the energy gaps of hybrid graphene—boron nitride nanoflakes with randomly distributed domains are predicted using artificial neural networks (ANNs). The training data is obtained by associating structural information to the target quantity of interest, i.e. the energy gap, obtained by DFT calculations. The selection of proper feature vectors is important for an accurate and efficient ANN model. However, finding an optimal set of features is generally not trivial. We compare different approaches for selecting the feature vectors, ranging from random selection of the features to guided approaches like removing the features with lowest variance and by using the mutual information regression selection technique. We show that the feature selection procedures provides a significant reduction of the input space dimensionality. In addition, a selection method based on the ranking of the cutting radius is proposed and evaluated. This may not only be important for establishing optimal ANN models, but may offer insights into the minimum information required to map certain targeted properties.}, + langid = {english}, + keywords = {ANN,DFT,featurele selection,ML} +} + +@article{neupertIntroductionMachineLearning2021, + title = {Introduction to {{Machine Learning}} for the {{Sciences}}}, + author = {Neupert, Titus and Fischer, Mark H. and Greplova, Eliska and Choo, Kenny and Denner, Michael}, + date = {2021-02-08}, + url = {https://arxiv.org/abs/2102.04883v1}, + urldate = {2021-12-14}, + abstract = {This is an introductory machine learning course specifically developed with STEM students in mind. We discuss supervised, unsupervised, and reinforcement learning. The notes start with an exposition of machine learning methods without neural networks, such as principle component analysis, t-SNE, and linear regression. We continue with an introduction to both basic and advanced neural network structures such as conventional neural networks, (variational) autoencoders, generative adversarial networks, restricted Boltzmann machines, and recurrent neural networks. Questions of interpretability are discussed using the examples of dreaming and adversarial attacks.}, + langid = {english}, + keywords = {general,ML,review}, + file = {/home/johannes/Nextcloud/Zotero/Neupert et al_2021_Introduction to Machine Learning for the Sciences.pdf;/home/johannes/Zotero/storage/GE7KJ34Q/2102.html} +} + +@article{nigamEquivariantRepresentationsMolecular2022, + title = {Equivariant Representations for Molecular {{Hamiltonians}} and {{N-center}} Atomic-Scale Properties}, + author = {Nigam, Jigyasa and Willatt, Michael J. and Ceriotti, Michele}, + date = {2022-01-07}, + journaltitle = {The Journal of Chemical Physics}, + shortjournal = {J. Chem. Phys.}, + volume = {156}, + number = {1}, + pages = {014115}, + publisher = {{American Institute of Physics}}, + issn = {0021-9606}, + doi = {10.1063/5.0072784}, + url = {https://aip.scitation.org/doi/full/10.1063/5.0072784}, + urldate = {2022-10-04}, + abstract = {Symmetry considerations are at the core of the major frameworks used to provide an effective mathematical representation of atomic configurations that is then used in machine-learning models to predict the properties associated with each structure. In most cases, the models rely on a description of atom-centered environments and are suitable to learn atomic properties or global observables that can be decomposed into atomic contributions. Many quantities that are relevant for quantum mechanical calculations, however—most notably the single-particle Hamiltonian matrix when written in an atomic orbital basis—are not associated with a single center, but with two (or more) atoms in the structure. We discuss a family of structural descriptors that generalize the very successful atom-centered density correlation features to the N-center case and show, in particular, how this construction can be applied to efficiently learn the matrix elements of the (effective) single-particle Hamiltonian written in an atom-centered orbital basis. These N-center features are fully equivariant—not only in terms of translations and rotations but also in terms of permutations of the indices associated with the atoms—and are suitable to construct symmetry-adapted machine-learning models of new classes of properties of molecules and materials.}, + keywords = {ACDC,equivariant,ML,ML-DFT,ML-ESM,N-center representation,NICE,prediction of Hamiltonian matrix,with-code}, + file = {/home/johannes/Nextcloud/Zotero/Nigam et al_2022_Equivariant representations for molecular Hamiltonians and N-center.pdf} +} + +@unpublished{nigamUnifiedTheoryAtomcentered2022, + title = {Unified Theory of Atom-Centered Representations and Graph Convolutional Machine-Learning Schemes}, + author = {Nigam, Jigyasa and Fraux, Guillaume and Ceriotti, Michele}, + date = {2022-02-03}, + eprint = {2202.01566}, + eprinttype = {arxiv}, + primaryclass = {physics, stat}, + url = {http://arxiv.org/abs/2202.01566}, + urldate = {2022-02-04}, + abstract = {Data-driven schemes that associate molecular and crystal structures with their microscopic properties share the need for a concise, effective description of the arrangement of their atomic constituents. Many types of models rely on descriptions of atom-centered environments, that are associated with an atomic property or with an atomic contribution to an extensive macroscopic quantity. Frameworks in this class can be understood in terms of atom-centered density correlations (ACDC), that are used as a basis for a body-ordered, symmetry-adapted expansion of the targets. Several other schemes, that gather information on the relationship between neighboring atoms using graph-convolutional (or message-passing) ideas, cannot be directly mapped to correlations centered around a single atom. We generalize the ACDC framework to include multi-centered information, generating representations that provide a complete linear basis to regress symmetric functions of atomic coordinates, and form the basis to systematize our understanding of both atom-centered and graph-convolutional machine-learning schemes.}, + archiveprefix = {arXiv}, + keywords = {_tablet,ACDC,ACE,descriptors,GCN,GNN,ML,MPNN,NequIP,NN,representation learning,SOAP,unified theory}, + file = {/home/johannes/Nextcloud/Zotero/false;/home/johannes/Nextcloud/Zotero/Nigam et al_2022_Unified theory of atom-centered representations and graph convolutional.pdf} +} + +@article{ohCompleteQuantumHall2013, + title = {The {{Complete Quantum Hall Trio}}}, + author = {Oh, Seongshik}, + date = {2013-04-12}, + journaltitle = {Science}, + volume = {340}, + number = {6129}, + pages = {153--154}, + publisher = {{American Association for the Advancement of Science}}, + doi = {10.1126/science.1237215}, + url = {https://www.science.org/doi/10.1126/science.1237215}, + urldate = {2022-05-13}, + keywords = {_tablet,Hall effect,Hall QAHE,Hall QHE,Hall QSHE,perspective}, + file = {/home/johannes/Nextcloud/Zotero/Oh_2013_The Complete Quantum Hall Trio.pdf} +} + +@article{oliveiraCECAMElectronicStructure2020, + title = {The {{CECAM}} Electronic Structure Library and the Modular Software Development Paradigm}, + author = {Oliveira, Micael J. T. and Papior, Nick and Pouillon, Yann and Blum, Volker and Artacho, Emilio and Caliste, Damien and Corsetti, Fabiano and de Gironcoli, Stefano and Elena, Alin M. and GarcÃa, Alberto and GarcÃa-Suárez, VÃctor M. and Genovese, Luigi and Huhn, William P. and Huhs, Georg and Kokott, Sebastian and Küçükbenli, Emine and Larsen, Ask H. and Lazzaro, Alfio and Lebedeva, Irina V. and Li, Yingzhou and López-Durán, David and López-Tarifa, Pablo and Lüders, Martin and Marques, Miguel A. L. and Minar, Jan and Mohr, Stephan and Mostofi, Arash A. and O’Cais, Alan and Payne, Mike C. and Ruh, Thomas and Smith, Daniel G. A. and Soler, José M. and Strubbe, David A. and Tancogne-Dejean, Nicolas and Tildesley, Dominic and Torrent, Marc and Yu, Victor Wen-zhe}, + options = {useprefix=true}, + date = {2020-07-14}, + journaltitle = {The Journal of Chemical Physics}, + shortjournal = {J. Chem. Phys.}, + volume = {153}, + number = {2}, + pages = {024117}, + publisher = {{American Institute of Physics}}, + issn = {0021-9606}, + doi = {10.1063/5.0012901}, + url = {https://aip.scitation.org/doi/10.1063/5.0012901}, + urldate = {2021-10-26}, + abstract = {First-principles electronic structure calculations are now accessible to a very large community of users across many disciplines, thanks to many successful software packages, some of which are described in this special issue. The traditional coding paradigm for such packages is monolithic, i.e., regardless of how modular its internal structure may be, the code is built independently from others, essentially from the compiler up, possibly with the exception of linear-algebra and message-passing libraries. This model has endured and been quite successful for decades. The successful evolution of the electronic structure methodology itself, however, has resulted in an increasing complexity and an ever longer list of features expected within all software packages, which implies a growing amount of replication between different packages, not only in the initial coding but, more importantly, every time a code needs to be re-engineered to adapt to the evolution of computer hardware architecture. The Electronic Structure Library (ESL) was initiated by CECAM (the European Centre for Atomic and Molecular Calculations) to catalyze a paradigm shift away from the monolithic model and promote modularization, with the ambition to extract common tasks from electronic structure codes and redesign them as open-source libraries available to everybody. Such libraries include “heavy-duty†ones that have the potential for a high degree of parallelization and adaptation to novel hardware within them, thereby separating the sophisticated computer science aspects of performance optimization and re-engineering from the computational science done by, e.g., physicists and chemists when implementing new ideas. We envisage that this modular paradigm will improve overall coding efficiency and enable specialists (whether they be computer scientists or computational scientists) to use their skills more effectively and will lead to a more dynamic evolution of software in the community as well as lower barriers to entry for new developers. The model comes with new challenges, though. The building and compilation of a code based on many interdependent libraries (and their versions) is a much more complex task than that of a code delivered in a single self-contained package. Here, we describe the state of the ESL, the different libraries it now contains, the short- and mid-term plans for further libraries, and the way the new challenges are faced. The ESL is a community initiative into which several pre-existing codes and their developers have contributed with their software and efforts, from which several codes are already benefiting, and which remains open to the community.}, + file = {/home/johannes/Nextcloud/Zotero/Oliveira et al_2020_The CECAM electronic structure library and the modular software development.pdf} +} + +@article{onatSensitivityDimensionalityAtomic2020, + title = {Sensitivity and Dimensionality of Atomic Environment Representations Used for Machine Learning Interatomic Potentials}, + author = {Onat, Berk and Ortner, Christoph and Kermode, James R.}, + date = {2020-10-12}, + journaltitle = {The Journal of Chemical Physics}, + shortjournal = {J. Chem. Phys.}, + volume = {153}, + number = {14}, + pages = {144106}, + publisher = {{American Institute of Physics}}, + issn = {0021-9606}, + doi = {10.1063/5.0016005}, + url = {https://aip.scitation.org/doi/10.1063/5.0016005}, + urldate = {2021-05-13}, + abstract = {Faithfully representing chemical environments is essential for describing materials and molecules with machine learning approaches. Here, we present a systematic classification of these representations and then investigate (i) the sensitivity to perturbations and (ii) the effective dimensionality of a variety of atomic environment representations and over a range of material datasets. Representations investigated include atom centered symmetry functions, Chebyshev Polynomial Symmetry Functions (CHSF), smooth overlap of atomic positions, many-body tensor representation, and atomic cluster expansion. In area (i), we show that none of the atomic environment representations are linearly stable under tangential perturbations and that for CHSF, there are instabilities for particular choices of perturbation, which we show can be removed with a slight redefinition of the representation. In area (ii), we find that most representations can be compressed significantly without loss of precision and, further, that selecting optimal subsets of a representation method improves the accuracy of regression models built for a given dataset.}, + keywords = {_tablet,ACE,ACSF,descriptors,MBTR,ML,MTP,SNAP,SOAP}, + file = {/home/johannes/Nextcloud/Zotero/Onat et al_2020_Sensitivity and dimensionality of atomic environment representations used for.pdf;/home/johannes/Zotero/storage/RQ8UAKFX/5.html} +} + +@online{OnlineCourseMachine, + title = {Online {{Course}}: {{Machine Learning}} for {{Physicists}} 2021 - {{HedgeDoc}}}, + shorttitle = {Online {{Course}}}, + url = {https://pad.gwdg.de/s/Machine_Learning_For_Physicists_2021#}, + urldate = {2021-05-13}, + abstract = {\# Online Course: Machine Learning for Physicists 2021 :::info **Lecture Series by Florian Marquard}, + keywords = {course,ML,notebook,with-code}, + file = {/home/johannes/Zotero/storage/6TZCQAXX/Machine_Learning_For_Physicists_2021.html} +} + +@article{ouyangSISSOCompressedsensingMethod2018, + title = {{{SISSO}}: {{A}} Compressed-Sensing Method for Identifying the Best Low-Dimensional Descriptor in an Immensity of Offered Candidates}, + shorttitle = {{{SISSO}}}, + author = {Ouyang, Runhai and Curtarolo, Stefano and Ahmetcik, Emre and Scheffler, Matthias and Ghiringhelli, Luca M.}, + date = {2018-08-07}, + journaltitle = {Physical Review Materials}, + shortjournal = {Phys. Rev. Materials}, + volume = {2}, + number = {8}, + pages = {083802}, + publisher = {{American Physical Society}}, + doi = {10.1103/PhysRevMaterials.2.083802}, + url = {https://link.aps.org/doi/10.1103/PhysRevMaterials.2.083802}, + urldate = {2021-05-19}, + abstract = {The lack of reliable methods for identifying descriptors—the sets of parameters capturing the underlying mechanisms of a material's property—is one of the key factors hindering efficient materials development. Here, we propose a systematic approach for discovering descriptors for materials' properties, within the framework of compressed-sensing-based dimensionality reduction. The sure independence screening and sparsifying operator (SISSO) tackles immense and correlated features spaces, and converges to the optimal solution from a combination of features relevant to the materials' property of interest. In addition, SISSO gives stable results also with small training sets. The methodology is benchmarked with the quantitative prediction of the ground-state enthalpies of octet binary materials (using ab initio data) and applied to the showcase example of predicting the metal/insulator classification of binaries (with experimental data). Accurate, predictive models are found in both cases. For the metal-insulator classification model, the predictive capability is tested beyond the training data: It rediscovers the available pressure-induced insulator-to-metal transitions and it allows for the prediction of yet unknown transition candidates, ripe for experimental validation. As a step forward with respect to previous model-identification methods, SISSO can become an effective tool for automatic materials development.}, + keywords = {compressed sensing,descriptors,descriptors analysis,ML}, + file = {/home/johannes/Nextcloud/Zotero/Ouyang et al_2018_SISSO.pdf;/home/johannes/Zotero/storage/FPEWTJ64/PhysRevMaterials.2.html} +} + +@article{oviedoInterpretableExplainableMachine2022, + title = {Interpretable and {{Explainable Machine Learning}} for {{Materials Science}} and {{Chemistry}}}, + author = {Oviedo, Felipe and Ferres, Juan Lavista and Buonassisi, Tonio and Butler, Keith T.}, + date = {2022-06-24}, + journaltitle = {Accounts of Materials Research}, + shortjournal = {Acc. Mater. Res.}, + volume = {3}, + number = {6}, + pages = {597--607}, + publisher = {{American Chemical Society}}, + doi = {10.1021/accountsmr.1c00244}, + url = {https://doi.org/10.1021/accountsmr.1c00244}, + urldate = {2022-07-11}, + abstract = {ConspectusMachine learning has become a common and powerful tool in materials research. As more data become available, with the use of high-performance computing and high-throughput experimentation, machine learning has proven potential to accelerate scientific research and technology development. Though the uptake of data-driven approaches for materials science is at an exciting, early stage, to realize the true potential of machine learning models for successful scientific discovery, they must have qualities beyond purely predictive power. The predictions and inner workings of models should provide a certain degree of explainability by human experts, permitting the identification of potential model issues or limitations, building trust in model predictions, and unveiling unexpected correlations that may lead to scientific insights. In this work, we summarize applications of interpretability and explainability techniques for materials science and chemistry and discuss how these techniques can improve the outcome of scientific studies. We start by defining the fundamental concepts of interpretability and explainability in machine learning and making them less abstract by providing examples in the field. We show how interpretability in scientific machine learning has additional constraints compared to general applications. Building upon formal definitions in machine learning, we formulate the basic trade-offs among the explainability, completeness, and scientific validity of model explanations in scientific problems. In the context of these trade-offs, we discuss how interpretable models can be constructed, what insights they provide, and what drawbacks they have. We present numerous examples of the application of interpretable machine learning in a variety of experimental and simulation studies, encompassing first-principles calculations, physicochemical characterization, materials development, and integration into complex systems. We discuss the varied impacts and uses of interpretabiltiy in these cases according to the nature and constraints of the scientific study of interest. We discuss various challenges for interpretable machine learning in materials science and, more broadly, in scientific settings. In particular, we emphasize the risks of inferring causation or reaching generalization by purely interpreting machine learning models and the need for uncertainty estimates for model explanations. Finally, we showcase a number of exciting developments in other fields that could benefit interpretability in material science problems. Adding interpretability to a machine learning model often requires no more technical know-how than building the model itself. By providing concrete examples of studies (many with associated open source code and data), we hope that this Account will encourage all practitioners of machine learning in materials science to look deeper into their models.}, + keywords = {AML,ML,XAI}, + file = {/home/johannes/Nextcloud/Zotero/Oviedo et al_2022_Interpretable and Explainable Machine Learning for Materials Science and.pdf;/home/johannes/Zotero/storage/9I3JM9FX/accountsmr.html} +} + +@article{paleicoBinHashMethod2021, + title = {A Bin and Hash Method for Analyzing Reference Data and Descriptors in Machine Learning Potentials}, + author = {Paleico, MartÃn Leandro and Behler, Jörg}, + date = {2021-04}, + journaltitle = {Machine Learning: Science and Technology}, + shortjournal = {Mach. Learn.: Sci. Technol.}, + volume = {2}, + number = {3}, + pages = {037001}, + publisher = {{IOP Publishing}}, + issn = {2632-2153}, + doi = {10.1088/2632-2153/abe663}, + url = {https://doi.org/10.1088/2632-2153/abe663}, + urldate = {2021-05-03}, + abstract = {In recent years the development of machine learning potentials (MLPs) has become a very active field of research. Numerous approaches have been proposed, which allow one to perform extended simulations of large systems at a small fraction of the computational costs of electronic structure calculations. The key to the success of modern MLPs is the close-to first principles quality description of the atomic interactions. This accuracy is reached by using very flexible functional forms in combination with high-level reference data from electronic structure calculations. These data sets can include up to hundreds of thousands of structures covering millions of atomic environments to ensure that all relevant features of the potential energy surface are well represented. The handling of such large data sets is nowadays becoming one of the main challenges in the construction of MLPs. In this paper we present a method, the bin-and-hash (BAH) algorithm, to overcome this problem by enabling the efficient identification and comparison of large numbers of multidimensional vectors. Such vectors emerge in multiple contexts in the construction of MLPs. Examples are the comparison of local atomic environments to identify and avoid unnecessary redundant information in the reference data sets that is costly in terms of both the electronic structure calculations as well as the training process, the assessment of the quality of the descriptors used as structural fingerprints in many types of MLPs, and the detection of possibly unreliable data points. The BAH algorithm is illustrated for the example of high-dimensional neural network potentials using atom-centered symmetry functions for the geometrical description of the atomic environments, but the method is general and can be combined with any current type of MLP.}, + langid = {english}, + keywords = {ACSF,descriptors,descriptors analysis,ML,MLP}, + annotation = {0 citations (Crossref) [2021-05-04]}, + file = {/home/johannes/Nextcloud/Zotero/Paleico_Behler_2021_A bin and hash method for analyzing reference data and descriptors in machine.pdf} +} + +@article{parsaeifardAssessmentStructuralResolution2021, + title = {An Assessment of the Structural Resolution of Various Fingerprints Commonly Used in Machine Learning}, + author = {Parsaeifard, Behnam and De, Deb Sankar and Christensen, Anders S. and Faber, Felix A. and Kocer, Emir and De, Sandip and Behler, Jörg and von Lilienfeld, O. Anatole and Goedecker, Stefan}, + date = {2021-03}, + journaltitle = {Machine Learning: Science and Technology}, + shortjournal = {Mach. Learn.: Sci. Technol.}, + volume = {2}, + number = {1}, + pages = {015018}, + publisher = {{IOP Publishing}}, + issn = {2632-2153}, + doi = {10.1088/2632-2153/abb212}, + url = {https://doi.org/10.1088/2632-2153/abb212}, + urldate = {2021-05-18}, + abstract = {Atomic environment fingerprints are widely used in computational materials science, from machine learning potentials to the quantification of similarities between atomic configurations. Many approaches to the construction of such fingerprints, also called structural descriptors, have been proposed. In this work, we compare the performance of fingerprints based on the overlap matrix, the smooth overlap of atomic positions, Behler–Parrinello atom-centered symmetry functions, modified Behler–Parrinello symmetry functions used in the ANI-1ccx potential and the Faber–Christensen–Huang–Lilienfeld fingerprint under various aspects. We study their ability to resolve differences in local environments and in particular examine whether there are certain atomic movements that leave the fingerprints exactly or nearly invariant. For this purpose, we introduce a sensitivity matrix whose eigenvalues quantify the effect of atomic displacement modes on the fingerprint. Further, we check whether these displacements correlate with the variation of localized physical quantities such as forces. Finally, we extend our examination to the correlation between molecular fingerprints obtained from the atomic fingerprints and global quantities of entire molecules.}, + langid = {english}, + keywords = {ACSF,BPSF,descriptor comparison,descriptors,FCHL,ML,SOAP}, + file = {/home/johannes/Nextcloud/Zotero/Parsaeifard et al_2021_An assessment of the structural resolution of various fingerprints commonly.pdf} +} + +@article{parsaeifardManifoldsQuasiconstantSOAP2022, + title = {Manifolds of Quasi-Constant {{SOAP}} and {{ACSF}} Fingerprints and the Resulting Failure to Machine Learn Four-Body Interactions}, + author = {Parsaeifard, Behnam and Goedecker, Stefan}, + date = {2022-01-21}, + journaltitle = {The Journal of Chemical Physics}, + shortjournal = {J. Chem. Phys.}, + volume = {156}, + number = {3}, + pages = {034302}, + publisher = {{American Institute of Physics}}, + issn = {0021-9606}, + doi = {10.1063/5.0070488}, + url = {https://aip.scitation.org/doi/full/10.1063/5.0070488}, + urldate = {2022-09-20}, + abstract = {Atomic fingerprints are commonly used for the characterization of local environments of atoms in machine learning and other contexts. In this work, we study the behavior of two widely used fingerprints, namely, the smooth overlap of atomic positions (SOAP) and the atom-centered symmetry functions (ACSFs), under finite changes of atomic positions and demonstrate the existence of manifolds of quasi-constant fingerprints. These manifolds are found numerically by following eigenvectors of the sensitivity matrix with quasi-zero eigenvalues. The existence of such manifolds in ACSF and SOAP causes a failure to machine learn four-body interactions, such as torsional energies that are part of standard force fields. No such manifolds can be found for the overlap matrix (OM) fingerprint due to its intrinsic many-body character.}, + keywords = {ACSF,descriptors,descriptors analysis,incompleteness,ML,SOAP}, + file = {/home/johannes/Nextcloud/Zotero/Parsaeifard_Goedecker_2022_Manifolds of quasi-constant SOAP and ACSF fingerprints and the resulting.pdf} +} + +@unpublished{pedersonMachineLearningDensity2022, + title = {Machine Learning and Density Functional Theory}, + author = {Pederson, Ryan and Kalita, Bhupalee and Burke, Kieron}, + date = {2022-05-03}, + eprint = {2205.01591}, + eprinttype = {arxiv}, + primaryclass = {cond-mat, physics:physics}, + url = {http://arxiv.org/abs/2205.01591}, + urldate = {2022-05-13}, + abstract = {Over the past decade machine learning has made significant advances in approximating density functionals, but whether this signals the end of human-designed functionals remains to be seen. Ryan Pederson, Bhupalee Kalita and Kieron Burke discuss the rise of machine learning for functional design.}, + archiveprefix = {arXiv}, + keywords = {DeepMind,density functional,DFT,DM21,ML,ML-DFT,ML-ESM}, + file = {/home/johannes/Nextcloud/Zotero/Pederson et al_2022_Machine learning and density functional theory.pdf;/home/johannes/Zotero/storage/UPT9RJEW/2205.html} +} + +@unpublished{pedregosaScikitlearnMachineLearning2018, + title = {Scikit-Learn: {{Machine Learning}} in {{Python}}}, + shorttitle = {Scikit-Learn}, + author = {Pedregosa, Fabian and Varoquaux, Gaël and Gramfort, Alexandre and Michel, Vincent and Thirion, Bertrand and Grisel, Olivier and Blondel, Mathieu and Müller, Andreas and Nothman, Joel and Louppe, Gilles and Prettenhofer, Peter and Weiss, Ron and Dubourg, Vincent and Vanderplas, Jake and Passos, Alexandre and Cournapeau, David and Brucher, Matthieu and Perrot, Matthieu and Duchesnay, Édouard}, + date = {2018-06-05}, + eprint = {1201.0490}, + eprinttype = {arxiv}, + primaryclass = {cs}, + url = {http://arxiv.org/abs/1201.0490}, + urldate = {2021-07-14}, + abstract = {Scikit-learn is a Python module integrating a wide range of state-of-the-art machine learning algorithms for medium-scale supervised and unsupervised problems. This package focuses on bringing machine learning to non-specialists using a general-purpose high-level language. Emphasis is put on ease of use, performance, documentation, and API consistency. It has minimal dependencies and is distributed under the simplified BSD license, encouraging its use in both academic and commercial settings. Source code, binaries, and documentation can be downloaded from http://scikit-learn.org.}, + archiveprefix = {arXiv}, + keywords = {library,ML,scikit-learn}, + file = {/home/johannes/Nextcloud/Zotero/Pedregosa et al_2018_Scikit-learn.pdf;/home/johannes/Zotero/storage/RDIVNKI6/1201.html} +} + +@article{peixotoNonlocalEffectImpurity2020, + title = {Non-Local Effect of Impurity States on the Exchange Coupling Mechanism in Magnetic Topological Insulators}, + author = {Peixoto, Thiago R. F. and Bentmann, Hendrik and Rüßmann, Philipp and Tcakaev, Abdul-Vakhab and Winnerlein, Martin and Schreyeck, Steffen and Schatz, Sonja and Vidal, Raphael Crespo and Stier, Fabian and Zabolotnyy, Volodymyr and Green, Robert J. and Min, Chul Hee and Fornari, Celso I. and Maaß, Henriette and Vasili, Hari Babu and Gargiani, Pierluigi and Valvidares, Manuel and Barla, Alessandro and Buck, Jens and Hoesch, Moritz and Diekmann, Florian and Rohlf, Sebastian and Kalläne, Matthias and Rossnagel, Kai and Gould, Charles and Brunner, Karl and Blügel, Stefan and Hinkov, Vladimir and Molenkamp, Laurens W. and Reinert, Friedrich}, + date = {2020-11-19}, + journaltitle = {npj Quantum Materials}, + shortjournal = {npj Quantum Mater.}, + volume = {5}, + number = {1}, + pages = {1--6}, + publisher = {{Nature Publishing Group}}, + issn = {2397-4648}, + doi = {10.1038/s41535-020-00288-0}, + url = {https://www.nature.com/articles/s41535-020-00288-0}, + urldate = {2022-10-05}, + abstract = {Since the discovery of the quantum anomalous Hall (QAH) effect in the magnetically doped topological insulators (MTI) Cr:(Bi,Sb)2Te3 and V:(Bi,Sb)2Te3, the search for the magnetic coupling mechanisms underlying the onset of ferromagnetism has been a central issue, and a variety of different scenarios have been put forward. By combining resonant photoemission, X-ray magnetic circular dichroism and density functional theory, we determine the local electronic and magnetic configurations of V and Cr impurities in (Bi,Sb)2Te3. State-of-the-art first-principles calculations find pronounced differences in their 3d densities of states, and show how these impurity states mediate characteristic short-range pd exchange interactions, whose strength sensitively varies with the position of the 3d states relative to the Fermi level. Measurements on films with varying host stoichiometry support this trend. Our results explain, in an unified picture, the origins of the observed magnetic properties, and establish the essential role of impurity-state-mediated exchange interactions in the magnetism of MTI.}, + issue = {1}, + langid = {english}, + keywords = {Electronic properties and materials,Ferromagnetism,Hall QAHE,Magnetic properties and materials,ruess,Spintronics,topological insulator,Topological insulators}, + file = {/home/johannes/Nextcloud/Zotero/Peixoto et al_2020_Non-local effect of impurity states on the exchange coupling mechanism in.pdf;/home/johannes/Zotero/storage/DDIQNTSB/s41535-020-00288-0.html} +} + +@article{pereiraChallengesTopologicalInsulator2021, + title = {Challenges of {{Topological Insulator Research}}: {{Bi2Te3 Thin Films}} and {{Magnetic Heterostructures}}}, + shorttitle = {Challenges of {{Topological Insulator Research}}}, + author = {Pereira, Vanda M. and Wu, Chi-Nan and Höfer, Katharina and Choa, Arnold and Knight, Cariad-A. and Swanson, Jesse and Becker, Christoph and Komarek, Alexander C. and Rata, A. Diana and Rößler, Sahana and Wirth, Steffen and Guo, Mengxin and Hong, Minghwei and Kwo, Jueinai and Tjeng, Liu Hao and Altendorf, Simone G.}, + date = {2021}, + journaltitle = {physica status solidi (b)}, + volume = {258}, + number = {1}, + pages = {2000346}, + issn = {1521-3951}, + doi = {10.1002/pssb.202000346}, + url = {https://onlinelibrary.wiley.com/doi/abs/10.1002/pssb.202000346}, + urldate = {2022-05-30}, + abstract = {Topological insulators (TIs) are of particular interest in the recent solid-state research because of their exceptional features stemming from the conducting, topologically protected surface states. The exotic properties include the occurrence of novel quantum phenomena and make them promising materials for spintronics and quantum computing applications. Theoretical studies have provided a vast amount of valuable predictions and proposals, whose experimental observation and implementation, to date, are often hindered by an insufficient sample quality. The effect of even a relatively low concentration of defects can make the access to purely topological surface states impossible. This points out the need of high-quality bulk-insulating materials with ultra-clean surfaces/interfaces, which requires sophisticated sample/device preparations as well as special precautions during the measurements. Herein, the challenging work on 3D TI thin films with a focus on Bi2Te3 is reported. It covers the optimization of the molecular beam epitaxy growth process, the in situ characterization of surface states and transport properties, the influence of exposure to ambient gases and of capping layers, as well as the effect of interfacing TI thin film with magnetic materials.}, + langid = {english}, + keywords = {angle-resolved photoelectron spectroscopy,in situ transport,molecular beam epitaxy,topological insulators}, + annotation = {\_eprint: https://onlinelibrary.wiley.com/doi/pdf/10.1002/pssb.202000346}, + file = {/home/johannes/Nextcloud/Zotero/Pereira et al_2021_Challenges of Topological Insulator Research.pdf;/home/johannes/Zotero/storage/CXAHTBAM/pssb.html} +} + +@inproceedings{pezoaFoundationsJSONSchema2016, + title = {Foundations of {{JSON Schema}}}, + booktitle = {Proceedings of the 25th {{International Conference}} on {{World Wide Web}}}, + author = {Pezoa, Felipe and Reutter, Juan L. and Suarez, Fernando and Ugarte, MartÃn and VrgoÄ, Domagoj}, + date = {2016-04-11}, + series = {{{WWW}} '16}, + pages = {263--273}, + publisher = {{International World Wide Web Conferences Steering Committee}}, + location = {{Republic and Canton of Geneva, CHE}}, + doi = {10.1145/2872427.2883029}, + url = {https://doi.org/10.1145/2872427.2883029}, + urldate = {2021-10-17}, + abstract = {JSON -- the most popular data format for sending API requests and responses -- is still lacking a standardized schema or meta-data definition that allows the developers to specify the structure of JSON documents. JSON Schema is an attempt to provide a general purpose schema language for JSON, but it is still work in progress, and the formal specification has not yet been agreed upon. Why this could be a problem becomes evident when examining the behaviour of numerous tools for validating JSON documents against this initial schema proposal: although they agree on most general cases, when presented with the greyer areas of the specification they tend to differ significantly. In this paper we provide the first formal definition of syntax and semantics for JSON Schema and use it to show that implementing this layer on top of JSON is feasible in practice. This is done both by analysing the theoretical aspects of the validation problem and by showing how to set up and validate a JSON Schema for Wikidata, the central storage for Wikimedia.}, + isbn = {978-1-4503-4143-1}, + keywords = {expressiveness of schema languages,JSON,JSON schema,JSON validation}, + file = {/home/johannes/Nextcloud/Zotero/Pezoa et al_2016_Foundations of JSON Schema.pdf} +} + +@article{pfauInitioSolutionManyelectron2020, + title = {Ab Initio Solution of the Many-Electron {{Schrödinger}} Equation with Deep Neural Networks}, + author = {Pfau, David}, + date = {2020}, + journaltitle = {Physical Review Research}, + shortjournal = {Phys. Rev. Research}, + volume = {2}, + number = {3}, + doi = {10.1103/PhysRevResearch.2.033429}, + keywords = {DeepMind,FermiNet,JAX,library,MC,ML,ML-ESM,ML-QMBP,NN,PauliNet,prediction of wavefunction,QMC,VMC,with-code}, + file = {/home/johannes/Nextcloud/Zotero/Pfau_2020_iAb initio-i solution of the many-electron Schrödinger equation with deep.pdf;/home/johannes/Zotero/storage/7HFHVNYZ/PhysRevResearch.2.html} +} + +@article{pilaniaDataBasedMethodsMaterials2020, + title = {Data-{{Based Methods}} for {{Materials Design}} and {{Discovery}}: {{Basic Ideas}} and {{General Methods}}}, + shorttitle = {Data-{{Based Methods}} for {{Materials Design}} and {{Discovery}}}, + author = {Pilania, Ghanshyam and Balachandran, Prasanna V. and Gubernatis, James E. and Lookman, Turab}, + date = {2020-03-06}, + journaltitle = {Synthesis Lectures on Materials and Optics}, + volume = {1}, + number = {1}, + pages = {1--188}, + publisher = {{Morgan \& Claypool Publishers}}, + issn = {2691-1930}, + doi = {10.2200/S00981ED1V01Y202001MOP001}, + url = {https://www.morganclaypool.com/doi/10.2200/S00981ED1V01Y202001MOP001}, + urldate = {2021-05-19}, + keywords = {book,materials informatics}, + file = {/home/johannes/Nextcloud/Zotero/Pilania et al_2020_Data-Based Methods for Materials Design and Discovery.pdf;/home/johannes/Zotero/storage/8YQEF8LU/S00981ED1V01Y202001MOP001.html} +} + +@article{poelkingBenchMLExtensiblePipelining2022, + title = {{{BenchML}}: An Extensible Pipelining Framework for Benchmarking Representations of Materials and Molecules at Scale}, + shorttitle = {{{BenchML}}}, + author = {Poelking, Carl and Faber, Felix and Cheng, Bingqing}, + date = {2022}, + journaltitle = {Machine Learning: Science and Technology}, + shortjournal = {Mach. Learn.: Sci. Technol.}, + issn = {2632-2153}, + doi = {10.1088/2632-2153/ac4d11}, + url = {http://iopscience.iop.org/article/10.1088/2632-2153/ac4d11}, + urldate = {2022-05-09}, + abstract = {We introduce a machine-learning (ML) framework for high-throughput benchmarking of diverse representations of chemical systems against datasets of materials and molecules. The guiding principle underlying the benchmarking approach is to evaluate raw descriptor performance by limiting model complexity to simple regression schemes while enforcing best ML practices, allowing for unbiased hyperparameter optimization, and assessing learning progress through learning curves along series of synchronized train-test splits. The resulting models are intended as baselines that can inform future method development, in addition to indicating how easily a given dataset can be learnt. Through a comparative analysis of the training outcome across a diverse set of physicochemical, topological and geometric representations, we glean insight into the relative merits of these representations as well as their interrelatedness.}, + langid = {english}, + keywords = {descriptor comparison}, + file = {/home/johannes/Nextcloud/Zotero/Poelking et al_2022_BenchML.pdf} +} + +@online{PossibleGameChanger, + title = {A Possible Game Changer for next Generation Microelectronics | {{Argonne National Laboratory}}}, + url = {https://www.anl.gov/article/a-possible-game-changer-for-next-generation-microelectronics}, + urldate = {2022-12-07}, + langid = {english}, + keywords = {/unread,condensed matter,physics,popular science,skyrmions,Spintronics}, + file = {/home/johannes/Zotero/storage/9GDZNCPD/a-possible-game-changer-for-next-generation-microelectronics.html} +} + +@article{pozdnyakovIncompletenessAtomicStructure2020, + title = {Incompleteness of {{Atomic Structure Representations}}}, + author = {Pozdnyakov, Sergey N. and Willatt, Michael J. and Bartók, Albert P. and Ortner, Christoph and Csányi, Gábor and Ceriotti, Michele}, + date = {2020-10-12}, + journaltitle = {Physical Review Letters}, + shortjournal = {Phys. Rev. Lett.}, + volume = {125}, + number = {16}, + pages = {166001}, + publisher = {{American Physical Society}}, + doi = {10.1103/PhysRevLett.125.166001}, + url = {https://link.aps.org/doi/10.1103/PhysRevLett.125.166001}, + urldate = {2021-05-13}, + abstract = {Many-body descriptors are widely used to represent atomic environments in the construction of machine-learned interatomic potentials and more broadly for fitting, classification, and embedding tasks on atomic structures. There is a widespread belief in the community that three-body correlations are likely to provide an overcomplete description of the environment of an atom. We produce several counterexamples to this belief, with the consequence that any classifier, regression, or embedding model for atom-centered properties that uses three- (or four)-body features will incorrectly give identical results for different configurations. Writing global properties (such as total energies) as a sum of many atom-centered contributions mitigates the impact of this fundamental deficiency—explaining the success of current “machine-learning†force fields. We anticipate the issues that will arise as the desired accuracy increases, and suggest potential solutions.}, + keywords = {_tablet,3-body order descriptors,descriptors,descriptors analysis,GPR,incompleteness,MBTR,ML,SOAP}, + file = {/home/johannes/Nextcloud/Zotero/Pozdnyakov et al_2020_Incompleteness of Atomic Structure Representations.pdf;/home/johannes/Zotero/storage/5QHMC4CR/PhysRevLett.125.html} +} + +@article{pozdnyakovIncompletenessGraphConvolutional2022, + title = {Incompleteness of Graph Convolutional Neural Networks for Points Clouds in Three Dimensions}, + author = {Pozdnyakov, Sergey N. and Ceriotti, Michele}, + date = {2022-01-18}, + doi = {10.48550/arXiv.2201.07136}, + url = {https://arxiv.org/abs/2201.07136v3}, + urldate = {2022-10-04}, + abstract = {Graph neural networks (GNN) are very popular methods in machine learning and have been applied very successfully to the prediction of the properties of molecules and materials. First-order GNNs are well known to be incomplete, i.e., there exist graphs that are distinct but appear identical when seen through the lens of the GNN. More complicated schemes have thus been designed to increase their resolving power. Applications to molecules (and more generally, point clouds), however, add a geometric dimension to the problem. The most straightforward and prevalent approach to construct graph representation for molecules regards atoms as vertices in a graph and draws a bond between each pair of atoms within a chosen cutoff. Bonds can be decorated with the distance between atoms, and the resulting "distance graph NNs" (dGNN) have empirically demonstrated excellent resolving power and are widely used in chemical ML, with all known indistinguishable graphs being resolved in the fully-connected limit. Here we show that even for the restricted case of fully-connected graphs induced by 3D atom clouds dGNNs are not complete. We construct pairs of distinct point clouds that generate graphs that, for any cutoff radius, are equivalent based on a first-order Weisfeiler-Lehman test. This class of degenerate structures includes chemically-plausible configurations, setting an ultimate limit to the expressive power of some of the well-established GNN architectures for atomistic machine learning. Models that explicitly use angular or directional information in the description of atomic environments can resolve these degeneracies.}, + langid = {english}, + keywords = {GCN,GNN,incompleteness,ML,MPNN,WL test}, + file = {/home/johannes/Nextcloud/Zotero/Pozdnyakov_Ceriotti_2022_Incompleteness of graph convolutional neural networks for points clouds in.pdf;/home/johannes/Zotero/storage/ZKHDUH3X/2201.html} +} + +@article{prodanNearsightednessElectronicMatter2005, + title = {Nearsightedness of Electronic Matter}, + author = {Prodan, E. and Kohn, W.}, + date = {2005-08-16}, + journaltitle = {Proceedings of the National Academy of Sciences}, + volume = {102}, + number = {33}, + pages = {11635--11638}, + publisher = {{Proceedings of the National Academy of Sciences}}, + doi = {10.1073/pnas.0505436102}, + url = {https://www.pnas.org/doi/full/10.1073/pnas.0505436102}, + urldate = {2022-10-05}, + keywords = {condensed matter,electronic structure,near-sightedness,NEM,original publication,physics}, + file = {/home/johannes/Nextcloud/Zotero/Prodan_Kohn_2005_Nearsightedness of electronic matter.pdf} +} + +@book{QuantumTheoryMagnetism, + title = {Quantum {{Theory}} of {{Magnetism}}}, + url = {https://link.springer.com/book/10.1007/978-3-540-85416-6}, + urldate = {2022-06-18}, + langid = {english}, + keywords = {_tablet,condensed matter,graduate,magnetism,textbook}, + file = {/home/johannes/Nextcloud/Zotero/Quantum Theory of Magnetism.pdf;/home/johannes/Zotero/storage/ULV44ULF/978-3-540-85416-6.html} +} + +@unpublished{rackersCrackingQuantumScaling2022, + title = {Cracking the {{Quantum Scaling Limit}} with {{Machine Learned Electron Densities}}}, + author = {Rackers, Joshua A. and Tecot, Lucas and Geiger, Mario and Smidt, Tess E.}, + date = {2022-02-10}, + number = {arXiv:2201.03726}, + eprint = {2201.03726}, + eprinttype = {arxiv}, + primaryclass = {cond-mat, physics:physics}, + publisher = {{arXiv}}, + doi = {10.48550/arXiv.2201.03726}, + url = {http://arxiv.org/abs/2201.03726}, + urldate = {2022-05-18}, + abstract = {A long-standing goal of science is to accurately solve the Schr\textbackslash "odinger equation for large molecular systems. The poor scaling of current quantum chemistry algorithms on classical computers imposes an effective limit of about a few dozen atoms for which we can calculate molecular electronic structure. We present a machine learning (ML) method to break through this scaling limit and make quantum chemistry calculations of very large systems possible. We show that Euclidean Neural Networks can be trained to predict the electron density with high fidelity from limited data. Learning the electron density allows us to train a machine learning model on small systems and make accurate predictions on large ones. We show that this ML electron density model can break through the quantum scaling limit and calculate the electron density of systems of thousands of atoms with quantum accuracy.}, + archiveprefix = {arXiv}, + keywords = {_tablet,charge density,e3nn,ENN,ML,ML-DFT,ML-ESM,molecules,prediction of electron density,script,target: density,transfer learning,with-code}, + file = {/home/johannes/Nextcloud/Zotero/Rackers et al_2022_Cracking the Quantum Scaling Limit with Machine Learned Electron Densities.pdf;/home/johannes/Zotero/storage/X9XGJLLI/2201.html} +} + +@misc{rackersCrackingQuantumScaling2022a, + title = {Cracking the {{Quantum Scaling Limit}} with {{Machine Learned Electron Densities}}}, + author = {Rackers, Joshua A. and Tecot, Lucas and Geiger, Mario and Smidt, Tess E.}, + date = {2022-02-10}, + number = {arXiv:2201.03726}, + eprint = {2201.03726}, + eprinttype = {arxiv}, + primaryclass = {cond-mat, physics:physics}, + publisher = {{arXiv}}, + doi = {10.48550/arXiv.2201.03726}, + url = {http://arxiv.org/abs/2201.03726}, + urldate = {2022-07-10}, + abstract = {A long-standing goal of science is to accurately solve the Schr\textbackslash "odinger equation for large molecular systems. The poor scaling of current quantum chemistry algorithms on classical computers imposes an effective limit of about a few dozen atoms for which we can calculate molecular electronic structure. We present a machine learning (ML) method to break through this scaling limit and make quantum chemistry calculations of very large systems possible. We show that Euclidean Neural Networks can be trained to predict the electron density with high fidelity from limited data. Learning the electron density allows us to train a machine learning model on small systems and make accurate predictions on large ones. We show that this ML electron density model can break through the quantum scaling limit and calculate the electron density of systems of thousands of atoms with quantum accuracy.}, + archiveprefix = {arXiv}, + keywords = {Condensed Matter - Soft Condensed Matter,Physics - Biological Physics,Physics - Chemical Physics}, + file = {/home/johannes/Nextcloud/Zotero/Rackers et al_2022_Cracking the Quantum Scaling Limit with Machine Learned Electron Densities2.pdf;/home/johannes/Zotero/storage/NL7QJTKF/2201.html} +} + +@article{raderTopologicalInsulatorsMaterials2021, + title = {Topological {{Insulators}}: {{Materials}} – {{Fundamental Properties}} – {{Devices}}}, + shorttitle = {Topological {{Insulators}}}, + author = {Rader, Oliver and Bihlmayer, Gustav and Fischer, Saskia F.}, + date = {2021}, + journaltitle = {physica status solidi (b)}, + volume = {258}, + number = {1}, + pages = {2170010}, + issn = {1521-3951}, + doi = {10.1002/pssb.202170010}, + url = {https://onlinelibrary.wiley.com/doi/abs/10.1002/pssb.202170010}, + urldate = {2022-05-30}, + abstract = {Topological insulators are materials that are electrically insulating in the bulk but can conduct electricity due to topologically protected electronic edge or surface states. Since 2013, the German Research Foundation (DFG) has been supporting the Priority Program “Topological Insulators: Materials – Fundamental Properties – Devices†(SPP 1666) and in the time since, topological insulators developed from a mere curiosity to a material class that entered many fields of applied research. This Special Issue presents in 20 articles reports of the Priority Program reflecting its three areas of activity: (i) Understanding and improvement of existing topological insulator materials, regarding the size of the band gap and intrinsic doping levels, to enable room temperature applications, (ii) exploration of fundamental properties necessary for the development of device structures, and (iii) discovery of new materials to overcome deficits of current materials and explore new properties. See also the Guest Editorial (article number 2000594). The cover shows a Bi2Te3/MnBi2Te4 heterostructure, a ferromagnetic topological insulator (data by M. Albu, H. Groiss, S. Wimmer, G. Kothleitner, O. Caha, and J. MichaliÄka; artwork by E. D. L. Rienks and O. Rader).}, + annotation = {\_eprint: https://onlinelibrary.wiley.com/doi/pdf/10.1002/pssb.202170010}, + file = {/home/johannes/Nextcloud/Zotero/Rader et al_2021_Topological Insulators.pdf;/home/johannes/Zotero/storage/3TIJLTAF/pssb.html} +} + +@article{raderTopologicalInsulatorsMaterials2021a, + title = {Topological {{Insulators}}: {{Materials}} – {{Fundamental Properties}} – {{Devices}}}, + shorttitle = {Topological {{Insulators}}}, + author = {Rader, Oliver and Bihlmayer, Gustav and Fischer, Saskia F.}, + date = {2021}, + journaltitle = {physica status solidi (b)}, + volume = {258}, + number = {1}, + pages = {2000594}, + issn = {1521-3951}, + doi = {10.1002/pssb.202000594}, + url = {https://onlinelibrary.wiley.com/doi/abs/10.1002/pssb.202000594}, + urldate = {2022-05-30}, + langid = {english}, + annotation = {\_eprint: https://onlinelibrary.wiley.com/doi/pdf/10.1002/pssb.202000594}, + file = {/home/johannes/Nextcloud/Zotero/Rader et al_2021_Topological Insulators2.pdf;/home/johannes/Zotero/storage/CXY5KUXP/pssb.html} +} + +@article{raissiPhysicsinformedNeuralNetworks2019, + title = {Physics-Informed Neural Networks: {{A}} Deep Learning Framework for Solving Forward and Inverse Problems Involving Nonlinear Partial Differential Equations}, + shorttitle = {Physics-Informed Neural Networks}, + author = {Raissi, M. and Perdikaris, P. and Karniadakis, G. E.}, + date = {2019-02-01}, + journaltitle = {Journal of Computational Physics}, + shortjournal = {Journal of Computational Physics}, + volume = {378}, + pages = {686--707}, + issn = {0021-9991}, + doi = {10.1016/j.jcp.2018.10.045}, + url = {https://www.sciencedirect.com/science/article/pii/S0021999118307125}, + urldate = {2022-03-23}, + abstract = {We introduce physics-informed neural networks – neural networks that are trained to solve supervised learning tasks while respecting any given laws of physics described by general nonlinear partial differential equations. In this work, we present our developments in the context of solving two main classes of problems: data-driven solution and data-driven discovery of partial differential equations. Depending on the nature and arrangement of the available data, we devise two distinct types of algorithms, namely continuous time and discrete time models. The first type of models forms a new family of data-efficient spatio-temporal function approximators, while the latter type allows the use of arbitrarily accurate implicit Runge–Kutta time stepping schemes with unlimited number of stages. The effectiveness of the proposed framework is demonstrated through a collection of classical problems in fluids, quantum mechanics, reaction–diffusion systems, and the propagation of nonlinear shallow-water waves.}, + langid = {english}, + keywords = {Nonlinear dynamics,original publication,PINN,Python,rec-by-bluegel,Runge–Kutta methods,with-code}, + file = {/home/johannes/Nextcloud/Zotero/Raissi et al_2019_Physics-informed neural networks.pdf;/home/johannes/Zotero/storage/YR4ICSGZ/S0021999118307125.html} +} + +@article{ramakrishnanBigDataMeets2015, + title = {Big {{Data Meets Quantum Chemistry Approximations}}: {{The Δ-Machine Learning Approach}}}, + shorttitle = {Big {{Data Meets Quantum Chemistry Approximations}}}, + author = {Ramakrishnan, Raghunathan and Dral, Pavlo O. and Rupp, Matthias and von Lilienfeld, O. Anatole}, + options = {useprefix=true}, + date = {2015-05-12}, + journaltitle = {Journal of Chemical Theory and Computation}, + shortjournal = {J. Chem. Theory Comput.}, + volume = {11}, + number = {5}, + pages = {2087--2096}, + publisher = {{American Chemical Society}}, + issn = {1549-9618}, + doi = {10.1021/acs.jctc.5b00099}, + url = {https://doi.org/10.1021/acs.jctc.5b00099}, + urldate = {2022-09-29}, + abstract = {Chemically accurate and comprehensive studies of the virtual space of all possible molecules are severely limited by the computational cost of quantum chemistry. We introduce a composite strategy that adds machine learning corrections to computationally inexpensive approximate legacy quantum methods. After training, highly accurate predictions of enthalpies, free energies, entropies, and electron correlation energies are possible, for significantly larger molecular sets than used for training. For thermochemical properties of up to 16k isomers of C7H10O2 we present numerical evidence that chemical accuracy can be reached. We also predict electron correlation energy in post Hartree–Fock methods, at the computational cost of Hartree–Fock, and we establish a qualitative relationship between molecular entropy and electron correlation. The transferability of our approach is demonstrated, using semiempirical quantum chemistry and machine learning models trained on 1 and 10\% of 134k organic molecules, to reproduce enthalpies of all remaining molecules at density functional theory level of accuracy.}, + keywords = {delta learning,Δ-machine learning}, + file = {/home/johannes/Nextcloud/Zotero/Ramakrishnan et al_2015_Big Data Meets Quantum Chemistry Approximations.pdf} +} + +@unpublished{ramsundarDifferentiablePhysicsPosition2021, + title = {Differentiable {{Physics}}: {{A Position Piece}}}, + shorttitle = {Differentiable {{Physics}}}, + author = {Ramsundar, Bharath and Krishnamurthy, Dilip and Viswanathan, Venkatasubramanian}, + date = {2021-09-14}, + number = {arXiv:2109.07573}, + eprint = {2109.07573}, + eprinttype = {arxiv}, + primaryclass = {physics}, + publisher = {{arXiv}}, + doi = {10.48550/arXiv.2109.07573}, + url = {http://arxiv.org/abs/2109.07573}, + urldate = {2022-05-18}, + abstract = {Differentiable physics provides a new approach for modeling and understanding the physical systems by pairing the new technology of differentiable programming with classical numerical methods for physical simulation. We survey the rapidly growing literature of differentiable physics techniques and highlight methods for parameter estimation, learning representations, solving differential equations, and developing what we call scientific foundation models using data and inductive priors. We argue that differentiable physics offers a new paradigm for modeling physical phenomena by combining classical analytic solutions with numerical methodology using the bridge of differentiable programming.}, + archiveprefix = {arXiv}, + keywords = {autodiff,ML,physics-informed ML}, + file = {/home/johannes/Nextcloud/Zotero/Ramsundar et al_2021_Differentiable Physics.pdf;/home/johannes/Zotero/storage/RGUHZPWB/2109.html} +} + +@article{reiserGraphNeuralNetworks2021, + title = {Graph Neural Networks in {{TensorFlow-Keras}} with {{RaggedTensor}} Representation (Kgcnn)}, + author = {Reiser, Patrick and Eberhard, André and Friederich, Pascal}, + date = {2021-08-01}, + journaltitle = {Software Impacts}, + shortjournal = {Software Impacts}, + volume = {9}, + pages = {100095}, + issn = {2665-9638}, + doi = {10.1016/j.simpa.2021.100095}, + url = {https://www.sciencedirect.com/science/article/pii/S266596382100035X}, + urldate = {2021-07-21}, + abstract = {Graph neural networks are a versatile machine learning architecture that received a lot of attention recently due to its wide range of applications. In this technical report, we present an implementation of graph convolution and graph pooling layers for TensorFlow-Keras models, which allows a seamless and flexible integration into standard Keras layers to set up graph models in a functional way. We developed the Keras Graph Convolutional Neural Network Python package kgcnn based on TensorFlow-Keras which focus on a transparent tensor structure passed between layers and an ease-of-use mindset.}, + langid = {english}, + keywords = {GCN,GNN,keras,library,ML,models,SchNet,with-code}, + file = {/home/johannes/Nextcloud/Zotero/Reiser et al_2021_Graph neural networks in TensorFlow-Keras with RaggedTensor representation.pdf;/home/johannes/Zotero/storage/6KMNG399/S266596382100035X.html} +} + +@misc{reiserGraphNeuralNetworks2022, + title = {Graph Neural Networks for Materials Science and Chemistry}, + author = {Reiser, Patrick and Neubert, Marlen and Eberhard, André and Torresi, Luca and Zhou, Chen and Shao, Chen and Metni, Houssam and van Hoesel, Clint and Schopmans, Henrik and Sommer, Timo and Friederich, Pascal}, + options = {useprefix=true}, + date = {2022-08-05}, + number = {arXiv:2208.09481}, + eprint = {2208.09481}, + eprinttype = {arxiv}, + primaryclass = {cond-mat, physics:physics}, + publisher = {{arXiv}}, + doi = {10.48550/arXiv.2208.09481}, + url = {http://arxiv.org/abs/2208.09481}, + urldate = {2022-09-27}, + abstract = {Machine learning plays an increasingly important role in many areas of chemistry and materials science, e.g. to predict materials properties, to accelerate simulations, to design new materials, and to predict synthesis routes of new materials. Graph neural networks (GNNs) are one of the fastest growing classes of machine learning models. They are of particular relevance for chemistry and materials science, as they directly work on a graph or structural representation of molecules and materials and therefore have full access to all relevant information required to characterize materials. In this review article, we provide an overview of the basic principles of GNNs, widely used datasets, and state-of-the-art architectures, followed by a discussion of a wide range of recent applications of GNNs in chemistry and materials science, and concluding with a road-map for the further development and application of GNNs.}, + archiveprefix = {arXiv}, + keywords = {_tablet,GCN,GNN,molecules,review,solids}, + file = {/home/johannes/Nextcloud/Zotero/Reiser et al_2022_Graph neural networks for materials science and chemistry.pdf;/home/johannes/Zotero/storage/IVEGXDHZ/2208.html} +} + +@article{RiseQuantumMaterials2016, + title = {The Rise of Quantum Materials}, + date = {2016-02}, + journaltitle = {Nature Physics}, + shortjournal = {Nature Phys}, + volume = {12}, + number = {2}, + pages = {105--105}, + publisher = {{Nature Publishing Group}}, + issn = {1745-2481}, + doi = {10.1038/nphys3668}, + url = {https://www.nature.com/articles/nphys3668}, + urldate = {2021-08-24}, + abstract = {Emergent phenomena are common in condensed matter. Their study now extends beyond strongly correlated electron systems, giving rise to the broader concept of quantum materials.}, + issue = {2}, + langid = {english}, + keywords = {quantum materials,review}, + annotation = {Bandiera\_abtest: a Cg\_type: Nature Research Journals Primary\_atype: Editorial Subject\_term: Condensed-matter physics;History;Quantum physics Subject\_term\_id: condensed-matter-physics;history;quantum-physics}, + file = {/home/johannes/Nextcloud/Zotero/2016_The rise of quantum materials.pdf;/home/johannes/Zotero/storage/YG3UAYEY/nphys3668.html} +} + +@article{ruppFastAccurateModeling2012, + title = {Fast and {{Accurate Modeling}} of {{Molecular Atomization Energies}} with {{Machine Learning}}}, + author = {Rupp, Matthias and Tkatchenko, Alexandre and Müller, Klaus-Robert and von Lilienfeld, O. Anatole}, + options = {useprefix=true}, + date = {2012-01-31}, + journaltitle = {Physical Review Letters}, + shortjournal = {Phys. Rev. Lett.}, + volume = {108}, + number = {5}, + pages = {058301}, + publisher = {{American Physical Society}}, + doi = {10.1103/PhysRevLett.108.058301}, + url = {https://link.aps.org/doi/10.1103/PhysRevLett.108.058301}, + urldate = {2021-07-10}, + abstract = {We introduce a machine learning model to predict atomization energies of a diverse set of organic molecules, based on nuclear charges and atomic positions only. The problem of solving the molecular Schrödinger equation is mapped onto a nonlinear statistical regression problem of reduced complexity. Regression models are trained on and compared to atomization energies computed with hybrid density-functional theory. Cross validation over more than seven thousand organic molecules yields a mean absolute error of ∼10 kcal/mol. Applicability is demonstrated for the prediction of molecular atomization potential energy curves.}, + keywords = {Coulomb matrix,descriptors,ML,original publication}, + file = {/home/johannes/Nextcloud/Zotero/Rupp et al_2012_Fast and Accurate Modeling of Molecular Atomization Energies with Machine.pdf;/home/johannes/Zotero/storage/AP7Y6JEW/PhysRevLett.108.html} +} + +@article{ruppMachineLearningQuantum2015, + title = {Machine Learning for Quantum Mechanics in a Nutshell}, + author = {Rupp, Matthias}, + date = {2015}, + journaltitle = {International Journal of Quantum Chemistry}, + volume = {115}, + number = {16}, + pages = {1058--1073}, + issn = {1097-461X}, + doi = {10.1002/qua.24954}, + url = {https://onlinelibrary.wiley.com/doi/abs/10.1002/qua.24954}, + urldate = {2022-05-13}, + abstract = {Models that combine quantum mechanics (QM) with machine learning (ML) promise to deliver the accuracy of QM at the speed of ML. This hands-on tutorial introduces the reader to QM/ML models based on kernel learning, an elegant, systematically nonlinear form of ML. Pseudocode and a reference implementation are provided, enabling the reader to reproduce results from recent publications where atomization energies of small organic molecules are predicted using kernel ridge regression. © 2015 Wiley Periodicals, Inc.}, + langid = {english}, + keywords = {_tablet,Coulomb matrix,GPR,kernel methods,KRR,ML,models,tutorial}, + annotation = {\_eprint: https://onlinelibrary.wiley.com/doi/pdf/10.1002/qua.24954}, + file = {/home/johannes/Nextcloud/Zotero/Rupp_2015_Machine learning for quantum mechanics in a nutshell.pdf;/home/johannes/Zotero/storage/7CP5YBAD/qua.html} +} + +@article{russmannAiiDAKKRPluginIts2021, + title = {The {{AiiDA-KKR}} Plugin and Its Application to High-Throughput Impurity Embedding into a Topological Insulator}, + author = {Rüßmann, Philipp and Bertoldo, Fabian and Blügel, Stefan}, + date = {2021-01-26}, + journaltitle = {npj Computational Materials}, + volume = {7}, + number = {1}, + pages = {1--9}, + publisher = {{Nature Publishing Group}}, + issn = {2057-3960}, + doi = {10.1038/s41524-020-00482-5}, + url = {https://www.nature.com/articles/s41524-020-00482-5}, + urldate = {2021-05-13}, + abstract = {The ever increasing availability of supercomputing resources led computer-based materials science into a new era of high-throughput calculations. Recently, Pizzi et al. introduced the AiiDA framework that provides a way to automate calculations while allowing to store the full provenance of complex workflows in a database. We present the development of the AiiDA-KKR plugin that allows to perform a large number of ab initio impurity embedding calculations based on the relativistic full-potential Korringa-Kohn-Rostoker Green function method. The capabilities of the AiiDA-KKR plugin are demonstrated with the calculation of several thousand impurities embedded into the prototypical topological insulator Sb2Te3. The results are collected in the JuDiT database which we use to investigate chemical trends as well as Fermi level and layer dependence of physical properties of impurities. This includes the study of spin moments, the impurity’s tendency to form in-gap states or its effect on the charge doping of the host-crystal. These properties depend on the detailed electronic structure of the impurity embedded into the host crystal which highlights the need for ab initio calculations in order to get accurate predictions.}, + issue = {1}, + langid = {english}, + keywords = {_tablet,AiiDA,aiida-kkr,defects,FZJ,impurity embedding,juKKR,KKR,PGI-1/IAS-1,physics,topological insulator}, + file = {/home/johannes/Nextcloud/Zotero/Rüßmann et al_2021_The AiiDA-KKR plugin and its application to high-throughput impurity embedding.pdf;/home/johannes/Zotero/storage/X4T36V7Q/s41524-020-00482-5.html} +} + +@article{russmannAiiDASpiritPluginAutomated2022, + title = {The {{AiiDA-Spirit Plugin}} for {{Automated Spin-Dynamics Simulations}} and {{Multi-Scale Modeling Based}} on {{First-Principles Calculations}}}, + author = {Rüßmann, Philipp and Ribas Sobreviela, Jordi and Sallermann, Moritz and Hoffmann, Markus and Rhiem, Florian and Blügel, Stefan}, + date = {2022}, + journaltitle = {Frontiers in Materials}, + volume = {9}, + issn = {2296-8016}, + doi = {10.3389/fmats.2022.825043}, + url = {https://www.frontiersin.org/articles/10.3389/fmats.2022.825043}, + urldate = {2022-08-11}, + abstract = {Landau-Lifshitz-Gilbert (LLG) spin-dynamics calculations based on the extended Heisenberg Hamiltonian is an important tool in computational materials science involving magnetic materials. LLG simulations allow to bridge the gap from expensive quantum mechanical calculations with small unit cells to large supercells where the collective behavior of millions of spins can be studied. In this work we present the AiiDA-Spirit plugin that connects the spin-dynamics code Spirit to the AiiDA framework. AiiDA provides a Python interface that facilitates performing high-throughput calculations while automatically augmenting the calculations with metadata describing the data provenance between calculations in a directed acyclic graph. The AiiDA-Spirit interface thus provides an easy way for high-throughput spin-dynamics calculations. The interface to the AiiDA infrastructure furthermore has the advantage that input parameters for the extended Heisenberg model can be extracted from high-throughput first-principles calculations including a proper treatment of the data provenance that ensures reproducibility of the calculation results in accordance to the FAIR principles. We describe the layout of the AiiDA-Spirit plugin and demonstrate its capabilities using selected examples for LLG spin-dynamics and Monte Carlo calculations. Furthermore, the integration with first-principles calculations through AiiDA is demonstrated at the example of γ–Fe, where the complex spin-spiral ground state is investigated.}, + keywords = {_tablet,AiiDA,aiida-kkr,Heisenberg model,Jij,KKR,library,PGI-1/IAS-1,rec-by-ruess,spin dynamics,Spirit,with-code}, + file = {/home/johannes/Nextcloud/Zotero/Rüßmann et al_2022_The AiiDA-Spirit Plugin for Automated Spin-Dynamics Simulations and Multi-Scale.pdf} +} + +@article{russmannInitioTheoryFourierTransformed2021, + title = {Ab {{Initio Theory}} of {{Fourier-Transformed Quasiparticle Interference Maps}} and {{Application}} to the {{Topological Insulator Bi2Te3}}}, + author = {Rüßmann, Philipp and Mavropoulos, Phivos and Blügel, Stefan}, + date = {2021}, + journaltitle = {physica status solidi (b)}, + volume = {258}, + number = {1}, + pages = {2000031}, + issn = {1521-3951}, + doi = {10.1002/pssb.202000031}, + url = {https://onlinelibrary.wiley.com/doi/abs/10.1002/pssb.202000031}, + urldate = {2022-05-30}, + abstract = {The quasiparticle interference (QPI) technique is a powerful tool that allows to uncover the structure and properties of electronic structure of a material combined with scattering properties of defects at surfaces. Recently, this technique has been pivotal in proving the unique properties of the surface state of topological insulators which manifests itself in the absence of backscattering. Herein, a Green function-based formalism is derived for the ab initio computation of Fourier-transformed QPI images. The efficiency of the new implementation is shown at the examples of QPI that forms around magnetic and nonmagnetic defects at the Bi2Te3 surface. This method allows a deepened understanding of the scattering properties of topologically protected electrons off defects and is a useful tool in the study of quantum materials in the future.}, + langid = {english}, + keywords = {_tablet,density functional theory,impurity scattering,Korringa–Kohn–Rostoker,quasiparticle interferences,topological insulators}, + annotation = {\_eprint: https://onlinelibrary.wiley.com/doi/pdf/10.1002/pssb.202000031}, + file = {/home/johannes/Nextcloud/Zotero/Rüßmann et al_2021_Ab Initio Theory of Fourier-Transformed Quasiparticle Interference Maps and.pdf} +} + +@thesis{russmannSpinScatteringTopologically2018, + title = {Spin Scattering of Topologically Protected Electrons at Defects}, + author = {Rüßmann, Philipp}, + date = {2018}, + number = {FZJ-2018-04348}, + institution = {{Forschungszentrum Jülich GmbH Zentralbibliothek, Verlag}}, + url = {http://hdl.handle.net/2128/19428}, + urldate = {2022-08-12}, + abstract = {This thesis provides a detailed microscopic understanding of the impurity scattering of topologically protected electrons, which are studied within the example of strong threedimensional topological insulators (Tls) and type-II Weyl semimetals. The immense research interest in the recent past in topological materials is to a large extend due to the fact that their unconventional electronic surface states are robust against perturbations, such as surface structural relaxations or defects. One of the most intringuing physical properties of topological surface states in Tls is the forbidden backscattering off time-reversal invariant defects, which makes Tl materials very promising candidates for future low-power electronics or quantum information technology. [...] Rüßmann, Philipp}, + isbn = {9783958063365}, + langid = {english}, + keywords = {juKKR,KKR,PGI-1/IAS-1,thesis}, + file = {/home/johannes/Nextcloud/Zotero/false;/home/johannes/Zotero/storage/T7V45S9S/850306.html} +} + +@article{ryczkoDeepLearningDensityfunctional2019, + title = {Deep Learning and Density-Functional Theory}, + author = {Ryczko, Kevin}, + date = {2019}, + journaltitle = {Physical Review A}, + shortjournal = {Phys. Rev. A}, + volume = {100}, + number = {2}, + doi = {10.1103/PhysRevA.100.022512}, + keywords = {CNN,Condensed Matter - Materials Science,DFT,featureless,ML,ML-DFT,ML-ESM,models,Physics - Computational Physics,rec-by-bluegel}, + file = {/home/johannes/Nextcloud/Zotero/Ryczko_2019_Deep learning and density-functional theory.pdf;/home/johannes/Zotero/storage/DYNSZ4CL/1811.html;/home/johannes/Zotero/storage/NJ67I8R7/PhysRevA.100.html} +} + +@article{saalMaterialsDesignDiscovery2013, + title = {Materials {{Design}} and {{Discovery}} with {{High-Throughput Density Functional Theory}}: {{The Open Quantum Materials Database}} ({{OQMD}})}, + shorttitle = {Materials {{Design}} and {{Discovery}} with {{High-Throughput Density Functional Theory}}}, + author = {Saal, James E. and Kirklin, Scott and Aykol, Muratahan and Meredig, Bryce and Wolverton, C.}, + date = {2013-11-01}, + journaltitle = {JOM}, + shortjournal = {JOM}, + volume = {65}, + number = {11}, + pages = {1501--1509}, + issn = {1543-1851}, + doi = {10.1007/s11837-013-0755-4}, + url = {https://doi.org/10.1007/s11837-013-0755-4}, + urldate = {2021-10-15}, + abstract = {High-throughput density functional theory (HT DFT) is fast becoming a powerful tool for accelerating materials design and discovery by the amassing tens and even hundreds of thousands of DFT calculations in large databases. Complex materials problems can be approached much more efficiently and broadly through the sheer quantity of structures and chemistries available in such databases. Our HT DFT database, the Open Quantum Materials Database (OQMD), contains over 200,000 DFT calculated crystal structures and will be freely available for public use at http://oqmd.org. In this review, we describe the OQMD and its use in five materials problems, spanning a wide range of applications and materials types: (I) Li-air battery combination catalyst/electrodes, (II) Li-ion battery anodes, (III) Li-ion battery cathode coatings reactive with HF, (IV) Mg-alloy long-period stacking ordered (LPSO) strengthening precipitates, and (V) training a machine learning model to predict new stable ternary compounds.}, + langid = {english}, + file = {/home/johannes/Nextcloud/Zotero/Saal et al_2013_Materials Design and Discovery with High-Throughput Density Functional Theory.pdf} +} + +@unpublished{samuelMachineLearningPipelines2020, + title = {Machine {{Learning Pipelines}}: {{Provenance}}, {{Reproducibility}} and {{FAIR Data Principles}}}, + shorttitle = {Machine {{Learning Pipelines}}}, + author = {Samuel, Sheeba and Löffler, Frank and König-Ries, Birgitta}, + date = {2020-06-22}, + eprint = {2006.12117}, + eprinttype = {arxiv}, + primaryclass = {cs, stat}, + url = {http://arxiv.org/abs/2006.12117}, + urldate = {2021-10-21}, + abstract = {Machine learning (ML) is an increasingly important scientific tool supporting decision making and knowledge generation in numerous fields. With this, it also becomes more and more important that the results of ML experiments are reproducible. Unfortunately, that often is not the case. Rather, ML, similar to many other disciplines, faces a reproducibility crisis. In this paper, we describe our goals and initial steps in supporting the end-to-end reproducibility of ML pipelines. We investigate which factors beyond the availability of source code and datasets influence reproducibility of ML experiments. We propose ways to apply FAIR data practices to ML workflows. We present our preliminary results on the role of our tool, ProvBook, in capturing and comparing provenance of ML experiments and their reproducibility using Jupyter Notebooks.}, + archiveprefix = {arXiv}, + keywords = {Computer Science - Artificial Intelligence,Computer Science - Machine Learning,Statistics - Machine Learning}, + file = {/home/johannes/Nextcloud/Zotero/Samuel et al_2020_Machine Learning Pipelines.pdf;/home/johannes/Zotero/storage/GTJD4NAB/2006.html} +} + +@inproceedings{satorrasEquivariantGraphNeural2021, + title = {E(n) {{Equivariant Graph Neural Networks}}}, + booktitle = {Proceedings of the 38th {{International Conference}} on {{Machine Learning}}}, + author = {Satorras, VıÌctor Garcia and Hoogeboom, Emiel and Welling, Max}, + date = {2021-07-01}, + pages = {9323--9332}, + publisher = {{PMLR}}, + issn = {2640-3498}, + url = {https://proceedings.mlr.press/v139/satorras21a.html}, + urldate = {2022-03-29}, + abstract = {This paper introduces a new model to learn graph neural networks equivariant to rotations, translations, reflections and permutations called E(n)-Equivariant Graph Neural Networks (EGNNs). In contrast with existing methods, our work does not require computationally expensive higher-order representations in intermediate layers while it still achieves competitive or better performance. In addition, whereas existing methods are limited to equivariance on 3 dimensional spaces, our model is easily scaled to higher-dimensional spaces. We demonstrate the effectiveness of our method on dynamical systems modelling, representation learning in graph autoencoders and predicting molecular properties.}, + eventtitle = {International {{Conference}} on {{Machine Learning}}}, + langid = {english}, + keywords = {EGNN,equivariant,GDL,GNN,ML,NN,original publication,rec-by-bluegel}, + file = {/home/johannes/Nextcloud/Zotero/Satorras et al_2021_E(n) Equivariant Graph Neural Networks2.pdf;/home/johannes/Zotero/storage/3ATM3ZJA/Satorras et al_2021_E(n) Equivariant Graph Neural Networks.pdf} +} + +@article{saucedaBIGDMLAccurateQuantum2022, + title = {{{BIGDML}}—{{Towards}} Accurate Quantum Machine Learning Force Fields for Materials}, + author = {Sauceda, Huziel E. and Gálvez-González, Luis E. and Chmiela, Stefan and Paz-Borbón, Lauro Oliver and Müller, Klaus-Robert and Tkatchenko, Alexandre}, + date = {2022-06-29}, + journaltitle = {Nature Communications}, + shortjournal = {Nat Commun}, + volume = {13}, + number = {1}, + pages = {3733}, + publisher = {{Nature Publishing Group}}, + issn = {2041-1723}, + doi = {10.1038/s41467-022-31093-x}, + url = {https://www.nature.com/articles/s41467-022-31093-x}, + urldate = {2023-01-02}, + abstract = {Machine-learning force fields (MLFF) should be accurate, computationally and data efficient, and applicable to molecules, materials, and interfaces thereof. Currently, MLFFs often introduce tradeoffs that restrict their practical applicability to small subsets of chemical space or require exhaustive datasets for training. Here, we introduce the Bravais-Inspired Gradient-Domain Machine Learning (BIGDML) approach and demonstrate its ability to construct reliable force fields using a training set with just 10–200 geometries for materials including pristine and defect-containing 2D and 3D semiconductors and metals, as well as chemisorbed and physisorbed atomic and molecular adsorbates on surfaces. The BIGDML model employs the full relevant symmetry group for a given material, does not assume artificial atom types or localization of atomic interactions and exhibits high data efficiency and state-of-the-art energy accuracies (errors substantially below 1 meV per atom) for an extended set of materials. Extensive path-integral molecular dynamics carried out with BIGDML models demonstrate the counterintuitive localization of benzene–graphene dynamics induced by nuclear quantum effects and their strong contributions to the hydrogen diffusion coefficient in a Pd crystal for a wide range of temperatures.}, + issue = {1}, + langid = {english}, + keywords = {/unread,Coulomb matrix,crystal symmetry,defects,force fields,interfaces and thin films,library,materials,MD,ML,MLFF,MLP,models,molecular dynamics,PyTorch,sGDML,surface physics,symmetry,with-code}, + file = {/home/johannes/Nextcloud/Zotero/Sauceda et al_2022_BIGDML—Towards accurate quantum machine learning force fields for materials.pdf} +} + +@unpublished{saucedaBIGDMLExactMachine2021, + title = {{{BIGDML}}: {{Towards Exact Machine Learning Force Fields}} for {{Materials}}}, + shorttitle = {{{BIGDML}}}, + author = {Sauceda, Huziel E. and Gálvez-González, Luis E. and Chmiela, Stefan and Paz-Borbón, Lauro Oliver and Müller, Klaus-Robert and Tkatchenko, Alexandre}, + date = {2021-06-08}, + eprint = {2106.04229}, + eprinttype = {arxiv}, + primaryclass = {cond-mat, physics:physics, physics:quant-ph}, + url = {http://arxiv.org/abs/2106.04229}, + urldate = {2021-06-17}, + abstract = {Machine-learning force fields (MLFF) should be accurate, computationally and data efficient, and applicable to molecules, materials, and interfaces thereof. Currently, MLFFs often introduce tradeoffs that restrict their practical applicability to small subsets of chemical space or require exhaustive datasets for training. Here, we introduce the Bravais-Inspired Gradient-Domain Machine Learning (BIGDML) approach and demonstrate its ability to construct reliable force fields using a training set with just 10-200 geometries for materials including pristine and defect-containing 2D and 3D semiconductors and metals, as well as chemisorbed and physisorbed atomic and molecular adsorbates on surfaces. The BIGDML model employs the full relevant symmetry group for a given material, does not assume artificial atom types or localization of atomic interactions and exhibits high data efficiency and state-of-the-art energy accuracies (errors substantially below 1 meV per atom) for an extended set of materials. Extensive path-integral molecular dynamics carried out with BIGDML models demonstrate the counterintuitive localization of benzene--graphene dynamics induced by nuclear quantum effects and allow to rationalize the Arrhenius behavior of hydrogen diffusion coefficient in a Pd crystal for a wide range of temperatures.}, + archiveprefix = {arXiv}, + keywords = {CM,ML,ML-FF,MLP,models,PES,sGDML}, + file = {/home/johannes/Nextcloud/Zotero/Sauceda et al_2021_BIGDML.pdf;/home/johannes/Zotero/storage/XVR5SBVI/2106.html} +} + +@unpublished{scherbelaSolvingElectronicSchr2021, + title = {Solving the Electronic {{Schr}}\textbackslash "odinger Equation for Multiple Nuclear Geometries with Weight-Sharing Deep Neural Networks}, + author = {Scherbela, Michael and Reisenhofer, Rafael and Gerard, Leon and Marquetand, Philipp and Grohs, Philipp}, + date = {2021-12-17}, + eprint = {2105.08351}, + eprinttype = {arxiv}, + primaryclass = {physics}, + url = {http://arxiv.org/abs/2105.08351}, + urldate = {2022-03-28}, + abstract = {Accurate numerical solutions for the Schr\textbackslash "odinger equation are of utmost importance in quantum chemistry. However, the computational cost of current high-accuracy methods scales poorly with the number of interacting particles. Combining Monte Carlo methods with unsupervised training of neural networks has recently been proposed as a promising approach to overcome the curse of dimensionality in this setting and to obtain accurate wavefunctions for individual molecules at a moderately scaling computational cost. These methods currently do not exploit the regularity exhibited by wavefunctions with respect to their molecular geometries. Inspired by recent successful applications of deep transfer learning in machine translation and computer vision tasks, we attempt to leverage this regularity by introducing a weight-sharing constraint when optimizing neural network-based models for different molecular geometries. That is, we restrict the optimization process such that up to 95 percent of weights in a neural network model are in fact equal across varying molecular geometries. We find that this technique can accelerate optimization when considering sets of nuclear geometries of the same molecule by an order of magnitude and that it opens a promising route towards pre-trained neural network wavefunctions that yield high accuracy even across different molecules.}, + archiveprefix = {arXiv}, + keywords = {Computer Science - Machine Learning,Physics - Chemical Physics,Physics - Computational Physics}, + file = {/home/johannes/Nextcloud/Zotero/Scherbela et al_2021_Solving the electronic Schr-odinger equation for multiple nuclear geometries.pdf;/home/johannes/Zotero/storage/PSDVAEAB/2105.html} +} + +@article{scherbelaSolvingElectronicSchrodinger2022, + title = {Solving the Electronic {{Schrödinger}} Equation for Multiple Nuclear Geometries with Weight-Sharing Deep Neural Networks}, + author = {Scherbela, Michael and Reisenhofer, Rafael and Gerard, Leon and Marquetand, Philipp and Grohs, Philipp}, + date = {2022-05}, + journaltitle = {Nature Computational Science}, + shortjournal = {Nat Comput Sci}, + volume = {2}, + number = {5}, + pages = {331--341}, + publisher = {{Nature Publishing Group}}, + issn = {2662-8457}, + doi = {10.1038/s43588-022-00228-x}, + url = {https://www.nature.com/articles/s43588-022-00228-x}, + urldate = {2022-08-16}, + abstract = {The Schrödinger equation describes the quantum-mechanical behaviour of particles, making it the most fundamental equation in chemistry. A solution for a given molecule allows computation of any of its properties. Finding accurate solutions for many different molecules and geometries is thus crucial to the discovery of new materials such as drugs or catalysts. Despite its importance, the Schrödinger equation is notoriously difficult to solve even for single molecules, as established methods scale exponentially with the number of particles. Combining Monte Carlo techniques with unsupervised optimization of neural networks was recently discovered as a promising approach to overcome this curse of dimensionality, but the corresponding methods do not exploit synergies that arise when considering multiple geometries. Here we show that sharing the vast majority of weights across neural network models for different geometries substantially accelerates optimization. Furthermore, weight-sharing yields pretrained models that require only a small number of additional optimization steps to obtain high-accuracy solutions for new geometries.}, + issue = {5}, + langid = {english}, + keywords = {Backflow,cusps,DeepErwin,FermiNet,JAX,library,MCMC,ML-ESM,ML-QMBP,molecules,PauliNet,prediction of wavefunction,QMC,VMC,weight-sharing,with-code}, + file = {/home/johannes/Nextcloud/Zotero/Scherbela et al_2022_Solving the electronic Schrödinger equation for multiple nuclear geometries.pdf;/home/johannes/Zotero/storage/JHT752NC/s43588-022-00228-x.html} +} + +@article{schmidtMachineLearningPhysical2019, + title = {Machine {{Learning}} the {{Physical Nonlocal Exchange}}–{{Correlation Functional}} of {{Density-Functional Theory}}}, + author = {Schmidt, Jonathan and Benavides-Riveros, Carlos L. and Marques, Miguel A. L.}, + date = {2019-10-17}, + journaltitle = {The Journal of Physical Chemistry Letters}, + shortjournal = {J. Phys. Chem. Lett.}, + volume = {10}, + number = {20}, + pages = {6425--6431}, + publisher = {{American Chemical Society}}, + doi = {10.1021/acs.jpclett.9b02422}, + url = {https://doi.org/10.1021/acs.jpclett.9b02422}, + urldate = {2022-07-05}, + abstract = {We train a neural network as the universal exchange–correlation functional of density-functional theory that simultaneously reproduces both the exact exchange–correlation energy and the potential. This functional is extremely nonlocal but retains the computational scaling of traditional local or semilocal approximations. It therefore holds the promise of solving some of the delocalization problems that plague density-functional theory, while maintaining the computational efficiency that characterizes the Kohn–Sham equations. Furthermore, by using automatic differentiation, a capability present in modern machine-learning frameworks, we impose the exact mathematical relation between the exchange–correlation energy and the potential, leading to a fully consistent method. We demonstrate the feasibility of our approach by looking at one-dimensional systems with two strongly correlated electrons, where density-functional methods are known to fail, and investigate the behavior and performance of our functional by varying the degree of nonlocality.}, + keywords = {autodiff,DFT,ML,ML-DFA,ML-DFT,ML-ESM,prediction from density,prediction of Exc,prediction of vxc,pytorch}, + file = {/home/johannes/Nextcloud/Zotero/Schmidt et al_2019_Machine Learning the Physical Nonlocal Exchange–Correlation Functional of.pdf;/home/johannes/Zotero/storage/QCMK7FSR/acs.jpclett.html} +} + +@article{schmidtRecentAdvancesApplications2019, + title = {Recent Advances and Applications of Machine Learning in Solid-State Materials Science}, + author = {Schmidt, Jonathan and Marques, Mário R. G. and Botti, Silvana and Marques, Miguel A. L.}, + date = {2019-08-08}, + journaltitle = {npj Computational Materials}, + shortjournal = {npj Comput Mater}, + volume = {5}, + number = {1}, + pages = {1--36}, + publisher = {{Nature Publishing Group}}, + issn = {2057-3960}, + doi = {10.1038/s41524-019-0221-0}, + url = {https://www.nature.com/articles/s41524-019-0221-0}, + urldate = {2021-06-29}, + abstract = {One of the most exciting tools that have entered the material science toolbox in recent years is machine learning. This collection of statistical methods has already proved to be capable of considerably speeding up both fundamental and applied research. At present, we are witnessing an explosion of works that develop and apply machine learning to solid-state systems. We provide a comprehensive overview and analysis of the most recent research in this topic. As a starting point, we introduce machine learning principles, algorithms, descriptors, and databases in materials science. We continue with the description of different machine learning approaches for the discovery of stable materials and the prediction of their crystal structure. Then we discuss research in numerous quantitative structure–property relationships and various approaches for the replacement of first-principle methods by machine learning. We review how active learning and surrogate-based optimization can be applied to improve the rational design process and related examples of applications. Two major questions are always the interpretability of and the physical understanding gained from machine learning models. We consider therefore the different facets of interpretability and their importance in materials science. Finally, we propose solutions and future research paths for various challenges in computational materials science.}, + issue = {1}, + langid = {english}, + annotation = {Bandiera\_abtest: a Cc\_license\_type: cc\_by Cg\_type: Nature Research Journals Primary\_atype: Reviews Subject\_term: Condensed-matter physics;Electronic structure;Materials science;Metals and alloys;Semiconductors Subject\_term\_id: condensed-matter-physics;electronic-structure;materials-science;metals-and-alloys;semiconductors}, + file = {/home/johannes/Nextcloud/Zotero/Schmidt et al_2019_Recent advances and applications of machine learning in solid-state materials.pdf;/home/johannes/Zotero/storage/BY9RESIZ/s41524-019-0221-0.html} +} + +@article{schmittIntegrationTopologicalInsulator2022, + title = {Integration of {{Topological Insulator Josephson Junctions}} in {{Superconducting Qubit Circuits}}}, + author = {Schmitt, Tobias W. and Connolly, Malcolm R. and Schleenvoigt, Michael and Liu, Chenlu and Kennedy, Oscar and Chávez-Garcia, José M. and Jalil, Abdur R. and Bennemann, Benjamin and Trellenkamp, Stefan and Lentz, Florian and Neumann, Elmar and Lindström, Tobias and de Graaf, Sebastian E. and Berenschot, Erwin and Tas, Niels and Mussler, Gregor and Petersson, Karl D. and Grützmacher, Detlev and Schüffelgen, Peter}, + options = {useprefix=true}, + date = {2022-04-13}, + journaltitle = {Nano Letters}, + shortjournal = {Nano Lett.}, + volume = {22}, + number = {7}, + pages = {2595--2602}, + publisher = {{American Chemical Society}}, + issn = {1530-6984}, + doi = {10.1021/acs.nanolett.1c04055}, + url = {https://doi.org/10.1021/acs.nanolett.1c04055}, + urldate = {2022-04-16}, + abstract = {The integration of semiconductor Josephson junctions (JJs) in superconducting quantum circuits provides a versatile platform for hybrid qubits and offers a powerful way to probe exotic quasiparticle excitations. Recent proposals for using circuit quantum electrodynamics (cQED) to detect topological superconductivity motivate the integration of novel topological materials in such circuits. Here, we report on the realization of superconducting transmon qubits implemented with (Bi0.06Sb0.94)2Te3 topological insulator (TI) JJs using ultrahigh vacuum fabrication techniques. Microwave losses on our substrates, which host monolithically integrated hardmasks used for the selective area growth of TI nanostructures, imply microsecond limits to relaxation times and, thus, their compatibility with strong-coupling cQED. We use the cavity–qubit interaction to show that the Josephson energy of TI-based transmons scales with their JJ dimensions and demonstrate qubit control as well as temporal quantum coherence. Our results pave the way for advanced investigations of topological materials in both novel Josephson and topological qubits.}, + keywords = {_tablet,experimental,FZJ,PGI,PGI-9,quantum computing,superconductor,topological insulator}, + file = {/home/johannes/Nextcloud/Zotero/Schmitt et al_2022_Integration of Topological Insulator Josephson Junctions in Superconducting.pdf;/home/johannes/Zotero/storage/ZVNVRDSF/acs.nanolett.html} +} + +@article{schuchComputationalComplexityInteracting2009, + title = {Computational Complexity of Interacting Electrons and Fundamental Limitations of Density Functional Theory}, + author = {Schuch, Norbert and Verstraete, Frank}, + date = {2009-10}, + journaltitle = {Nature Physics}, + shortjournal = {Nature Phys}, + volume = {5}, + number = {10}, + pages = {732--735}, + publisher = {{Nature Publishing Group}}, + issn = {1745-2481}, + doi = {10.1038/nphys1370}, + url = {https://www.nature.com/articles/nphys1370}, + urldate = {2022-10-05}, + abstract = {Using arguments from computational complexity theory, fundamental limitations are found for how efficient it is to calculate the ground-state energy of many-electron systems using density functional theory.}, + issue = {10}, + langid = {english}, + keywords = {computational complexity,DFT,NP-hard,Schrödinger equation}, + file = {/home/johannes/Nextcloud/Zotero/Schuch_Verstraete_2009_Computational complexity of interacting electrons and fundamental limitations.pdf;/home/johannes/Zotero/storage/3TCEDPTF/nphys1370.html} +} + +@book{schuttMachineLearningMeets2020, + title = {Machine {{Learning Meets Quantum Physics}}}, + editor = {Schütt, Kristof T. and Chmiela, Stefan and von Lilienfeld, O. Anatole and Tkatchenko, Alexandre and Tsuda, Koji and Müller, Klaus-Robert}, + options = {useprefix=true}, + date = {2020}, + series = {Lecture {{Notes}} in {{Physics}}}, + volume = {968}, + publisher = {{Springer International Publishing}}, + location = {{Cham}}, + doi = {10.1007/978-3-030-40245-7}, + url = {http://link.springer.com/10.1007/978-3-030-40245-7}, + urldate = {2021-05-13}, + isbn = {978-3-030-40244-0 978-3-030-40245-7}, + langid = {english}, + keywords = {chemistry,descriptors,general,ML,models,physics,review}, + file = {/home/johannes/Books/scientific_machine_learning/Schütt et al_2020_Machine Learning Meets Quantum Physics.pdf} +} + +@article{schuttSchNetDeepLearning2018, + title = {{{SchNet}} – {{A}} Deep Learning Architecture for Molecules and Materials}, + author = {Schütt, K. T. and Sauceda, H. E. and Kindermans, P.-J. and Tkatchenko, A. and Müller, K.-R.}, + date = {2018-06-28}, + journaltitle = {The Journal of Chemical Physics}, + shortjournal = {J. Chem. Phys.}, + volume = {148}, + number = {24}, + pages = {241722}, + publisher = {{American Institute of Physics}}, + issn = {0021-9606}, + doi = {10.1063/1.5019779}, + url = {https://aip.scitation.org/doi/full/10.1063/1.5019779}, + urldate = {2022-10-03}, + abstract = {Deep learning has led to a paradigm shift in artificial intelligence, including web, text, and image search, speech recognition, as well as bioinformatics, with growing impact in chemical physics. Machine learning, in general, and deep learning, in particular, are ideally suitable for representing quantum-mechanical interactions, enabling us to model nonlinear potential-energy surfaces or enhancing the exploration of chemical compound space. Here we present the deep learning architecture SchNet that is specifically designed to model atomistic systems by making use of continuous-filter convolutional layers. We demonstrate the capabilities of SchNet by accurately predicting a range of properties across chemical space for molecules and materials, where our model learns chemically plausible embeddings of atom types across the periodic table. Finally, we employ SchNet to predict potential-energy surfaces and energy-conserving force fields for molecular dynamics simulations of small molecules and perform an exemplary study on the quantum-mechanical properties of C20-fullerene that would have been infeasible with regular ab initio molecular dynamics.}, + keywords = {DTNN,MD,ML,MLP,NN,SchNet}, + file = {/home/johannes/Nextcloud/Zotero/Schütt et al_2018_SchNet – A deep learning architecture for molecules and materials.pdf} +} + +@misc{schuttSchNetPackNeuralNetwork2022, + title = {{{SchNetPack}} 2.0: {{A}} Neural Network Toolbox for Atomistic Machine Learning}, + shorttitle = {{{SchNetPack}} 2.0}, + author = {Schütt, Kristof T. and Hessmann, Stefaan S. P. and Gebauer, Niklas W. A. and Lederer, Jonas and Gastegger, Michael}, + date = {2022-12-11}, + number = {arXiv:2212.05517}, + eprint = {2212.05517}, + eprinttype = {arxiv}, + primaryclass = {physics, stat}, + publisher = {{arXiv}}, + doi = {10.48550/arXiv.2212.05517}, + url = {http://arxiv.org/abs/2212.05517}, + urldate = {2022-12-27}, + abstract = {SchNetPack is a versatile neural networks toolbox that addresses both the requirements of method development and application of atomistic machine learning. Version 2.0 comes with an improved data pipeline, modules for equivariant neural networks as well as a PyTorch implementation of molecular dynamics. An optional integration with PyTorch Lightning and the Hydra configuration framework powers a flexible command-line interface. This makes SchNetPack 2.0 easily extendable with custom code and ready for complex training task such as generation of 3d molecular structures.}, + archiveprefix = {arXiv}, + keywords = {/unread,Deep learning,equivariant,Hydra,library,MLP,models,PAiNN,pytorch,SchNet,SO(3),with-code}, + file = {/home/johannes/Nextcloud/Zotero/Schütt et al_2022_SchNetPack 2.pdf;/home/johannes/Zotero/storage/AHBKQSBM/2212.html} +} + +@article{schuttUnifyingMachineLearning2019, + title = {Unifying Machine Learning and Quantum Chemistry with a Deep Neural Network for Molecular Wavefunctions}, + author = {Schütt, K. T. and Gastegger, M. and Tkatchenko, A. and Müller, K.-R. and Maurer, R. J.}, + date = {2019-11-15}, + journaltitle = {Nature Communications}, + shortjournal = {Nat Commun}, + volume = {10}, + number = {1}, + pages = {5024}, + publisher = {{Nature Publishing Group}}, + issn = {2041-1723}, + doi = {10.1038/s41467-019-12875-2}, + url = {https://www.nature.com/articles/s41467-019-12875-2}, + urldate = {2021-06-09}, + abstract = {Machine learning advances chemistry and materials science by enabling large-scale exploration of chemical space based on quantum chemical calculations. While these models supply fast and accurate predictions of atomistic chemical properties, they do not explicitly capture the electronic degrees of freedom of a molecule, which limits their applicability for reactive chemistry and chemical analysis. Here we present a deep learning framework for the prediction of the quantum mechanical wavefunction in a local basis of atomic orbitals from which all other ground-state properties can be derived. This approach retains full access to the electronic structure via the wavefunction at force-field-like efficiency and captures quantum mechanics in an analytically differentiable representation. On several examples, we demonstrate that this opens promising avenues to perform inverse design of molecular structures for targeting electronic property optimisation and a clear path towards increased synergy of machine learning and quantum chemistry.}, + issue = {1}, + langid = {english}, + keywords = {ANN,ML,ML-ESM,models,molecules,original publication,prediction of wavefunction,SchNet,SchNOrb,WFT}, + file = {/home/johannes/Nextcloud/Zotero/Schütt et al_2019_Unifying machine learning and quantum chemistry with a deep neural network for.pdf;/home/johannes/Zotero/storage/ADRZDHRZ/s41467-019-12875-2.html} +} + +@article{sendekMachineLearningModeling, + title = {Machine {{Learning Modeling}} for {{Accelerated Battery Materials Design}} in the {{Small Data Regime}}}, + author = {Sendek, Austin D. and Ransom, Brandi and Cubuk, Ekin D. and Pellouchoud, Lenson A. and Nanda, Jagjit and Reed, Evan J.}, + journaltitle = {Advanced Energy Materials}, + volume = {n/a}, + number = {n/a}, + pages = {2200553}, + issn = {1614-6840}, + doi = {10.1002/aenm.202200553}, + url = {https://onlinelibrary.wiley.com/doi/abs/10.1002/aenm.202200553}, + urldate = {2022-07-03}, + abstract = {Machine learning (ML)-based approaches to battery design are relatively new but demonstrate significant promise for accelerating the timeline for new materials discovery, process optimization, and cell lifetime prediction. Battery modeling represents an interesting and unconventional application area for ML, as datasets are often small but some degree of physical understanding of the underlying processes may exist. This review article provides discussion and analysis of several important and increasingly common questions: how ML-based battery modeling works, how much data are required, how to judge model performance, and recommendations for building models in the small data regime. This article begins with an introduction to ML in general, highlighting several important concepts for small data applications. Previous ionic conductivity modeling efforts are discussed in depth as a case study to illustrate these modeling concepts. Finally, an overview of modeling efforts in major areas of battery design is provided and several areas for promising future efforts are identified, within the context of typical small data constraints.}, + langid = {english}, + keywords = {_tablet,chemistry,materials informatics,ML,small data,tutorial}, + annotation = {\_eprint: https://onlinelibrary.wiley.com/doi/pdf/10.1002/aenm.202200553}, + file = {/home/johannes/Nextcloud/Zotero/Sendek et al_Machine Learning Modeling for Accelerated Battery Materials Design in the Small.pdf;/home/johannes/Zotero/storage/55KE647F/aenm.html} +} + +@misc{shenRepresentationindependentElectronicCharge2021, + title = {A Representation-Independent Electronic Charge Density Database for Crystalline Materials}, + author = {Shen, Jimmy-Xuan and Munro, Jason M. and Horton, Matthew K. and Huck, Patrick and Dwaraknath, Shyam and Persson, Kristin A.}, + date = {2021-07-07}, + number = {arXiv:2107.03540}, + eprint = {2107.03540}, + eprinttype = {arxiv}, + primaryclass = {cond-mat}, + publisher = {{arXiv}}, + doi = {10.48550/arXiv.2107.03540}, + url = {http://arxiv.org/abs/2107.03540}, + urldate = {2022-12-31}, + abstract = {In addition to being the core quantity in density functional theory, the charge density can be used in many tertiary analyses in materials sciences from bonding to assigning charge to specific atoms. The charge density is data-rich since it contains information about all the electrons in the system. With increasing utilization of machine-learning tools in materials sciences, a data-rich object like the charge density can be utilized in a wide range of applications. The database presented here provides a modern and user-friendly interface for a large and continuously updated collection of charge densities as part of the Materials Project. In addition to the charge density data, we provide the theory and code for changing the representation of the charge density which should enable more advanced machine-learning studies for the broader community.}, + archiveprefix = {arXiv}, + keywords = {/unread,charge density,data repositories,Database,dimensionality reduction of target,electronic structure,library,materials,materials database,materials project,ML,ML-DFT,prediction from density,prediction of electron density,with-code}, + file = {/home/johannes/Nextcloud/Zotero/Shen et al_2021_A representation-independent electronic charge density database for crystalline.pdf;/home/johannes/Zotero/storage/9A3MUVVK/2107.html} +} + +@article{shmilovichOrbitalMixerUsing2022, + title = {Orbital {{Mixer}}: {{Using Atomic Orbital Features}} for {{Basis-Dependent Prediction}} of {{Molecular Wavefunctions}}}, + shorttitle = {Orbital {{Mixer}}}, + author = {Shmilovich, Kirill and Willmott, Devin and Batalov, Ivan and Kornbluth, Mordechai and Mailoa, Jonathan and Kolter, J. Zico}, + date = {2022-09-19}, + journaltitle = {Journal of Chemical Theory and Computation}, + shortjournal = {J. Chem. Theory Comput.}, + publisher = {{American Chemical Society}}, + issn = {1549-9618}, + doi = {10.1021/acs.jctc.2c00555}, + url = {https://doi.org/10.1021/acs.jctc.2c00555}, + urldate = {2022-09-29}, + abstract = {Leveraging ab initio data at scale has enabled the development of machine learning models capable of extremely accurate and fast molecular property prediction. A central paradigm of many previous studies focuses on generating predictions for only a fixed set of properties. Recent lines of research instead aim to explicitly learn the electronic structure via molecular wavefunctions, from which other quantum chemical properties can be directly derived. While previous methods generate predictions as a function of only the atomic configuration, in this work we present an alternate approach that directly purposes basis-dependent information to predict molecular electronic structure. Our model, Orbital Mixer, is composed entirely of multi-layer perceptrons (MLPs) using MLP-Mixer layers within a simple, intuitive, and scalable architecture that achieves competitive Hamiltonian and molecular orbital energy and coefficient prediction accuracies compared to the state-of-the-art.}, + keywords = {ML,ML-ESM,MLP,molecules,Orbital Mixer,original publication,PhiSNet,prediction of wavefunction,SchNOrb,with-code}, + file = {/home/johannes/Nextcloud/Zotero/Shmilovich et al_2022_Orbital Mixer.pdf} +} + +@article{simmhanSurveyDataProvenance2005, + title = {A Survey of Data Provenance in E-Science}, + author = {Simmhan, Yogesh L. and Plale, Beth and Gannon, Dennis}, + date = {2005-09-01}, + journaltitle = {ACM SIGMOD Record}, + shortjournal = {SIGMOD Rec.}, + volume = {34}, + number = {3}, + pages = {31--36}, + issn = {0163-5808}, + doi = {10.1145/1084805.1084812}, + url = {https://doi.org/10.1145/1084805.1084812}, + urldate = {2021-10-17}, + abstract = {Data management is growing in complexity as large-scale applications take advantage of the loosely coupled resources brought together by grid middleware and by abundant storage capacity. Metadata describing the data products used in and generated by these applications is essential to disambiguate the data and enable reuse. Data provenance, one kind of metadata, pertains to the derivation history of a data product starting from its original sources.In this paper we create a taxonomy of data provenance characteristics and apply it to current research efforts in e-science, focusing primarily on scientific workflow approaches. The main aspect of our taxonomy categorizes provenance systems based on why they record provenance, what they describe, how they represent and store provenance, and ways to disseminate it. The survey culminates with an identification of open research problems in the field.}, + file = {/home/johannes/Nextcloud/Zotero/Simmhan et al_2005_A survey of data provenance in e-science.pdf} +} + +@article{singraberParallelMultistreamTraining2019, + title = {Parallel {{Multistream Training}} of {{High-Dimensional Neural Network Potentials}}}, + author = {Singraber, Andreas and Morawietz, Tobias and Behler, Jörg and Dellago, Christoph}, + date = {2019-05-14}, + journaltitle = {Journal of Chemical Theory and Computation}, + shortjournal = {J. Chem. Theory Comput.}, + volume = {15}, + number = {5}, + pages = {3075--3092}, + publisher = {{American Chemical Society}}, + issn = {1549-9618}, + doi = {10.1021/acs.jctc.8b01092}, + url = {https://doi.org/10.1021/acs.jctc.8b01092}, + urldate = {2021-05-18}, + abstract = {Over the past years high-dimensional neural network potentials (HDNNPs), fitted to accurately reproduce ab initio potential energy surfaces, have become a powerful tool in chemistry, physics and materials science. Here, we focus on the training of the neural networks that lies at the heart of the HDNNP method. We present an efficient approach for optimizing the weight parameters of the neural network via multistream Kalman filtering, using potential energies and forces as reference data. In this procedure, the choice of the free parameters of the Kalman filter can have a significant impact on the fit quality. Carrying out a large parameter study, we determine optimal settings and demonstrate how to optimize training results of HDNNPs. Moreover, we illustrate our HDNNP training approach by revisiting previously presented fits for water and developing a new potential for copper sulfide. This material, accessible in computer simulations so far only via first-principles methods, forms a particularly complex solid structure at low temperatures and undergoes a phase transition to a superionic state upon heating. Analyzing MD simulations carried out with the Cu2S HDNNP, we confirm that the underlying ab initio reference method indeed reproduces this behavior.}, + keywords = {HDNNP,ML,MLP,models,parallelization}, + file = {/home/johannes/Nextcloud/Zotero/Singraber et al_2019_Parallel Multistream Training of High-Dimensional Neural Network Potentials.pdf} +} + +@article{sivaramanMachinelearnedInteratomicPotentials2020, + title = {Machine-Learned Interatomic Potentials by Active Learning: Amorphous and Liquid Hafnium Dioxide}, + shorttitle = {Machine-Learned Interatomic Potentials by Active Learning}, + author = {Sivaraman, Ganesh and Krishnamoorthy, Anand Narayanan and Baur, Matthias and Holm, Christian and Stan, Marius and Csányi, Gábor and Benmore, Chris and Vázquez-Mayagoitia, Ãlvaro}, + date = {2020-07-23}, + journaltitle = {npj Computational Materials}, + volume = {6}, + number = {1}, + pages = {1--8}, + publisher = {{Nature Publishing Group}}, + issn = {2057-3960}, + doi = {10.1038/s41524-020-00367-7}, + url = {https://www.nature.com/articles/s41524-020-00367-7}, + urldate = {2021-05-13}, + abstract = {We propose an active learning scheme for automatically sampling a minimum number of uncorrelated configurations for fitting the Gaussian Approximation Potential (GAP). Our active learning scheme consists of an unsupervised machine learning (ML) scheme coupled with a Bayesian optimization technique that evaluates the GAP model. We apply this scheme to a Hafnium dioxide (HfO2) dataset generated from a “melt-quench†ab initio molecular dynamics (AIMD) protocol. Our results show that the active learning scheme, with no prior knowledge of the dataset, is able to extract a configuration that reaches the required energy fit tolerance. Further, molecular dynamics (MD) simulations performed using this active learned GAP model on 6144 atom systems of amorphous and liquid state elucidate the structural properties of HfO2 with near ab initio precision and quench rates (i.e., 1.0\,K/ps) not accessible via AIMD. The melt and amorphous X-ray structural factors generated from our simulation are in good agreement with experiment. In addition, the calculated diffusion constants are in good agreement with previous ab initio studies.}, + issue = {1}, + langid = {english}, + keywords = {active learning,AIMD,GAP,MD,ML}, + file = {/home/johannes/Nextcloud/Zotero/Sivaraman et al_2020_Machine-learned interatomic potentials by active learning.pdf;/home/johannes/Zotero/storage/WMMU2G78/s41524-020-00367-7.html} +} + +@article{smidtFindingSymmetryBreaking2021, + title = {Finding Symmetry Breaking Order Parameters with {{Euclidean}} Neural Networks}, + author = {Smidt, Tess E. and Geiger, Mario and Miller, Benjamin Kurt}, + date = {2021-01-04}, + journaltitle = {Physical Review Research}, + shortjournal = {Phys. Rev. Research}, + volume = {3}, + number = {1}, + pages = {L012002}, + publisher = {{American Physical Society}}, + doi = {10.1103/PhysRevResearch.3.L012002}, + url = {https://link.aps.org/doi/10.1103/PhysRevResearch.3.L012002}, + urldate = {2022-10-17}, + abstract = {Curie's principle states that “when effects show certain asymmetry, this asymmetry must be found in the causes that gave rise to them.†We demonstrate that symmetry equivariant neural networks uphold Curie's principle and can be used to articulate many symmetry-relevant scientific questions as simple optimization problems. We prove these properties mathematically and demonstrate them numerically by training a Euclidean symmetry equivariant neural network to learn symmetry breaking input to deform a square into a rectangle and to generate octahedra tilting patterns in perovskites.}, + keywords = {TODO}, + file = {/home/johannes/Nextcloud/Zotero/Smidt et al_2021_Finding symmetry breaking order parameters with Euclidean neural networks.pdf;/home/johannes/Zotero/storage/TUYBWD9X/Smidt et al. - 2021 - Finding symmetry breaking order parameters with Eu.pdf;/home/johannes/Zotero/storage/4XA7SRBQ/PhysRevResearch.3.html} +} + +@inproceedings{smithMachineLearningBazaar2020, + title = {The {{Machine Learning Bazaar}}: {{Harnessing}} the {{ML Ecosystem}} for {{Effective System Development}}}, + shorttitle = {The {{Machine Learning Bazaar}}}, + booktitle = {Proceedings of the 2020 {{ACM SIGMOD International Conference}} on {{Management}} of {{Data}}}, + author = {Smith, Micah J. and Sala, Carles and Kanter, James Max and Veeramachaneni, Kalyan}, + date = {2020-06-11}, + pages = {785--800}, + publisher = {{ACM}}, + location = {{Portland OR USA}}, + doi = {10.1145/3318464.3386146}, + url = {https://dl.acm.org/doi/10.1145/3318464.3386146}, + urldate = {2021-10-08}, + eventtitle = {{{SIGMOD}}/{{PODS}} '20: {{International Conference}} on {{Management}} of {{Data}}}, + isbn = {978-1-4503-6735-6}, + langid = {english}, + file = {/home/johannes/Nextcloud/Zotero/Smith et al_2020_The Machine Learning Bazaar.pdf} +} + +@article{snyderFindingDensityFunctionals2012, + title = {Finding {{Density Functionals}} with {{Machine Learning}}}, + author = {Snyder, John C. and Rupp, Matthias and Hansen, Katja and Müller, Klaus-Robert and Burke, Kieron}, + date = {2012-06-19}, + journaltitle = {Physical Review Letters}, + shortjournal = {Phys. Rev. Lett.}, + volume = {108}, + number = {25}, + pages = {253002}, + publisher = {{American Physical Society}}, + doi = {10.1103/PhysRevLett.108.253002}, + url = {https://link.aps.org/doi/10.1103/PhysRevLett.108.253002}, + urldate = {2021-10-15}, + abstract = {Machine learning is used to approximate density functionals. For the model problem of the kinetic energy of noninteracting fermions in 1D, mean absolute errors below 1 kcal/mol on test densities similar to the training set are reached with fewer than 100 training densities. A predictor identifies if a test density is within the interpolation region. Via principal component analysis, a projected functional derivative finds highly accurate self-consistent densities. The challenges for application of our method to real electronic structure problems are discussed.}, + keywords = {_tablet,2-step model,DFT,dimensionality reduction,KRR,ML,ML-DFA,ML-DFT,ML-ESM,ML-OF,models,orbital-free DFT,original publication,PCA,prediction from density,prediction from density functional,prediction of electron density,prediction of kinetic energy}, + file = {/home/johannes/Nextcloud/Zotero/Snyder et al_2012_Finding Density Functionals with Machine Learning.pdf;/home/johannes/Zotero/storage/6NMNCTQB/Snyder et al. - 2012 - Finding Density Functionals with Machine Learning.tex;/home/johannes/Zotero/storage/TBZPF93I/Snyder et al_2012_Finding Density Functionals with Machine Learning.pdf;/home/johannes/Zotero/storage/RRS5SC4P/PhysRevLett.108.html} +} + +@article{soiland-reyesPackagingResearchArtefacts2022, + title = {Packaging Research Artefacts with {{RO-Crate}}}, + author = {Soiland-Reyes, Stian and Sefton, Peter and Crosas, Mercè and Castro, Leyla Jael and Coppens, Frederik and Fernández, José M. and Garijo, Daniel and Grüning, Björn and La Rosa, Marco and Leo, Simone and Ó Carragáin, Eoghan and Portier, Marc and Trisovic, Ana and RO-Crate Community and Groth, Paul and Goble, Carole}, + date = {2022-01-01}, + journaltitle = {Data Science}, + volume = {Preprint}, + pages = {1--42}, + publisher = {{IOS Press}}, + issn = {2451-8484}, + doi = {10.3233/DS-210053}, + url = {https://content.iospress.com/articles/data-science/ds210053}, + urldate = {2022-05-24}, + abstract = {An increasing number of researchers support reproducibility by including pointers to and descriptions of datasets, software and methods in their publications. However, scientific articles may be ambiguous, incomplete and difficult to process by autom}, + issue = {Preprint}, + langid = {english}, + keywords = {FAIR,FDO,JSON-LD,PID,RDM,workflows}, + file = {/home/johannes/Nextcloud/Zotero/Soiland-Reyes et al_2022_Packaging research artefacts with RO-Crate.pdf;/home/johannes/Zotero/storage/X2IWHLC7/ds210053.html} +} + +@inproceedings{souzaProvenanceDataMachine2019, + title = {Provenance {{Data}} in the {{Machine Learning Lifecycle}} in {{Computational Science}} and {{Engineering}}}, + booktitle = {2019 {{IEEE}}/{{ACM Workflows}} in {{Support}} of {{Large-Scale Science}} ({{WORKS}})}, + author = {Souza, Renan and Azevedo, Leonardo and Lourenço, VÃtor and Soares, Elton and Thiago, Raphael and Brandão, Rafael and Civitarese, Daniel and Brazil, Emilio and Moreno, Marcio and Valduriez, Patrick and Mattoso, Marta and Cerqueira, Renato and Netto, Marco A.S.}, + date = {2019-11}, + pages = {1--10}, + doi = {10.1109/WORKS49585.2019.00006}, + abstract = {Machine Learning (ML) has become essential in several industries. In Computational Science and Engineering (CSE), the complexity of the ML lifecycle comes from the large variety of data, scientists' expertise, tools, and workflows. If data are not tracked properly during the lifecycle, it becomes unfeasible to recreate a ML model from scratch or to explain to stackholders how it was created. The main limitation of provenance tracking solutions is that they cannot cope with provenance capture and integration of domain and ML data processed in the multiple workflows in the lifecycle, while keeping the provenance capture overhead low. To handle this problem, in this paper we contribute with a detailed characterization of provenance data in the ML lifecycle in CSE; a new provenance data representation, called PROV-ML, built on top of W3C PROV and ML Schema; and extensions to a system that tracks provenance from multiple workflows to address the characteristics of ML and CSE, and to allow for provenance queries with a standard vocabulary. We show a practical use in a real case in the O\&G industry, along with its evaluation using 239,616 CUDA cores in parallel.}, + eventtitle = {2019 {{IEEE}}/{{ACM Workflows}} in {{Support}} of {{Large-Scale Science}} ({{WORKS}})}, + keywords = {Computational Science and Engineering,Machine Learning Lifecycle,Workflow Provenance}, + file = {/home/johannes/Nextcloud/Zotero/Souza et al_2019_Provenance Data in the Machine Learning Lifecycle in Computational Science and.pdf;/home/johannes/Zotero/storage/NXAA6T76/8943505.html} +} + +@book{spaldinMagneticMaterialsFundamentals2010, + title = {Magnetic {{Materials}}: {{Fundamentals}} and {{Applications}}}, + shorttitle = {Magnetic {{Materials}}}, + author = {Spaldin, Nicola A.}, + date = {2010}, + edition = {2}, + publisher = {{Cambridge University Press}}, + location = {{Cambridge}}, + doi = {10.1017/CBO9780511781599}, + url = {https://www.cambridge.org/core/books/magnetic-materials/4C8C2C5DF32C9E8D528E1E8D26381C1F}, + urldate = {2022-08-30}, + abstract = {Magnetic Materials is an excellent introduction to the basics of magnetism, magnetic materials and their applications in modern device technologies. Retaining the concise style of the original, this edition has been thoroughly revised to address significant developments in the field, including the improved understanding of basic magnetic phenomena, new classes of materials, and changes to device paradigms. With homework problems, solutions to selected problems and a detailed list of references, Magnetic Materials continues to be the ideal book for a one-semester course and as a self-study guide for researchers new to the field. New to this edition:Entirely new chapters on Exchange Bias Coupling, Multiferroic and Magnetoelectric Materials, Magnetic InsulatorsRevised throughout, with substantial updates to the chapters on Magnetic Recording and Magnetic Semiconductors, incorporating the latest advances in the fieldNew example problems with worked solutions}, + isbn = {978-0-521-88669-7}, + keywords = {condensed matter,graduate,magnetism,textbook}, + file = {/home/johannes/Nextcloud/Zotero/Spaldin_2010_Magnetic Materials.pdf;/home/johannes/Zotero/storage/3A42DP7U/4C8C2C5DF32C9E8D528E1E8D26381C1F.html} +} + +@misc{spencerBetterFasterFermionic2020, + title = {Better, {{Faster Fermionic Neural Networks}}}, + author = {Spencer, James S. and Pfau, David and Botev, Aleksandar and Foulkes, W. M. C.}, + date = {2020-11-13}, + number = {arXiv:2011.07125}, + eprint = {2011.07125}, + eprinttype = {arxiv}, + primaryclass = {physics}, + publisher = {{arXiv}}, + doi = {10.48550/arXiv.2011.07125}, + url = {http://arxiv.org/abs/2011.07125}, + urldate = {2022-06-25}, + abstract = {The Fermionic Neural Network (FermiNet) is a recently-developed neural network architecture that can be used as a wavefunction Ansatz for many-electron systems, and has already demonstrated high accuracy on small systems. Here we present several improvements to the FermiNet that allow us to set new records for speed and accuracy on challenging systems. We find that increasing the size of the network is sufficient to reach chemical accuracy on atoms as large as argon. Through a combination of implementing FermiNet in JAX and simplifying several parts of the network, we are able to reduce the number of GPU hours needed to train the FermiNet on large systems by an order of magnitude. This enables us to run the FermiNet on the challenging transition of bicyclobutane to butadiene and compare against the PauliNet on the automerization of cyclobutadiene, and we achieve results near the state of the art for both.}, + archiveprefix = {arXiv}, + keywords = {autodiff,DeepMind,FermiNet,JAX,library,MC,ML,ML-ESM,ML-QMBP,NN,PauliNet,prediction of wavefunction,QMC,VMC,with-code}, + file = {/home/johannes/Nextcloud/Zotero/Spencer et al_2020_Better, Faster Fermionic Neural Networks.pdf;/home/johannes/Zotero/storage/SCSQGZ4K/2011.html} +} + +@article{spitalerPerspectivesTheoryDefects2018, + title = {Perspectives on the {{Theory}} of {{Defects}}}, + author = {Spitaler, Jürgen and Estreicher, Stefan K.}, + date = {2018}, + journaltitle = {Frontiers in Materials}, + shortjournal = {Front. Mater.}, + volume = {5}, + publisher = {{Frontiers}}, + issn = {2296-8016}, + doi = {10.3389/fmats.2018.00070}, + url = {https://www.frontiersin.org/articles/10.3389/fmats.2018.00070/full#h3}, + urldate = {2021-05-21}, + abstract = {Our understanding of defects in materials science has changed tremendously over the last century. While one hundred years ago they were often ignored by scientists, nowadays they are in the spotlight of scientific interest and whole branches of technology have emerged from their skillful handling. The first part of this article gives a historical overview and discusses why defects are so important for modern material science. In the second part, we review the treatment of defects in semiconductors. We start by explaining the assumptions and approximations involved in ab-initio calculations and then discuss the treatment of defects in materials. In the third part, we focus on defects in metals. We discuss the theoretical treatment of vacancies in metals starting from experimental findings. The impact of improved theoretical techniques on the predictive power is discussed. This is illustrated with the role of vacancies in intermetallic compounds and random alloys. The last section deals with dislocations.}, + langid = {english}, + keywords = {defects,First-principles theory,Metals and alloys,review,Semiconductors}, + file = {/home/johannes/Nextcloud/Zotero/Spitaler_Estreicher_2018_Perspectives on the Theory of Defects.pdf} +} + +@article{staackeKernelChargeEquilibration2022, + title = {Kernel Charge Equilibration: Efficient and Accurate Prediction of Molecular Dipole Moments with a Machine-Learning Enhanced Electron Density Model}, + shorttitle = {Kernel Charge Equilibration}, + author = {Staacke, Carsten G. and Wengert, Simon and Kunkel, Christian and Csányi, Gábor and Reuter, Karsten and Margraf, Johannes T.}, + date = {2022-03}, + journaltitle = {Machine Learning: Science and Technology}, + shortjournal = {Mach. Learn.: Sci. Technol.}, + volume = {3}, + number = {1}, + pages = {015032}, + publisher = {{IOP Publishing}}, + issn = {2632-2153}, + doi = {10.1088/2632-2153/ac568d}, + url = {https://doi.org/10.1088/2632-2153/ac568d}, + urldate = {2022-09-30}, + abstract = {State-of-the-art machine learning (ML) interatomic potentials use local representations of atomic environments to ensure linear scaling and size-extensivity. This implies a neglect of long-range interactions, most prominently related to electrostatics. To overcome this limitation, we herein present a ML framework for predicting charge distributions and their interactions termed kernel charge equilibration (kQEq). This model is based on classical charge equilibration (QEq) models expanded with an environment-dependent electronegativity. In contrast to previously reported neural network models with a similar concept, kQEq takes advantage of the linearity of both QEq and Kernel Ridge Regression to obtain a closed-form linear algebra expression for training the models. Furthermore, we avoid the ambiguity of charge partitioning schemes by using dipole moments as reference data. As a first application, we show that kQEq can be used to generate accurate and highly data-efficient models for molecular dipole moments.}, + langid = {english}, + keywords = {charge equilibration,charge transfer,electronegativity,kQEq,KRR,ML,molecules}, + file = {/home/johannes/Nextcloud/Zotero/Staacke et al_2022_Kernel charge equilibration.pdf} +} + +@unpublished{steinbachReproducibilityDataScience2022, + type = {presentation}, + title = {Reproducibility in {{Data Science}} and {{Machine Learning}}}, + author = {Steinbach, Peter}, + date = {2022-06-09}, + publisher = {{figshare}}, + doi = {10.6084/m9.figshare.20036651.v1}, + url = {https://figshare.com/articles/presentation/Reproducibility_in_Data_Science_and_Machine_Learning/20036651/1}, + urldate = {2022-06-09}, + abstract = {Machine Learning is becoming ubiquitous in many scientific domains. However, practitioners struggle to apply every new addition to the Machine Learning market on their data with comparable effects than published. In this talk, I'd like to present recent observations on reproducibility of Machine Learning results and how the community strives to tackle related challenges. Given at https://events.hifis.net/event/426/timetable/}, + langid = {english}, + file = {/home/johannes/Nextcloud/Zotero/Steinbach_2022_Reproducibility in Data Science and Machine Learning.pdf;/home/johannes/Zotero/storage/MPANKYEK/1.html} +} + +@article{suttonIdentifyingDomainsApplicability2020, + title = {Identifying Domains of Applicability of Machine Learning Models for Materials Science}, + author = {Sutton, Christopher and Boley, Mario and Ghiringhelli, Luca M. and Rupp, Matthias and Vreeken, Jilles and Scheffler, Matthias}, + date = {2020-09-04}, + journaltitle = {Nature Communications}, + volume = {11}, + number = {1}, + pages = {4428}, + publisher = {{Nature Publishing Group}}, + issn = {2041-1723}, + doi = {10.1038/s41467-020-17112-9}, + url = {https://www.nature.com/articles/s41467-020-17112-9}, + urldate = {2021-05-19}, + abstract = {Although machine learning (ML) models promise to substantially accelerate the discovery of novel materials, their performance is often still insufficient to draw reliable conclusions. Improved ML models are therefore actively researched, but their design is currently guided mainly by monitoring the average model test error. This can render different models indistinguishable although their performance differs substantially across materials, or it can make a model appear generally insufficient while it actually works well in specific sub-domains. Here, we present a method, based on subgroup discovery, for detecting domains of applicability (DA) of models within a materials class. The utility of this approach is demonstrated by analyzing three state-of-the-art ML models for predicting the formation energy of transparent conducting oxides. We find that, despite having a mutually indistinguishable and unsatisfactory average error, the models have DAs with distinctive features and notably improved performance.}, + issue = {1}, + langid = {english}, + keywords = {descriptors,Domains of applicability (DA),ML,models,subgroup discovery}, + file = {/home/johannes/Nextcloud/Zotero/Sutton et al_2020_Identifying domains of applicability of machine learning models for materials.pdf;/home/johannes/Zotero/storage/MPRCKUUI/s41467-020-17112-9.html} +} + +@article{talirzMaterialsCloudPlatform2020, + title = {Materials {{Cloud}}, a Platform for Open Computational Science}, + author = {Talirz, Leopold and Kumbhar, Snehal and Passaro, Elsa and Yakutovich, Aliaksandr V. and Granata, Valeria and Gargiulo, Fernando and Borelli, Marco and Uhrin, Martin and Huber, Sebastiaan P. and Zoupanos, Spyros and Adorf, Carl S. and Andersen, Casper Welzel and Schütt, Ole and Pignedoli, Carlo A. and Passerone, Daniele and VandeVondele, Joost and Schulthess, Thomas C. and Smit, Berend and Pizzi, Giovanni and Marzari, Nicola}, + date = {2020-09-08}, + journaltitle = {Scientific Data}, + shortjournal = {Sci Data}, + volume = {7}, + number = {1}, + pages = {299}, + publisher = {{Nature Publishing Group}}, + issn = {2052-4463}, + doi = {10.1038/s41597-020-00637-5}, + url = {https://www.nature.com/articles/s41597-020-00637-5}, + urldate = {2021-10-15}, + abstract = {Materials Cloud is a platform designed to enable open and seamless sharing of resources for computational science, driven by applications in materials modelling. It hosts (1) archival and dissemination services for raw and curated data, together with their provenance graph, (2) modelling services and virtual machines, (3) tools for data analytics, and pre-/post-processing, and (4) educational materials. Data is citable and archived persistently, providing a comprehensive embodiment of entire simulation pipelines (calculations performed, codes used, data generated) in the form of graphs that allow retracing and reproducing any computed result. When an AiiDA database is shared on Materials Cloud, peers can browse the interconnected record of simulations, download individual files or the full database, and start their research from the results of the original authors. The infrastructure is agnostic to the specific simulation codes used and can support diverse applications in computational science that transcend its initial materials domain.}, + issue = {1}, + langid = {english}, + annotation = {Bandiera\_abtest: a Cc\_license\_type: cc\_by Cg\_type: Nature Research Journals Primary\_atype: Research Subject\_term: Databases;Materials science Subject\_term\_id: databases;materials-science}, + file = {/home/johannes/Nextcloud/Zotero/Talirz et al_2020_Materials Cloud, a platform for open computational science.pdf;/home/johannes/Zotero/storage/TEZC6LT2/s41597-020-00637-5.html} +} + +@unpublished{talirzTrendsAtomisticSimulation2021, + title = {Trends in Atomistic Simulation Software Usage}, + author = {Talirz, Leopold and Ghiringhelli, Luca M. and Smit, Berend}, + date = {2021-08-27}, + eprint = {2108.12350}, + eprinttype = {arxiv}, + primaryclass = {cond-mat, physics:physics}, + url = {http://arxiv.org/abs/2108.12350}, + urldate = {2021-09-11}, + abstract = {Driven by the unprecedented computational power available to scientific research, the use of computers in solid-state physics, chemistry and materials science has been on a continuous rise. This review focuses on the software used for the simulation of matter at the atomic scale. We provide a comprehensive overview of major codes in the field, and analyze how citations to these codes in the academic literature have evolved since 2010. An interactive version of the underlying data set is available at https://atomistic.software .}, + archiveprefix = {arXiv}, + keywords = {DFT,DFT codes comparison}, + file = {/home/johannes/Nextcloud/Zotero/Talirz et al_2021_Trends in atomistic simulation software usage.pdf;/home/johannes/Zotero/storage/SCDVYPXG/2108.html} +} + +@article{tealeDFTExchangeSharing2022, + title = {{{DFT Exchange}}: {{Sharing Perspectives}} on the {{Workhorse}} of {{Quantum Chemistry}} and {{Materials Science}}}, + shorttitle = {{{DFT Exchange}}}, + author = {Teale, Andrew and Helgaker, Trygve and Savin, Andreas and Adamo, Carlo and Aradi, Balint and Arbuznikov, Alexei and Ayers, Paul and Baerends, Evert Jan and Barone, Vincenzo and Calaminici, Patrizia and Cances, Eric and Carter, Emily and Chattaraj, Pratim and Chermette, Henry and Ciofini, Ilaria and Crawford, Daniel and Proft, Frank De and Dobson, John and Draxl, Caludia and Frauenheim, Thomas and Fromager, Emmanuel and Fuentealba, Patricio and Gagliardi, Laura and Galli, Giulia and Gao, Jiali and Geerlings, Paul and Gidopoulous, Nikitas and Gill, Peter and Gori-Giorgi, Paola and Gorling, Andreas and Gould, TIm and Grimme, Stefan and Gritsenko, Oleg and Jensen, Hans Jorgen and Johnson, Erin and Jones, Robert and Kaupp, Martin and Koster, Andreas and Kronik, Leeor and Krylov, Anna and Kvaal, Simen and Laestadius, Andre and Levy, Mel and Lewin, Mathieu and Liu, Shubin and Loos, Pierre-Francois and Maitra, Neepa and Neese, Frank and Perdew, John and Pernal, Katarzyna and Pernot, Pascal and Piecuch, Piotr and Rebolini, Elisa and Reining, Lucia and Romaniello, Pina and Ruzsinszky, Adrienn and Salahub, Dennis and Scheffler, Matthias and Schwerdtfeger, Peter and Staroverov, Vicktor and Sun, Jianwei and Tellgren, Erik and Tozer, David and Trickey, Sam and Ullrich, Carsten and Vela, Alberto and Vignale, Giovanni and Wesolowski, Tomasz and Xu, Xin and Yang, Weitao}, + date = {2022-06-17}, + doi = {10.26434/chemrxiv-2022-13j2v}, + url = {https://chemrxiv.org/engage/chemrxiv/article-details/62974da519595958f0bcc339}, + urldate = {2022-06-23}, + abstract = {In this paper, the history, present status, and future of density-functional theory (DFT) is informally reviewed and discussed by 70 workers in the field, including molecular scientists, materials scientists, method developers and practitioners. The format of the paper is that of a roundtable discussion, in which the participants express and exchange views on DFT in the form of 300 individual contributions, formulated as responses to a preset list of 26 questions. Supported by a bibliography of 776 entries, the paper represents a broad snapshot of DFT, anno 2022.}, + langid = {english}, + keywords = {DFT,PGI-1/IAS-1,review}, + file = {/home/johannes/Nextcloud/Zotero/Teale et al_2022_DFT Exchange.pdf;/home/johannes/Zotero/storage/JIDYX9CC/62974da519595958f0bcc339.html} +} + +@thesis{thiessDevelopmentApplicationMassively2011, + title = {Development and Application of a Massively Parallel {{KKR Green}} Function Method for Large Scale Systems}, + author = {Thieß, Alexander R. and Blügel, Stefan}, + date = {2011}, + institution = {{Publikationsserver der RWTH Aachen University}}, + location = {{Aachen}}, + langid = {english}, + pagetotal = {173}, + keywords = {density functional theory,Dichtefunktional,dilute magnetic semiconductors,Festkörperphysik,KKR-Methode,Korring Kohn Rostoker Green functions,phase change materials,Phase-Change-Technologie,Physik,Supercomputer,supercomputing,Verdünnte magnetische Legierung} +} + +@thesis{thiessDevelopmentApplicationMassively2013, + title = {Development and Application of a Massively Parallel {{KKR Green}} Function Method for Large Scale Systems}, + author = {Thieß, Alexander R.}, + date = {2013}, + number = {PreJuSER-19395}, + institution = {{Forschungszentrum Jülich GmbH Zentralbibliothek, Verlag}}, + url = {http://hdl.handle.net/2128/18578}, + urldate = {2022-08-12}, + abstract = {The impact of structural and functional materials on society is often overlooked but can in fact hardly be overestimated: In numerous examples, ranging from the improvement of steel to the invention of light emitting diodes, carbon fibers as well as cheaper and larger memories for data storage, novel materials are a key to successfully face global challenges on mobility, energy, communication and sustainability. Most strikingly visible is this influence for technologies based on electronic, optical, and magnetic materials, technologies that revo- lutionize computing and communication excelling mankind into the information age. With the miniaturization of devices, made possible by the invention of the transistor and the integrated circuit, enormous and still exponentially growing computing and communication capabilities are fundamentally changing how we interact, work and live. Material science and condensed matter physics are at the heart of the invention, development, design and improvement of novel materials and subsequently of novel physical phenomena and processes and are thus an excellent demonstration of the interdependence of science, technology and society. Advances in modern material design and technology are closely linked to advances in understanding on the basis of condensed matter physics, statistical physics and quantum mechanics of the many particle problem as well as the development of powerful methods. High-performance experimental tools combined with extraordinary progress in theory and computational power provide insight on the microscopic phenomena in materials and have paved new roads towards understanding as well as raising and answering new questions. On the theory side, density functional theory takes a central position in this process. The ab initio description of materials from the first principles of quantum mechanics holds fun- damental and highly valuable information on the interactions and interplay of electrons in solids and contributes such to the advancement of knowledge on the structural, mechanical, optical, thermal, electrical, magnetic, ferroic or transport properties in bulk solids, surfaces, thin films, heterostructures, quantum wells, clusters and molecules. The complicated task to compute material properties on the quantum mechanical level of myriad of atoms in solids became first accessible by exploiting the periodicity of crystalline solids and high symmetry of idealized systems. Density functional theory calculations exploiting the periodic boundary [...] Thieß, Alexander R.}, + isbn = {9783893369065}, + langid = {english}, + keywords = {juKKR,KKR,KKRnano,PGI-1/IAS-1,thesis}, + file = {/home/johannes/Nextcloud/Zotero/Thieß_2013_Development and application of a massively parallel KKR Green function method.pdf;/home/johannes/Zotero/storage/XL6HDNB2/19395.html} +} + +@article{thiessMassivelyParallelDensity2012, + title = {Massively Parallel Density Functional Calculations for Thousands of Atoms: {{KKRnano}}}, + shorttitle = {Massively Parallel Density Functional Calculations for Thousands of Atoms}, + author = {Thiess, A.}, + date = {2012}, + journaltitle = {Physical Review B}, + shortjournal = {Phys. Rev. B}, + volume = {85}, + number = {23}, + doi = {10.1103/PhysRevB.85.235103}, + keywords = {_tablet,juKKR,KKR,KKRnano,PGI-1/IAS-1}, + file = {/home/johannes/Nextcloud/Zotero/Thiess_2012_Massively parallel density functional calculations for thousands of atoms.pdf;/home/johannes/Zotero/storage/PM97ULPL/PhysRevB.85.html} +} + +@article{thiessMassivelyParallelDensity2012a, + title = {Massively Parallel Density Functional Calculations for Thousands of Atoms: {{KKRnano}}}, + shorttitle = {Massively Parallel Density Functional Calculations for Thousands of Atoms}, + author = {Thiess, A. and Zeller, Rudolf and Bolten, M. and Dederichs, Peter H. and Blügel, Stefan}, + date = {2012}, + journaltitle = {Physical Review B}, + shortjournal = {Phys. Rev. B}, + volume = {85}, + number = {23}, + doi = {10.1103/PhysRevB.85.235103}, + file = {/home/johannes/Nextcloud/Zotero/Thiess_2012_Massively parallel density functional calculations for thousands of atoms2.pdf;/home/johannes/Zotero/storage/NJSJUCGL/PhysRevB.85.html} +} + +@article{thompsonSpectralNeighborAnalysis2015, + title = {Spectral Neighbor Analysis Method for Automated Generation of Quantum-Accurate Interatomic Potentials}, + author = {Thompson, A. P. and Swiler, L. P. and Trott, C. R. and Foiles, S. M. and Tucker, G. J.}, + date = {2015-03-15}, + journaltitle = {Journal of Computational Physics}, + shortjournal = {Journal of Computational Physics}, + volume = {285}, + pages = {316--330}, + issn = {0021-9991}, + doi = {10.1016/j.jcp.2014.12.018}, + url = {https://www.sciencedirect.com/science/article/pii/S0021999114008353}, + urldate = {2021-12-05}, + abstract = {We present a new interatomic potential for solids and liquids called Spectral Neighbor Analysis Potential (SNAP). The SNAP potential has a very general form and uses machine-learning techniques to reproduce the energies, forces, and stress tensors of a large set of small configurations of atoms, which are obtained using high-accuracy quantum electronic structure (QM) calculations. The local environment of each atom is characterized by a set of bispectrum components of the local neighbor density projected onto a basis of hyperspherical harmonics in four dimensions. The bispectrum components are the same bond-orientational order parameters employed by the GAP potential [1]. The SNAP potential, unlike GAP, assumes a linear relationship between atom energy and bispectrum components. The linear SNAP coefficients are determined using weighted least-squares linear regression against the full QM training set. This allows the SNAP potential to be fit in a robust, automated manner to large QM data sets using many bispectrum components. The calculation of the bispectrum components and the SNAP potential are implemented in the LAMMPS parallel molecular dynamics code. We demonstrate that a previously unnoticed symmetry property can be exploited to reduce the computational cost of the force calculations by more than one order of magnitude. We present results for a SNAP potential for tantalum, showing that it accurately reproduces a range of commonly calculated properties of both the crystalline solid and the liquid phases. In addition, unlike simpler existing potentials, SNAP correctly predicts the energy barrier for screw dislocation migration in BCC tantalum.}, + langid = {english}, + keywords = {descriptors,GAP,LAMMPS,MD,ML,PES,SNAP}, + file = {/home/johannes/Nextcloud/Zotero/Thompson et al_2015_Spectral neighbor analysis method for automated generation of quantum-accurate.pdf} +} + +@unpublished{togoSpglibSoftwareLibrary2018, + title = {Spglib: A Software Library for Crystal Symmetry Search}, + shorttitle = {\$\textbackslash texttt\{\vphantom\}{{Spglib}}\vphantom\{\}\$}, + author = {Togo, Atsushi and Tanaka, Isao}, + date = {2018-08-05}, + number = {arXiv:1808.01590}, + eprint = {1808.01590}, + eprinttype = {arxiv}, + primaryclass = {cond-mat}, + publisher = {{arXiv}}, + doi = {10.48550/arXiv.1808.01590}, + url = {http://arxiv.org/abs/1808.01590}, + urldate = {2022-05-18}, + abstract = {A computer algorithm to search crystal symmetries of crystal structures has been implemented in software \$\textbackslash texttt\{spglib\}\$. An iterative algorithm is employed to find a set of space group operations that belongs to any one of space group types by accepting certain amount of distortion for input unit cell structures. The source code is distributed under the BSD 3-Clause License that is a permissive free software licence. Although \$\textbackslash texttt\{spglib\}\$ is a small code, the iteration loops made the source code complicated. The aim of this text is to provide the algorithm details to those people who are interested in inside-\$\textbackslash texttt\{spglib\}\$. This text is written for \$\textbackslash texttt\{spglib\}\$ v1.10.4.}, + archiveprefix = {arXiv}, + keywords = {condensed matter,crystal symmetry,library}, + file = {/home/johannes/Nextcloud/Zotero/Togo_Tanaka_2018_$-texttt Spglib $.pdf;/home/johannes/Zotero/storage/67C8WPLU/1808.html} +} + +@article{tokuraEmergentFunctionsQuantum2017, + title = {Emergent Functions of Quantum Materials}, + author = {Tokura, Yoshinori and Kawasaki, Masashi and Nagaosa, Naoto}, + date = {2017-11}, + journaltitle = {Nature Physics}, + shortjournal = {Nature Phys}, + volume = {13}, + number = {11}, + pages = {1056--1068}, + publisher = {{Nature Publishing Group}}, + issn = {1745-2481}, + doi = {10.1038/nphys4274}, + url = {https://www.nature.com/articles/nphys4274}, + urldate = {2021-08-24}, + abstract = {Materials can harbour quantum many-body systems, most typically in the form of strongly correlated electrons in solids, that lead to novel and remarkable functions thanks to emergence—collective behaviours that arise from strong interactions among the elements. These include the Mott transition, high-temperature superconductivity, topological superconductivity, colossal magnetoresistance, giant magnetoelectric effect, and topological insulators. These phenomena will probably be crucial for developing the next-generation quantum technologies that will meet the urgent technological demands for achieving a sustainable and safe society. Dissipationless electronics using topological currents and quantum spins, energy harvesting such as photovoltaics and thermoelectrics, and secure quantum computing and communication are the three major fields of applications working towards this goal. Here, we review the basic principles and the current status of the emergent phenomena and functions in materials from the viewpoint of strong correlation and topology.}, + issue = {11}, + langid = {english}, + annotation = {Bandiera\_abtest: a Cg\_type: Nature Research Journals Primary\_atype: Reviews Subject\_term: Electronic devices;Electronic properties and materials;Ferroelectrics and multiferroics;Superconducting properties and materials;Topological matter Subject\_term\_id: electronic-devices;electronic-properties-and-materials;ferroelectrics-and-multiferroics;superconducting-properties-and-materials;topological-matter}, + file = {/home/johannes/Nextcloud/Zotero/Tokura et al_2017_Emergent functions of quantum materials.pdf} +} + +@article{tokuraMagneticSkyrmionMaterials2021, + title = {Magnetic {{Skyrmion Materials}}}, + author = {Tokura, Yoshinori and Kanazawa, Naoya}, + date = {2021-03-10}, + journaltitle = {Chemical Reviews}, + shortjournal = {Chem. Rev.}, + volume = {121}, + number = {5}, + pages = {2857--2897}, + publisher = {{American Chemical Society}}, + issn = {0009-2665}, + doi = {10.1021/acs.chemrev.0c00297}, + url = {https://doi.org/10.1021/acs.chemrev.0c00297}, + urldate = {2021-08-23}, + abstract = {Skyrmion, a concept originally proposed in particle physics half a century ago, can now find the most fertile field for its applicability, that is, the magnetic skyrmion realized in helimagnetic materials. The spin swirling vortex-like texture of the magnetic skyrmion can define the particle nature by topology; that is, all the constituent spin moments within the two-dimensional sheet wrap the sphere just one time. Such a topological nature of the magnetic skyrmion can lead to extraordinary metastability via topological protection and the driven motion with low electric-current excitation, which may promise future application to spintronics. The skyrmions in the magnetic materials frequently show up as the crystal lattice form, e.g., hexagonal lattice, but sometimes as isolated or independent particles. These skyrmions in magnets were initially found in acentric magnets, such as chiral, polar, and bilayered magnets endowed with antisymmetric spin exchange interaction, while the skyrmion host materials have been explored in a broader family of compounds including centrosymmetric magnets. This review describes the materials science and materials chemistry of magnetic skyrmions using the classification scheme of the skyrmion forming microscopic mechanisms. The emergent phenomena and functions mediated by skyrmions are described, including the generation of emergent magnetic and electric field by statics and dynamics of skrymions and the inherent magnetoelectric effect. The other important magnetic topological defects in two or three dimensions, such as biskyrmions, antiskyrmions, merons, and hedgehogs, are also reviewed in light of their interplay with the skyrmions.} +} + +@article{townsendDataDrivenAccelerationCoupledCluster2019, + title = {Data-{{Driven Acceleration}} of the {{Coupled-Cluster Singles}} and {{Doubles Iterative Solver}}}, + author = {Townsend, Jacob and Vogiatzis, Konstantinos D.}, + date = {2019-07-18}, + journaltitle = {The Journal of Physical Chemistry Letters}, + shortjournal = {J. Phys. Chem. Lett.}, + volume = {10}, + number = {14}, + pages = {4129--4135}, + publisher = {{American Chemical Society}}, + doi = {10.1021/acs.jpclett.9b01442}, + url = {https://doi.org/10.1021/acs.jpclett.9b01442}, + urldate = {2022-05-13}, + abstract = {Solving the coupled-cluster (CC) equations is a cost-prohibitive process that exhibits poor scaling with system size. These equations are solved by determining the set of amplitudes (t) that minimize the system energy with respect to the coupled-cluster equations at the selected level of truncation. Here, a novel approach to predict the converged coupled-cluster singles and doubles (CCSD) amplitudes, thus the coupled-cluster wave function, is explored by using machine learning and electronic structure properties inherent to the MP2 level. Features are collected from quantum chemical data, such as orbital energies, one-electron Hamiltonian, Coulomb, and exchange terms. The data-driven CCSD (DDCCSD) is not an alchemical method because the actual iterative coupled-cluster equations are solved. However, accurate energetics can also be obtained by bypassing solving the CC equations entirely. Our preliminary data show that it is possible to achieve remarkable speedups in solving the CCSD equations, especially when the correct physics are encoded and used for training of machine learning models.}, + keywords = {ML,surrogate model}, + file = {/home/johannes/Nextcloud/Zotero/Townsend_Vogiatzis_2019_Data-Driven Acceleration of the Coupled-Cluster Singles and Doubles Iterative.pdf;/home/johannes/Zotero/storage/RVTRBAZI/acs.jpclett.html} +} + +@article{uhrinWorkflowsAiiDAEngineering2021, + title = {Workflows in {{AiiDA}}: {{Engineering}} a High-Throughput, Event-Based Engine for Robust and Modular Computational Workflows}, + shorttitle = {Workflows in {{AiiDA}}}, + author = {Uhrin, Martin and Huber, Sebastiaan P. and Yu, Jusong and Marzari, Nicola and Pizzi, Giovanni}, + date = {2021-02-01}, + journaltitle = {Computational Materials Science}, + shortjournal = {Computational Materials Science}, + volume = {187}, + pages = {110086}, + issn = {0927-0256}, + doi = {10.1016/j.commatsci.2020.110086}, + url = {https://www.sciencedirect.com/science/article/pii/S0927025620305772}, + urldate = {2021-06-29}, + abstract = {Over the last two decades, the field of computational science has seen a dramatic shift towards incorporating high-throughput computation and big-data analysis as fundamental pillars of the scientific discovery process. This has necessitated the development of tools and techniques to deal with the generation, storage and processing of large amounts of data. In this work we present an in-depth look at the workflow engine powering AiiDA, a widely adopted, highly flexible and database-backed informatics infrastructure with an emphasis on data reproducibility. We detail many of the design choices that were made which were informed by several important goals: the ability to scale from running on individual laptops up to high-performance supercomputers, managing jobs with runtimes spanning from fractions of a second to weeks and scaling up to thousands of jobs concurrently, and all this while maximising robustness. In short, AiiDA aims to be a Swiss army knife for high-throughput computational science. As well as the architecture, we outline important API design choices made to give workflow writers a great deal of liberty whilst guiding them towards writing robust and modular workflows, ultimately enabling them to encode their scientific knowledge to the benefit of the wider scientific community.}, + langid = {english}, + keywords = {AiiDA,Computational workflows,Data management,Data sharing,Database,Event-based,High-throughput,Provenance,Robust computation}, + file = {/home/johannes/Nextcloud/Zotero/Uhrin et al_2021_Workflows in AiiDA.pdf;/home/johannes/Zotero/storage/KDEGTQ46/S0927025620305772.html} +} + +@online{unitedstatesMaterialsGenomeInitiative, + title = {About the {{Materials Genome Initiative}}}, + author = {United States, National Science {and} Technology Council}, + url = {https://obamawhitehouse.archives.gov/mgi}, + urldate = {2021-10-15}, + abstract = {The Materials Genome Initiative is a multi-agency initiative designed to create a new era of policy, resources, and infrastructure that support U.S. institutions in the effort to discover, manufacture, and deploy advanced materials twice as fast, at a fraction of the cost.}, + langid = {english}, + organization = {{The White House}}, + file = {/home/johannes/Zotero/storage/LEWHVD66/Materials Genome Initiative for Global Competitiveness.pdf;/home/johannes/Zotero/storage/9KCC6KRJ/mgi.html} +} + +@article{unkeMachineLearningForce2021, + title = {Machine {{Learning Force Fields}}}, + author = {Unke, Oliver T. and Chmiela, Stefan and Sauceda, Huziel E. and Gastegger, Michael and Poltavsky, Igor and Schütt, Kristof T. and Tkatchenko, Alexandre and Müller, Klaus-Robert}, + date = {2021-08-25}, + journaltitle = {Chemical Reviews}, + shortjournal = {Chem. Rev.}, + volume = {121}, + number = {16}, + pages = {10142--10186}, + publisher = {{American Chemical Society}}, + issn = {0009-2665}, + doi = {10.1021/acs.chemrev.0c01111}, + url = {https://doi.org/10.1021/acs.chemrev.0c01111}, + urldate = {2021-10-22}, + abstract = {In recent years, the use of machine learning (ML) in computational chemistry has enabled numerous advances previously out of reach due to the computational complexity of traditional electronic-structure methods. One of the most promising applications is the construction of ML-based force fields (FFs), with the aim to narrow the gap between the accuracy of ab initio methods and the efficiency of classical FFs. The key idea is to learn the statistical relation between chemical structure and potential energy without relying on a preconceived notion of fixed chemical bonds or knowledge about the relevant interactions. Such universal ML approximations are in principle only limited by the quality and quantity of the reference data used to train them. This review gives an overview of applications of ML-FFs and the chemical insights that can be obtained from them. The core concepts underlying ML-FFs are described in detail, and a step-by-step guide for constructing and testing them from scratch is given. The text concludes with a discussion of the challenges that remain to be overcome by the next generation of ML-FFs.}, + keywords = {chemistry,GAP,GDML,GPR,hyperparameters,ML,ML-DFT,ML-ESM,ML-FF,MLP,models,MPNN,NNP,PES,regression,regularization,review,SchNet,sGDML,tutorial}, + file = {/home/johannes/Nextcloud/Zotero/Unke et al_2021_Machine Learning Force Fields2.pdf} +} + +@inproceedings{unkeSEEquivariantPrediction2021, + title = {{{SE}}(3)-Equivariant Prediction of Molecular Wavefunctions and Electronic Densities}, + booktitle = {Advances in {{Neural Information Processing Systems}}}, + author = {Unke, Oliver and Bogojeski, Mihail and Gastegger, Michael and Geiger, Mario and Smidt, Tess and Müller, Klaus-Robert}, + date = {2021}, + volume = {34}, + pages = {14434--14447}, + publisher = {{Curran Associates, Inc.}}, + url = {https://proceedings.neurips.cc/paper/2021/hash/78f1893678afbeaa90b1fa01b9cfb860-Abstract.html}, + urldate = {2022-08-21}, + abstract = {Machine learning has enabled the prediction of quantum chemical properties with high accuracy and efficiency, allowing to bypass computationally costly ab initio calculations. Instead of training on a fixed set of properties, more recent approaches attempt to learn the electronic wavefunction (or density) as a central quantity of atomistic systems, from which all other observables can be derived. This is complicated by the fact that wavefunctions transform non-trivially under molecular rotations, which makes them a challenging prediction target. To solve this issue, we introduce general SE(3)-equivariant operations and building blocks for constructing deep learning architectures for geometric point cloud data and apply them to reconstruct wavefunctions of atomistic systems with unprecedented accuracy. Our model achieves speedups of over three orders of magnitude compared to ab initio methods and reduces prediction errors by up to two orders of magnitude compared to the previous state-of-the-art. This accuracy makes it possible to derive properties such as energies and forces directly from the wavefunction in an end-to-end manner. We demonstrate the potential of our approach in a transfer learning application, where a model trained on low accuracy reference wavefunctions implicitly learns to correct for electronic many-body interactions from observables computed at a higher level of theory. Such machine-learned wavefunction surrogates pave the way towards novel semi-empirical methods, offering resolution at an electronic level while drastically decreasing computational cost. Additionally, the predicted wavefunctions can serve as initial guess in conventional ab initio methods, decreasing the number of iterations required to arrive at a converged solution, thus leading to significant speedups without any loss of accuracy or robustness. While we focus on physics applications in this contribution, the proposed equivariant framework for deep learning on point clouds is promising also beyond, say, in computer vision or graphics.}, + keywords = {_tablet,EGNN,equivariant,initial guess,ML-ESM,original publication,PhiSNet,prediction of electron density,prediction of wavefunction,SchNOrb,with-review}, + file = {/home/johannes/Nextcloud/Zotero/Unke et al_2021_SE(3)-equivariant prediction of molecular wavefunctions and electronic densities.pdf} +} + +@article{vannoordenTop100Papers2014, + title = {The Top 100 Papers}, + author = {Van Noorden, Richard and Maher, Brendan and Nuzzo, Regina}, + date = {2014-10-30}, + journaltitle = {Nature News}, + volume = {514}, + number = {7524}, + pages = {550}, + doi = {10.1038/514550a}, + url = {http://www.nature.com/news/the-top-100-papers-1.16224}, + urldate = {2022-10-05}, + abstract = {Nature explores the most-cited research of all time.}, + langid = {english}, + keywords = {DFT,for introductions,literature analysis,Nature,popular science}, + annotation = {Cg\_type: Nature News}, + file = {/home/johannes/Zotero/storage/5L7C8V9H/the-top-100-papers-1.html} +} + +@online{vanrossumPEPStyleGuide, + title = {{{PEP}} 8 -- {{Style Guide}} for {{Python Code}}}, + author = {van Rossum, Guido and Warsaw, Barry and Coghlan, Nick}, + options = {useprefix=true}, + url = {https://www.python.org/dev/peps/pep-0008/}, + urldate = {2021-09-23}, + abstract = {The official home of the Python Programming Language}, + langid = {english}, + organization = {{Python.org}}, + keywords = {coding style guide,PEP,Python,software engineering}, + file = {/home/johannes/Zotero/storage/A4H9CLJ5/pep-0008.html} +} + +@article{vanschorenOpenMLNetworkedScience2014, + title = {{{OpenML}}: Networked Science in Machine Learning}, + shorttitle = {{{OpenML}}}, + author = {Vanschoren, Joaquin and van Rijn, Jan N. and Bischl, Bernd and Torgo, Luis}, + options = {useprefix=true}, + date = {2014-06-16}, + journaltitle = {ACM SIGKDD Explorations Newsletter}, + shortjournal = {SIGKDD Explor. Newsl.}, + volume = {15}, + number = {2}, + eprint = {1407.7722}, + eprinttype = {arxiv}, + pages = {49--60}, + issn = {1931-0145, 1931-0153}, + doi = {10.1145/2641190.2641198}, + url = {http://arxiv.org/abs/1407.7722}, + urldate = {2022-01-02}, + abstract = {Many sciences have made significant breakthroughs by adopting online tools that help organize, structure and mine information that is too detailed to be printed in journals. In this paper, we introduce OpenML, a place for machine learning researchers to share and organize data in fine detail, so that they can work more effectively, be more visible, and collaborate with others to tackle harder problems. We discuss how OpenML relates to other examples of networked science and what benefits it brings for machine learning research, individual scientists, as well as students and practitioners.}, + archiveprefix = {arXiv}, + keywords = {Computer Science - Computers and Society,Computer Science - Machine Learning}, + file = {/home/johannes/Nextcloud/Zotero/Vanschoren et al_2014_OpenML.pdf;/home/johannes/Zotero/storage/YFGEM7US/1407.html} +} + +@inproceedings{vartakModelDBSystemMachine2016, + title = {{{ModelDB}}: A System for Machine Learning Model Management}, + shorttitle = {M{\textsc{odel}}{{DB}}}, + booktitle = {Proceedings of the {{Workshop}} on {{Human-In-the-Loop Data Analytics}}}, + author = {Vartak, Manasi and Subramanyam, Harihar and Lee, Wei-En and Viswanathan, Srinidhi and Husnoo, Saadiyah and Madden, Samuel and Zaharia, Matei}, + date = {2016-06-26}, + series = {{{HILDA}} '16}, + pages = {1--3}, + publisher = {{Association for Computing Machinery}}, + location = {{New York, NY, USA}}, + doi = {10.1145/2939502.2939516}, + url = {https://doi.org/10.1145/2939502.2939516}, + urldate = {2021-10-23}, + abstract = {Building a machine learning model is an iterative process. A data scientist will build many tens to hundreds of models before arriving at one that meets some acceptance criteria (e.g. AUC cutoff, accuracy threshold). However, the current style of model building is ad-hoc and there is no practical way for a data scientist to manage models that are built over time. As a result, the data scientist must attempt to "remember" previously constructed models and insights obtained from them. This task is challenging for more than a handful of models and can hamper the process of sensemaking. Without a means to manage models, there is no easy way for a data scientist to answer questions such as "Which models were built using an incorrect feature?", "Which model performed best on American customers?" or "How did the two top models compare?" In this paper, we describe our ongoing work on ModelDB, a novel end-to-end system for the management of machine learning models. ModelDB clients automatically track machine learning models in their native environments (e.g. scikit-learn, spark.ml), the ModelDB backend introduces a common layer of abstractions to represent models and pipelines, and the ModelDB frontend allows visual exploration and analyses of models via a web-based interface.}, + isbn = {978-1-4503-4207-0}, + file = {/home/johannes/Nextcloud/Zotero/Vartak et al_2016_Mspan class=smallcaps smallerCapitalodel-spanDB.pdf} +} + +@article{vedmedenko2020MagnetismRoadmap2020, + title = {The 2020 Magnetism Roadmap}, + author = {Vedmedenko, E. Y. and Kawakami, R. K. and Sheka, D. D. and Gambardella, P. and Kirilyuk, A. and Hirohata, A. and Binek, C. and Chubykalo-Fesenko, O. and Sanvito, S. and Kirby, B. J. and Grollier, J. and Everschor-Sitte, K. and Kampfrath, T. and You, C.-Y. and Berger, A.}, + date = {2020-08}, + journaltitle = {Journal of Physics D: Applied Physics}, + shortjournal = {J. Phys. D: Appl. Phys.}, + volume = {53}, + number = {45}, + pages = {453001}, + publisher = {{IOP Publishing}}, + issn = {0022-3727}, + doi = {10.1088/1361-6463/ab9d98}, + url = {https://doi.org/10.1088/1361-6463/ab9d98}, + urldate = {2022-09-20}, + abstract = {Andreas Berger CICnanoGUNE BRTA Following the success and relevance of the 2014 and 2017 Magnetism Roadmap articles, this 2020 Magnetism Roadmap edition takes yet another timely look at newly relevant and highly active areas in magnetism research. The overall layout of this article is unchanged, given that it has proved the most appropriate way to convey the most relevant aspects of today’s magnetism research in a wide variety of sub-fields to a broad readership. A different group of experts has again been selected for this article, representing both the breadth of new research areas, and the desire to incorporate different voices and viewpoints. The latter is especially relevant for thistype of article, in which one’s field of expertise has to be accommodated on two printed pages only, so that personal selection preferences are naturally rather more visible than in other types of articles. Most importantly, the very relevant advances in the field of magnetism research in recent years make the publication of yet another Magnetism Roadmap a very sensible and timely endeavour, allowing its authors and readers to take another broad-based, but concise look at the most significant developments in magnetism, their precise status, their challenges, and their anticipated future developments. While many of the contributions in this 2020 Magnetism Roadmap edition have significant associations with different aspects of magnetism, the general layout can nonetheless be classified in terms of three main themes: (i) phenomena, (ii) materials and characterization, and (iii) applications and devices. While these categories are unsurprisingly rather similar to the 2017 Roadmap, the order is different, in that the 2020 Roadmap considers phenomena first, even if their occurrences are naturally very difficult to separate from the materials exhibiting such phenomena. Nonetheless, the specifically selected topics seemed to be best displayed in the order presented here, in particular, because many of the phenomena or geometries discussed in (i) can be found or designed into a large variety of materials, so that the progression of the article embarks from more general concepts to more specific classes of materials in the selected order. Given that applications and devices are based on both phenomena and materials, it seemed most appropriate to close the article with the application and devices section (iii) once again. The 2020 Magnetism Roadmap article contains 14 sections, all of which were written by individual authors and experts, specifically addressing a subject in terms of its status, advances, challenges and perspectives in just two pages. Evidently, this two-page format limits the depth to which each subject can be described. Nonetheless, the most relevant and key aspects of each field are touched upon, which enables the Roadmap as whole to give its readership an initial overview of and outlook into a wide variety of topics and fields in a fairly condensed format. Correspondingly, the Roadmap pursues the goal of giving each reader a brief reference frame of relevant and current topics in modern applied magnetism research, even if not all sub-fields can be represented here. The first block of this 2020 Magnetism Roadmap, which is focussed on (i) phenomena, contains five contributions, which address the areas of interfacial Dzyaloshinskii–Moriya interactions, and two-dimensional and curvilinear magnetism, as well as spin-orbit torque phenomena and all optical magnetization reversal. All of these contributions describe cutting edge aspects of rather fundamental physical processes and properties, associated with new and improved magnetic materials’ properties, together with potential developments in terms of future devices and technology. As such, they form part of a widening magnetism ‘phenomena reservoir’ for utilization in applied magnetism and related device technology. The final block (iii) of this article focuses on such applications and device-related fields in four contributions relating to currently active areas of research, which are of course utilizing magnetic phenomena to enable specific functions. These contributions highlight the role of magnetism or spintronics in the field of neuromorphic and reservoir computing, terahertz technology, and domain wall-based logic. One aspect common to all of these application-related contributions is that they are not yet being utilized in commercially available technology; it is currently still an open question, whether or not such technological applications will be magnetism-based at all in the future, or if other types of materials and phenomena will yet outperform magnetism. This last point is actually a very good indication of the vibrancy of applied magnetism research today, given that it demonstrates that magnetism research is able to venture into novel application fields, based upon its portfolio of phenomena, effects and materials. This materials portfolio in particular defines the central block (ii) of this article, with its five contributions interconnecting phenomena with devices, for which materials and the characterization of their properties is the decisive discriminator between purely academically interesting aspects and the true viability of real-life devices, because only available materials and their associated fabrication and characterization methods permit reliable technological implementation. These five contributions specifically address magnetic films and multiferroic heterostructures for the purpose of spin electronic utilization, multi-scale materials modelling, and magnetic materials design based upon machine-learning, as well as materials characterization via polarized neutron measurements. As such, these contributions illustrate the balanced relevance of research into experimental and modelling magnetic materials, as well the importance of sophisticated characterization methods that allow for an ever-more refined understanding of materials. As a combined and integrated article, this 2020 Magnetism Roadmap is intended to be a reference point for current, novel and emerging research directions in modern magnetism, just as its 2014 and 2017 predecessors have been in previous years.}, + langid = {english}, + keywords = {magnetism,physics,review,roadmap}, + file = {/home/johannes/Nextcloud/Zotero/Vedmedenko et al_2020_The 2020 magnetism roadmap.pdf} +} + +@article{vojvodicExploringLimitsLowpressure2014, + title = {Exploring the Limits: {{A}} Low-Pressure, Low-Temperature {{Haber}}–{{Bosch}} Process}, + shorttitle = {Exploring the Limits}, + author = {Vojvodic, Aleksandra and Medford, Andrew James and Studt, Felix and Abild-Pedersen, Frank and Khan, Tuhin Suvra and Bligaard, T. and Nørskov, J. K.}, + date = {2014-04-08}, + journaltitle = {Chemical Physics Letters}, + shortjournal = {Chemical Physics Letters}, + volume = {598}, + pages = {108--112}, + issn = {0009-2614}, + doi = {10.1016/j.cplett.2014.03.003}, + url = {https://www.sciencedirect.com/science/article/pii/S000926141400147X}, + urldate = {2021-10-21}, + abstract = {The Haber–Bosch process for ammonia synthesis has been suggested to be the most important invention of the 20th century, and called the ‘Bellwether reaction in heterogeneous catalysis’. We examine the catalyst requirements for a new low-pressure, low-temperature synthesis process. We show that the absence of such a process for conventional transition metal catalysts can be understood as a consequence of a scaling relation between the activation energy for N2 dissociation and N adsorption energy found at the surface of these materials. A better catalyst cannot obey this scaling relation. We define the ideal scaling relation characterizing the most active catalyst possible, and show that it is theoretically possible to have a low pressure, low-temperature Haber–Bosch process. The challenge is to find new classes of catalyst materials with properties approaching the ideal, and we discuss the possibility that transition metal compounds have such properties.}, + langid = {english}, + keywords = {applications of DFT,DFT,master-thesis}, + file = {/home/johannes/Zotero/storage/86UUIB6S/S000926141400147X.html} +} + +@article{voskoAccurateSpindependentElectron1980, + title = {Accurate Spin-Dependent Electron Liquid Correlation Energies for Local Spin Density Calculations: A Critical Analysis}, + shorttitle = {Accurate Spin-Dependent Electron Liquid Correlation Energies for Local Spin Density Calculations}, + author = {Vosko, S. H. and Wilk, L. and Nusair, M.}, + date = {1980-08-01}, + journaltitle = {Canadian Journal of Physics}, + shortjournal = {Can. J. Phys.}, + volume = {58}, + number = {8}, + pages = {1200--1211}, + publisher = {{NRC Research Press}}, + issn = {0008-4204}, + doi = {10.1139/p80-159}, + url = {https://cdnsciencepub.com/doi/abs/10.1139/p80-159}, + urldate = {2021-10-18}, + keywords = {DFT,LDA,LSDA,original publication,xc functional}, + file = {/home/johannes/Nextcloud/Zotero/Vosko et al_1980_Accurate spin-dependent electron liquid correlation energies for local spin.pdf} +} + +@article{vuUnderstandingKernelRidge2015, + title = {Understanding Kernel Ridge Regression: {{Common}} Behaviors from Simple Functions to Density Functionals}, + shorttitle = {Understanding Kernel Ridge Regression}, + author = {Vu, Kevin and Snyder, John C. and Li, Li and Rupp, Matthias and Chen, Brandon F. and Khelif, Tarek and Müller, Klaus-Robert and Burke, Kieron}, + date = {2015}, + journaltitle = {International Journal of Quantum Chemistry}, + volume = {115}, + number = {16}, + pages = {1115--1128}, + issn = {1097-461X}, + doi = {10.1002/qua.24939}, + url = {https://onlinelibrary.wiley.com/doi/abs/10.1002/qua.24939}, + urldate = {2021-10-19}, + abstract = {Accurate approximations to density functionals have recently been obtained via machine learning (ML). By applying ML to a simple function of one variable without any random sampling, we extract the qualitative dependence of errors on hyperparameters. We find universal features of the behavior in extreme limits, including both very small and very large length scales, and the noise-free limit. We show how such features arise in ML models of density functionals. © 2015 Wiley Periodicals, Inc.}, + langid = {english}, + keywords = {_tablet,density functional theory,DFT,extreme behaviors,hyperparameters optimization,KRR,machine learning,ML,models,noise-free curve,tutorial}, + annotation = {\_eprint: https://onlinelibrary.wiley.com/doi/pdf/10.1002/qua.24939}, + file = {/home/johannes/Nextcloud/Zotero/false;/home/johannes/Nextcloud/Zotero/false;/home/johannes/Nextcloud/Zotero/Vu et al_2015_Understanding kernel ridge regression.pdf;/home/johannes/Zotero/storage/5INUIEQC/qua.html} +} + +@book{vvedenskySymmetryGroupsRepresentations2010, + title = {Symmetry, Groups, and Representations in Physics}, + author = {Vvedensky, Dimitri D. and Evans, Timothy S.}, + date = {2010}, + publisher = {{World Scientific}}, + location = {{Singapore}}, + abstract = {Presents an introduction to symmetry in physics based on discrete and continuous groups. This book includes exercises that illustrate the concepts introduced in the main text, to extend some of the main results, and to introduce fresh ideas. It is suitable for both beginning and advanced graduate students.}, + isbn = {978-1-84816-371-3}, + langid = {english}, + pagetotal = {350}, + keywords = {/unread}, + annotation = {OCLC: 633422775} +} + +@misc{wangGraphNetsPartial2019, + title = {Graph {{Nets}} for {{Partial Charge Prediction}}}, + author = {Wang, Yuanqing and Fass, Josh and Stern, Chaya D. and Luo, Kun and Chodera, John}, + date = {2019-09-17}, + number = {arXiv:1909.07903}, + eprint = {1909.07903}, + eprinttype = {arxiv}, + primaryclass = {physics}, + publisher = {{arXiv}}, + doi = {10.48550/arXiv.1909.07903}, + url = {http://arxiv.org/abs/1909.07903}, + urldate = {2022-09-27}, + abstract = {Atomic partial charges are crucial parameters for Molecular Dynamics (MD) simulations, molecular mechanics calculations, and virtual screening, as they determine the electrostatic contributions to interaction energies. Current methods for calculating partial charges, however, are either slow and scale poorly with molecular size (quantum chemical methods) or unreliable (empirical methods). Here, we present a new charge derivation method based on Graph Nets---a set of update and aggregate functions that operate on molecular topologies and propagate information thereon---that could approximate charges derived from Density Functional Theory (DFT) calculations with high accuracy and an over 500-fold speed up.}, + archiveprefix = {arXiv}, + keywords = {GCN,GNN,molecules,prediction of partial charge}, + file = {/home/johannes/Nextcloud/Zotero/Wang et al_2019_Graph Nets for Partial Charge Prediction.pdf;/home/johannes/Zotero/storage/5MD2WVP3/1909.html} +} + +@article{wangMachineLearningMaterials2020, + title = {Machine {{Learning}} for {{Materials Scientists}}: {{An Introductory Guide}} toward {{Best Practices}}}, + shorttitle = {Machine {{Learning}} for {{Materials Scientists}}}, + author = {Wang, Anthony Yu-Tung and Murdock, Ryan J. and Kauwe, Steven K. and Oliynyk, Anton O. and Gurlo, Aleksander and Brgoch, Jakoah and Persson, Kristin A. and Sparks, Taylor D.}, + date = {2020-06-23}, + journaltitle = {Chemistry of Materials}, + shortjournal = {Chem. Mater.}, + volume = {32}, + number = {12}, + pages = {4954--4965}, + publisher = {{American Chemical Society}}, + issn = {0897-4756}, + doi = {10.1021/acs.chemmater.0c01907}, + url = {https://doi.org/10.1021/acs.chemmater.0c01907}, + urldate = {2021-05-13}, + abstract = {This Methods/Protocols article is intended for materials scientists interested in performing machine learning-centered research. We cover broad guidelines and best practices regarding the obtaining and treatment of data, feature engineering, model training, validation, evaluation and comparison, popular repositories for materials data and benchmarking data sets, model and architecture sharing, and finally publication. In addition, we include interactive Jupyter notebooks with example Python code to demonstrate some of the concepts, workflows, and best practices discussed. Overall, the data-driven methods and machine learning workflows and considerations are presented in a simple way, allowing interested readers to more intelligently guide their machine learning research using the suggested references, best practices, and their own materials domain expertise.}, + keywords = {ML,models,notebooks,review,with-code}, + file = {/home/johannes/Nextcloud/Zotero/Wang et al_2020_Machine Learning for Materials Scientists.pdf;/home/johannes/Zotero/storage/PY7PFU35/acs.chemmater.html} +} + +@article{wangTopologicalStatesCondensed2017, + title = {Topological States of Condensed Matter}, + author = {Wang, Jing and Zhang, Shou-Cheng}, + date = {2017-11}, + journaltitle = {Nature Materials}, + shortjournal = {Nature Mater}, + volume = {16}, + number = {11}, + pages = {1062--1067}, + publisher = {{Nature Publishing Group}}, + issn = {1476-4660}, + doi = {10.1038/nmat5012}, + url = {https://www.nature.com/articles/nmat5012}, + urldate = {2021-08-24}, + abstract = {Topological states of quantum matter have been investigated intensively in recent years in materials science and condensed matter physics. The field developed explosively largely because of the precise theoretical predictions, well-controlled materials processing, and novel characterization techniques. In this Perspective, we review recent progress in topological insulators, the quantum anomalous Hall effect, chiral topological superconductors, helical topological superconductors and Weyl semimetals.}, + issue = {11}, + langid = {english}, + annotation = {Bandiera\_abtest: a Cg\_type: Nature Research Journals Primary\_atype: Reviews Subject\_term: Electronic properties and materials;Quantum Hall;Superconducting properties and materials;Topological matter Subject\_term\_id: electronic-properties-and-materials;quantum-hall;superconducting-properties-and-materials;topological-matter}, + file = {/home/johannes/Nextcloud/Zotero/Wang_Zhang_2017_Topological states of condensed matter.pdf} +} + +@article{wangWillAnyCrap2020, + title = {Will {{Any Crap We Put}} into {{Graphene Increase Its Electrocatalytic Effect}}?}, + author = {Wang, Lu and Sofer, Zdenek and Pumera, Martin}, + date = {2020-01-28}, + journaltitle = {ACS Nano}, + shortjournal = {ACS Nano}, + volume = {14}, + number = {1}, + pages = {21--25}, + publisher = {{American Chemical Society}}, + issn = {1936-0851}, + doi = {10.1021/acsnano.9b00184}, + url = {https://doi.org/10.1021/acsnano.9b00184}, + urldate = {2022-10-21}, + keywords = {physics,rec-by-ghosh,skeptics}, + file = {/home/johannes/Nextcloud/Zotero/Wang et al_2020_Will Any Crap We Put into Graphene Increase Its Electrocatalytic Effect.pdf;/home/johannes/Zotero/storage/5YHQ95UH/acsnano.html} +} + +@article{wardGeneralpurposeMachineLearning2016, + title = {A General-Purpose Machine Learning Framework for Predicting Properties of Inorganic Materials}, + author = {Ward, Logan and Agrawal, Ankit and Choudhary, Alok and Wolverton, Christopher}, + date = {2016-08-26}, + journaltitle = {npj Computational Materials}, + volume = {2}, + number = {1}, + pages = {1--7}, + publisher = {{Nature Publishing Group}}, + issn = {2057-3960}, + doi = {10.1038/npjcompumats.2016.28}, + url = {https://www.nature.com/articles/npjcompumats201628}, + urldate = {2021-05-13}, + abstract = {A very active area of materials research is to devise methods that use machine learning to automatically extract predictive models from existing materials data. While prior examples have demonstrated successful models for some applications, many more applications exist where machine learning can make a strong impact. To enable faster development of machine-learning-based models for such applications, we have created a framework capable of being applied to a broad range of materials data. Our method works by using a chemically diverse list of attributes, which we demonstrate are suitable for describing a wide variety of properties, and a novel method for partitioning the data set into groups of similar materials to boost the predictive accuracy. In this manuscript, we demonstrate how this new method can be used to predict diverse properties of crystalline and amorphous materials, such as band gap energy and glass-forming ability.}, + issue = {1}, + langid = {english}, + keywords = {compositional descriptors,descriptors,feature engineering,library,magpie,ML,models,rec-by-ruess}, + file = {/home/johannes/Nextcloud/Zotero/Ward et al_2016_A general-purpose machine learning framework for predicting properties of.pdf;/home/johannes/Zotero/storage/39BDW9ZZ/npjcompumats201628.html} +} + +@article{wardMatminerOpenSource2018, + title = {Matminer: {{An}} Open Source Toolkit for Materials Data Mining}, + shorttitle = {Matminer}, + author = {Ward, Logan and Dunn, Alexander and Faghaninia, Alireza and Zimmermann, Nils E. R. and Bajaj, Saurabh and Wang, Qi and Montoya, Joseph and Chen, Jiming and Bystrom, Kyle and Dylla, Maxwell and Chard, Kyle and Asta, Mark and Persson, Kristin A. and Snyder, G. Jeffrey and Foster, Ian and Jain, Anubhav}, + date = {2018-09-01}, + journaltitle = {Computational Materials Science}, + shortjournal = {Computational Materials Science}, + volume = {152}, + pages = {60--69}, + issn = {0927-0256}, + doi = {10.1016/j.commatsci.2018.05.018}, + url = {https://www.sciencedirect.com/science/article/pii/S0927025618303252}, + urldate = {2021-10-18}, + abstract = {As materials data sets grow in size and scope, the role of data mining and statistical learning methods to analyze these materials data sets and build predictive models is becoming more important. This manuscript introduces matminer, an open-source, Python-based software platform to facilitate data-driven methods of analyzing and predicting materials properties. Matminer provides modules for retrieving large data sets from external databases such as the Materials Project, Citrination, Materials Data Facility, and Materials Platform for Data Science. It also provides implementations for an extensive library of feature extraction routines developed by the materials community, with 47 featurization classes that can generate thousands of individual descriptors and combine them into mathematical functions. Finally, matminer provides a visualization module for producing interactive, shareable plots. These functions are designed in a way that integrates closely with machine learning and data analysis packages already developed and in use by the Python data science community. We explain the structure and logic of matminer, provide a description of its various modules, and showcase several examples of how matminer can be used to collect data, reproduce data mining studies reported in the literature, and test new methodologies.}, + langid = {english}, + keywords = {Data mining,Machine learning,Materials informatics,Open source software}, + file = {/home/johannes/Nextcloud/Zotero/Ward et al_2018_Matminer.pdf;/home/johannes/Zotero/storage/P7KMD3SZ/S0927025618303252.html} +} + +@article{waroquiersStatisticalAnalysisCoordination2017, + title = {Statistical {{Analysis}} of {{Coordination Environments}} in {{Oxides}}}, + author = {Waroquiers, David and Gonze, Xavier and Rignanese, Gian-Marco and Welker-Nieuwoudt, Cathrin and Rosowski, Frank and Göbel, Michael and Schenk, Stephan and Degelmann, Peter and André, Rute and Glaum, Robert and Hautier, Geoffroy}, + date = {2017-10-10}, + journaltitle = {Chemistry of Materials}, + shortjournal = {Chem. Mater.}, + volume = {29}, + number = {19}, + pages = {8346--8360}, + publisher = {{American Chemical Society}}, + issn = {0897-4756}, + doi = {10.1021/acs.chemmater.7b02766}, + url = {https://doi.org/10.1021/acs.chemmater.7b02766}, + urldate = {2021-07-20}, + abstract = {Coordination or local environments (e.g., tetrahedra and octahedra) are powerful descriptors of the crystalline structure of materials. These structural descriptors are essential to the understanding of crystal chemistry and the design of new materials. However, extensive statistics on the occurrence of local environment are not available even on common chemistries such as oxides. Here, we present the first large-scale statistical analysis of the coordination environments of cations in oxides using a large set of experimentally observed compounds (about 8000). Using a newly developed method, we provide the distribution of local environment for each cation in oxides. We discuss our results highlighting previously known trends and unexpected coordination environments, as well as compounds presenting very rare coordinations. Our work complements the know-how of the solid state chemist with a statistically sound analysis and paves the way for further data mining efforts linking, for instance, coordination environments to materials properties.}, + keywords = {ChemEnv,continuous symmetry measure,coordination environments,descriptors,library,pymatgen,rec-by-kovacik,voronoi analysis,voronoi tessellation,with-code}, + file = {/home/johannes/Nextcloud/Zotero/Waroquiers et al_2017_Statistical Analysis of Coordination Environments in Oxides.pdf} +} + +@software{wasmerAiidajutools2022, + title = {Aiida-Jutools}, + author = {Wasmer, Johannes}, + date = {2022-01-19T08:36:40Z}, + origdate = {2020-01-08T15:38:01Z}, + url = {https://github.com/JuDFTteam/aiida-jutools}, + urldate = {2022-10-16}, + abstract = {Tools for managing high-throughput experiments with AiiDA.}, + organization = {{JuDFTteam}} +} + +@unpublished{wasmerComparisonStructuralRepresentations2021, + type = {Poster}, + title = {Comparison of Structural Representations for Machine Learning-Accelerated Ab Initio Calculations}, + author = {Wasmer, Johannes and Rüßmann, Philipp and Blügel, Stefan}, + date = {2021}, + url = {https://juser.fz-juelich.de/record/901958}, + urldate = {2022-10-16}, + abstract = {Wasmer, Johannes; Rüßmann, Philipp; Blügel, Stefan}, + eventtitle = {{{DPG SKM21}}}, + langid = {english}, + venue = {{online}}, + file = {/home/johannes/Zotero/storage/RJIXGQPU/901958.html} +} + +@thesis{wasmerDevelopmentSurrogateMachine2021, + type = {mathesis}, + title = {Development of a Surrogate Machine Learning Model for the Acceleration of Density Functional Calculations with the {{Korringa-Kohn-Rostoker}} Method}, + author = {Wasmer, Johannes}, + date = {2021-10-27}, + institution = {{RWTH Aachen University}}, + url = {https://iffgit.fz-juelich.de/phd-project-wasmer/theses/master-thesis}, + urldate = {2022-08-08}, + langid = {english}, + pagetotal = {99}, + keywords = {_tablet,master-thesis,PGI-1/IAS-1,thesis}, + file = {/home/johannes/Nextcloud/Zotero/false;/home/johannes/Nextcloud/Zotero/Wasmer_2021_Development of a surrogate machine learning model for the acceleration of.pdf;/home/johannes/Zotero/storage/AC483X2N/master-thesis.html} +} + +@article{weinertSolutionPoissonEquation1981, + title = {Solution of {{Poisson}}’s Equation: {{Beyond Ewald}}â€type Methods}, + shorttitle = {Solution of {{Poisson}}’s Equation}, + author = {Weinert, M.}, + date = {1981-11}, + journaltitle = {Journal of Mathematical Physics}, + shortjournal = {J. Math. Phys.}, + volume = {22}, + number = {11}, + pages = {2433--2439}, + publisher = {{American Institute of Physics}}, + issn = {0022-2488}, + doi = {10.1063/1.524800}, + url = {https://aip.scitation.org/doi/10.1063/1.524800}, + urldate = {2022-10-01}, + keywords = {DFT,FLEUR,Poisson equation,potential}, + file = {/home/johannes/Nextcloud/Zotero/Weinert_1981_Solution of Poisson’s equation.pdf} +} + +@misc{wellawattePerspectiveExplanationsMolecular2022, + title = {A {{Perspective}} on {{Explanations}} of {{Molecular Prediction Models}}}, + author = {Wellawatte, Geemi P. and Gandhi, Heta A. and Seshadri, Aditi and White, Andrew D.}, + date = {2022-12-09}, + publisher = {{ChemRxiv}}, + doi = {10.26434/chemrxiv-2022-qfv02}, + url = {https://chemrxiv.org/engage/chemrxiv/article-details/639222a114d92d7cd6a65e90}, + urldate = {2022-12-29}, + abstract = {Chemists can be skeptical in using deep learning (DL) in decision making, due to the lack of interpretability in "black-box" models. Explainable artificial intelligence (XAI) is a branch of AI which addresses this drawback by providing tools to interpret DL models and their predictions. We review the principles of XAI in the domain of chemistry and emerging methods for creating and evaluating explanations. Then we focus methods developed by our group and their application to predicting solubility, blood-brain barrier permeability, and the scent of molecules. We show that XAI methods like chemical counterfactuals and descriptor explanations can both explain DL predictions and give insight into structure-property relationships. Finally, we discuss how a two step process of highly accurate black-box modeling and then creating explanations gives both highly accurate predictions and clear structure-property relationships.}, + langid = {english}, + keywords = {counterfactual explanation,Deep learning,GNN,molecules,XAI}, + file = {/home/johannes/Nextcloud/Zotero/Wellawatte et al_2022_A Perspective on Explanations of Molecular Prediction Models.pdf} +} + +@article{westermayrMachineLearningElectronically2021, + title = {Machine {{Learning}} for {{Electronically Excited States}} of {{Molecules}}}, + author = {Westermayr, Julia and Marquetand, Philipp}, + date = {2021-08-25}, + journaltitle = {Chemical Reviews}, + shortjournal = {Chem. Rev.}, + volume = {121}, + number = {16}, + pages = {9873--9926}, + publisher = {{American Chemical Society}}, + issn = {0009-2665}, + doi = {10.1021/acs.chemrev.0c00749}, + url = {https://doi.org/10.1021/acs.chemrev.0c00749}, + urldate = {2021-12-14}, + abstract = {Electronically excited states of molecules are at the heart of photochemistry, photophysics, as well as photobiology and also play a role in material science. Their theoretical description requires highly accurate quantum chemical calculations, which are computationally expensive. In this review, we focus on not only how machine learning is employed to speed up such excited-state simulations but also how this branch of artificial intelligence can be used to advance this exciting research field in all its aspects. Discussed applications of machine learning for excited states include excited-state dynamics simulations, static calculations of absorption spectra, as well as many others. In order to put these studies into context, we discuss the promises and pitfalls of the involved machine learning techniques. Since the latter are mostly based on quantum chemistry calculations, we also provide a short introduction into excited-state electronic structure methods and approaches for nonadiabatic dynamics simulations and describe tricks and problems when using them in machine learning for excited states of molecules.}, + keywords = {DFT,excited states,MD,ML,molecules,review,TDDFT}, + file = {/home/johannes/Nextcloud/Zotero/Westermayr_Marquetand_2021_Machine Learning for Electronically Excited States of Molecules.pdf} +} + +@unpublished{westermayrPerspectiveIntegratingMachine2021, + title = {Perspective on Integrating Machine Learning into Computational Chemistry and Materials Science}, + author = {Westermayr, Julia and Gastegger, Michael and Schütt, Kristof T. and Maurer, Reinhard J.}, + date = {2021-04-20}, + eprint = {2102.08435}, + eprinttype = {arxiv}, + primaryclass = {physics}, + url = {http://arxiv.org/abs/2102.08435}, + urldate = {2021-05-13}, + abstract = {Machine learning (ML) methods are being used in almost every conceivable area of electronic structure theory and molecular simulation. In particular, ML has become firmly established in the construction of high-dimensional interatomic potentials. Not a day goes by without another proof of principle being published on how ML methods can represent and predict quantum mechanical properties -- be they observable, such as molecular polarizabilities, or not, such as atomic charges. As ML is becoming pervasive in electronic structure theory and molecular simulation, we provide an overview of how atomistic computational modeling is being transformed by the incorporation of ML approaches. From the perspective of the practitioner in the field, we assess how common workflows to predict structure, dynamics, and spectroscopy are affected by ML. Finally, we discuss how a tighter and lasting integration of ML methods with computational chemistry and materials science can be achieved and what it will mean for research practice, software development, and postgraduate training.}, + archiveprefix = {arXiv}, + keywords = {ML,physics,review}, + file = {/home/johannes/Nextcloud/Zotero/Westermayr et al_2021_Perspective on integrating machine learning into computational chemistry and.pdf;/home/johannes/Nextcloud/Zotero/Westermayr et al_2021_Perspective on integrating machine learning into computational chemistry and2.pdf;/home/johannes/Zotero/storage/FHJLNQAU/2102.html} +} + +@article{whiteDeepLearningMolecules2021, + title = {Deep {{Learning}} for {{Molecules}} and {{Materials}}}, + author = {White, Andrew D.}, + date = {2021}, + journaltitle = {Living Journal of Computational Molecular Science}, + volume = {3}, + number = {1}, + pages = {1499--1499}, + issn = {2575-6524}, + doi = {10.33011/livecoms.3.1.1499}, + url = {https://livecomsjournal.org/index.php/livecoms/article/view/v3i1e1499}, + urldate = {2022-07-10}, + abstract = {Deep learning is becoming a standard tool in chemistry and materials science. Although there are learning materials available for deep learning, none cover the applications in chemistry and materials science or the peculiarities of working with molecules. The textbook described here provides a systematic and applied introduction to the latest research in deep learning in chemistry and materials science. It covers the math fundamentals, the requisite machine learning, the common neural network architectures used today, and the details necessary to be a practitioner of deep learning. The textbook is a living document and will be updated as the rapidly changing deep learning field evolves.}, + issue = {1}, + langid = {english}, + keywords = {_tablet,book,GNN,ML-DFT,ML-ESM,MLP,MPNN,tutorial}, + file = {/home/johannes/Nextcloud/Zotero/White_2021_Deep Learning for Molecules and Materials.pdf} +} + +@article{wiederCompactReviewMolecular2020, + title = {A Compact Review of Molecular Property Prediction with Graph Neural Networks}, + author = {Wieder, Oliver and Kohlbacher, Stefan and Kuenemann, Mélaine and Garon, Arthur and Ducrot, Pierre and Seidel, Thomas and Langer, Thierry}, + date = {2020-12-01}, + journaltitle = {Drug Discovery Today: Technologies}, + shortjournal = {Drug Discovery Today: Technologies}, + volume = {37}, + pages = {1--12}, + issn = {1740-6749}, + doi = {10.1016/j.ddtec.2020.11.009}, + url = {https://www.sciencedirect.com/science/article/pii/S1740674920300305}, + urldate = {2022-09-27}, + abstract = {As graph neural networks are becoming more and more powerful and useful in the field of drug discovery, many pharmaceutical companies are getting interested in utilizing these methods for their own in-house frameworks. This is especially compelling for tasks such as the prediction of molecular properties which is often one of the most crucial tasks in computer-aided drug discovery workflows. The immense hype surrounding these kinds of algorithms has led to the development of many different types of promising architectures and in this review we try to structure this highly dynamic field of AI-research by collecting and classifying 80 GNNs that have been used to predict more than 20 molecular properties using 48 different datasets.}, + langid = {english}, + keywords = {_tablet,chemistry,GCN,GNN,molecules,review}, + file = {/home/johannes/Nextcloud/Zotero/Wieder et al_2020_A compact review of molecular property prediction with graph neural networks.pdf;/home/johannes/Zotero/storage/KHCYV2ZB/S1740674920300305.html} +} + +@article{wilkinsAccurateMolecularPolarizabilities2019, + title = {Accurate Molecular Polarizabilities with Coupled Cluster Theory and Machine Learning}, + author = {Wilkins, David M. and Grisafi, Andrea and Yang, Yang and Lao, Ka Un and DiStasio, Robert A. and Ceriotti, Michele}, + date = {2019-02-26}, + journaltitle = {Proceedings of the National Academy of Sciences}, + volume = {116}, + number = {9}, + pages = {3401--3406}, + publisher = {{Proceedings of the National Academy of Sciences}}, + doi = {10.1073/pnas.1816132116}, + url = {https://www.pnas.org/doi/10.1073/pnas.1816132116}, + urldate = {2022-08-16}, + keywords = {ML-ESM,SA-GPR}, + file = {/home/johannes/Nextcloud/Zotero/Wilkins et al_2019_Accurate molecular polarizabilities with coupled cluster theory and machine.pdf} +} + +@article{wilkinsonFAIRGuidingPrinciples2016, + title = {The {{FAIR Guiding Principles}} for Scientific Data Management and Stewardship}, + author = {Wilkinson, Mark D. and Dumontier, Michel and Aalbersberg, IJsbrand Jan and Appleton, Gabrielle and Axton, Myles and Baak, Arie and Blomberg, Niklas and Boiten, Jan-Willem and da Silva Santos, Luiz Bonino and Bourne, Philip E. and Bouwman, Jildau and Brookes, Anthony J. and Clark, Tim and Crosas, Mercè and Dillo, Ingrid and Dumon, Olivier and Edmunds, Scott and Evelo, Chris T. and Finkers, Richard and Gonzalez-Beltran, Alejandra and Gray, Alasdair J. G. and Groth, Paul and Goble, Carole and Grethe, Jeffrey S. and Heringa, Jaap and ’t Hoen, Peter A. C. and Hooft, Rob and Kuhn, Tobias and Kok, Ruben and Kok, Joost and Lusher, Scott J. and Martone, Maryann E. and Mons, Albert and Packer, Abel L. and Persson, Bengt and Rocca-Serra, Philippe and Roos, Marco and van Schaik, Rene and Sansone, Susanna-Assunta and Schultes, Erik and Sengstag, Thierry and Slater, Ted and Strawn, George and Swertz, Morris A. and Thompson, Mark and van der Lei, Johan and van Mulligen, Erik and Velterop, Jan and Waagmeester, Andra and Wittenburg, Peter and Wolstencroft, Katherine and Zhao, Jun and Mons, Barend}, + options = {useprefix=true}, + date = {2016-03-15}, + journaltitle = {Scientific Data}, + shortjournal = {Sci Data}, + volume = {3}, + number = {1}, + pages = {160018}, + publisher = {{Nature Publishing Group}}, + issn = {2052-4463}, + doi = {10.1038/sdata.2016.18}, + url = {https://www.nature.com/articles/sdata201618}, + urldate = {2021-10-15}, + abstract = {There is an urgent need to improve the infrastructure supporting the reuse of scholarly data. A diverse set of stakeholders—representing academia, industry, funding agencies, and scholarly publishers—have come together to design and jointly endorse a concise and measureable set of principles that we refer to as the FAIR Data Principles. The intent is that these may act as a guideline for those wishing to enhance the reusability of their data holdings. Distinct from peer initiatives that focus on the human scholar, the FAIR Principles put specific emphasis on enhancing the ability of machines to automatically find and use the data, in addition to supporting its reuse by individuals. This Comment is the first formal publication of the FAIR Principles, and includes the rationale behind them, and some exemplar implementations in the community.}, + issue = {1}, + langid = {english}, + keywords = {FAIR,original publication}, + annotation = {Bandiera\_abtest: a Cg\_type: Nature Research Journals Primary\_atype: Comments \& Opinion Subject\_term: Publication characteristics;Research data Subject\_term\_id: publication-characteristics;research-data}, + file = {/home/johannes/Nextcloud/Zotero/Wilkinson et al_2016_The FAIR Guiding Principles for scientific data management and stewardship.pdf;/home/johannes/Zotero/storage/7QCVD3LB/sdata201618.html} +} + +@article{willattAtomdensityRepresentationsMachine2019, + title = {Atom-Density Representations for Machine Learning}, + author = {Willatt, Michael J. and Musil, Félix and Ceriotti, Michele}, + date = {2019-04-17}, + journaltitle = {The Journal of Chemical Physics}, + shortjournal = {J. Chem. Phys.}, + volume = {150}, + number = {15}, + pages = {154110}, + publisher = {{American Institute of Physics}}, + issn = {0021-9606}, + doi = {10.1063/1.5090481}, + url = {https://aip.scitation.org/doi/10.1063/1.5090481}, + urldate = {2021-05-30}, + abstract = {The applications of machine learning techniques to chemistry and materials science become more numerous by the day. The main challenge is to devise representations of atomic systems that are at the same time complete and concise, so as to reduce the number of reference calculations that are needed to predict the properties of different types of materials reliably. This has led to a proliferation of alternative ways to convert an atomic structure into an input for a machine-learning model. We introduce an abstract definition of chemical environments that is based on a smoothed atomic density, using a bra-ket notation to emphasize basis set independence and to highlight the connections with some popular choices of representations for describing atomic systems. The correlations between the spatial distribution of atoms and their chemical identities are computed as inner products between these feature kets, which can be given an explicit representation in terms of the expansion of the atom density on orthogonal basis functions, that is equivalent to the smooth overlap of atomic positions power spectrum, but also in real space, corresponding to n-body correlations of the atom density. This formalism lays the foundations for a more systematic tuning of the behavior of the representations, by introducing operators that represent the correlations between structure, composition, and the target properties. It provides a unifying picture of recent developments in the field and indicates a way forward toward more effective and computationally affordable machine-learning schemes for molecules and materials.}, + keywords = {ACSF,alchemical,chemical species scaling problem,descriptor dimred,descriptors,descriptors analysis,dimensionality reduction,MBTR,ML,SOAP}, + file = {/home/johannes/Nextcloud/Zotero/Willatt et al_2019_Atom-density representations for machine learning.pdf} +} + +@article{willattFeatureOptimizationAtomistic2018, + title = {Feature Optimization for Atomistic Machine Learning Yields a Data-Driven Construction of the Periodic Table of the Elements}, + author = {Willatt, Michael J. and Musil, Félix and Ceriotti, Michele}, + date = {2018-12-05}, + journaltitle = {Physical Chemistry Chemical Physics}, + shortjournal = {Phys. Chem. Chem. Phys.}, + volume = {20}, + number = {47}, + pages = {29661--29668}, + publisher = {{The Royal Society of Chemistry}}, + issn = {1463-9084}, + doi = {(}, + url = {https://pubs.rsc.org/en/content/articlelanding/2018/cp/c8cp05921g}, + urldate = {2021-05-13}, + abstract = {Machine-learning of atomic-scale properties amounts to extracting correlations between structure, composition and the quantity that one wants to predict. Representing the input structure in a way that best reflects such correlations makes it possible to improve the accuracy of the model for a given amount of reference data. When using a description of the structures that is transparent and well-principled, optimizing the representation might reveal insights into the chemistry of the data set. Here we show how one can generalize the SOAP kernel to introduce a distance-dependent weight that accounts for the multi-scale nature of the interactions, and a description of correlations between chemical species. We show that this improves substantially the performance of ML models of molecular and materials stability, while making it easier to work with complex, multi-component systems and to extend SOAP to coarse-grained intermolecular potentials. The element correlations that give the best performing model show striking similarities with the conventional periodic table of the elements, providing an inspiring example of how machine learning can rediscover, and generalize, intuitive concepts that constitute the foundations of chemistry.}, + langid = {english}, + keywords = {alchemical,chemical species scaling problem,descriptor dimred,descriptors,dimensionality reduction,ML,SOAP}, + file = {/home/johannes/Nextcloud/Zotero/Willatt et al_2018_Feature optimization for atomistic machine learning yields a data-driven.pdf;/home/johannes/Zotero/storage/ZY2VC9JE/C8CP05921G.html} +} + +@unpublished{winterUnsupervisedLearningGroup2022, + title = {Unsupervised {{Learning}} of {{Group Invariant}} and {{Equivariant Representations}}}, + author = {Winter, Robin and Bertolini, Marco and Le, Tuan and Noé, Frank and Clevert, Djork-Arné}, + date = {2022-02-15}, + eprint = {2202.07559}, + eprinttype = {arxiv}, + primaryclass = {cs}, + url = {http://arxiv.org/abs/2202.07559}, + urldate = {2022-05-11}, + abstract = {Equivariant neural networks, whose hidden features transform according to representations of a group G acting on the data, exhibit training efficiency and an improved generalisation performance. In this work, we extend group invariant and equivariant representation learning to the field of unsupervised deep learning. We propose a general learning strategy based on an encoder-decoder framework in which the latent representation is disentangled in an invariant term and an equivariant group action component. The key idea is that the network learns the group action on the data space and thus is able to solve the reconstruction task from an invariant data representation, hence avoiding the necessity of ad-hoc group-specific implementations. We derive the necessary conditions on the equivariant encoder, and we present a construction valid for any G, both discrete and continuous. We describe explicitly our construction for rotations, translations and permutations. We test the validity and the robustness of our approach in a variety of experiments with diverse data types employing different network architectures.}, + archiveprefix = {arXiv}, + keywords = {Computer Science - Machine Learning}, + file = {/home/johannes/Nextcloud/Zotero/Winter et al_2022_Unsupervised Learning of Group Invariant and Equivariant Representations.pdf;/home/johannes/Zotero/storage/5PYE8XM2/2202.html} +} + +@online{woodQuantumComplexityTamed2022, + title = {Quantum {{Complexity Tamed}} by {{Machine Learning}}}, + author = {Wood, Charlie}, + date = {2022-02-07T15:54+00:00}, + url = {https://www.quantamagazine.org/quantum-complexity-tamed-by-machine-learning-20220207/}, + urldate = {2022-10-05}, + abstract = {If only scientists understood exactly how electrons act in molecules, they’d be able to predict the behavior of everything from experimental drugs to high-temperature superconductors.}, + langid = {english}, + organization = {{Quanta Magazine}}, + keywords = {DeepMind,DFT,DM21,for introductions,ML-DFA,ML-DFT,ML-ESM,molecules,popular science}, + file = {/home/johannes/Zotero/storage/Y7SURWT5/quantum-complexity-tamed-by-machine-learning-20220207.html} +} + +@article{wurgerExploringStructurepropertyRelationships2021, + title = {Exploring Structure-Property Relationships in Magnesium Dissolution Modulators}, + author = {Würger, Tim and Mei, Di and Vaghefinazari, Bahram and Winkler, David A. and Lamaka, Sviatlana V. and Zheludkevich, Mikhail L. and Meißner, Robert H. and Feiler, Christian}, + date = {2021-01-08}, + journaltitle = {npj Materials Degradation}, + shortjournal = {npj Mater Degrad}, + volume = {5}, + number = {1}, + pages = {1--10}, + publisher = {{Nature Publishing Group}}, + issn = {2397-2106}, + doi = {10.1038/s41529-020-00148-z}, + url = {https://www.nature.com/articles/s41529-020-00148-z}, + urldate = {2021-07-20}, + abstract = {Small organic molecules that modulate the degradation behavior of Mg constitute benign and useful materials to modify the service environment of light metal materials for specific applications. The vast chemical space of potentially effective compounds can be explored by machine learning-based quantitative structure-property relationship models, accelerating the discovery of potent dissolution modulators. Here, we demonstrate how unsupervised clustering of a large number of potential Mg dissolution modulators by structural similarities and sketch-maps can predict their experimental performance using a kernel ridge regression model. We compare the prediction accuracy of this approach to that of a prior artificial neural networks study. We confirm the robustness of our data-driven model by blind prediction of the dissolution modulating performance of 10 untested compounds. Finally, a workflow is presented that facilitates the automated discovery of chemicals with desired dissolution modulating properties from a commercial database. We subsequently prove this concept by blind validation of five chemicals.}, + issue = {1}, + langid = {english}, + keywords = {descriptor comparison,descriptors,dimensionality reduction,KRR,materials discovery,ML,models,sketchmap,SOAP}, + annotation = {Bandiera\_abtest: a Cc\_license\_type: cc\_by Cg\_type: Nature Research Journals Primary\_atype: Research Subject\_term: Computational methods;Corrosion;Mathematics and computing;Theoretical chemistry Subject\_term\_id: computational-methods;corrosion;mathematics-and-computing;theoretical-chemistry}, + file = {/home/johannes/Nextcloud/Zotero/Würger et al_2021_Exploring structure-property relationships in magnesium dissolution modulators.pdf;/home/johannes/Zotero/storage/NM6RVQRY/s41529-020-00148-z.html} +} + +@article{xieCrystalGraphConvolutional2018, + title = {Crystal {{Graph Convolutional Neural Networks}} for an {{Accurate}} and {{Interpretable Prediction}} of {{Material Properties}}}, + author = {Xie, Tian and Grossman, Jeffrey C.}, + date = {2018-04-06}, + journaltitle = {Physical Review Letters}, + shortjournal = {Phys. Rev. Lett.}, + volume = {120}, + number = {14}, + pages = {145301}, + publisher = {{American Physical Society}}, + doi = {10.1103/PhysRevLett.120.145301}, + url = {https://link.aps.org/doi/10.1103/PhysRevLett.120.145301}, + urldate = {2022-09-27}, + abstract = {The use of machine learning methods for accelerating the design of crystalline materials usually requires manually constructed feature vectors or complex transformation of atom coordinates to input the crystal structure, which either constrains the model to certain crystal types or makes it difficult to provide chemical insights. Here, we develop a crystal graph convolutional neural networks framework to directly learn material properties from the connection of atoms in the crystal, providing a universal and interpretable representation of crystalline materials. Our method provides a highly accurate prediction of density functional theory calculated properties for eight different properties of crystals with various structure types and compositions after being trained with 104 data points. Further, our framework is interpretable because one can extract the contributions from local chemical environments to global properties. Using an example of perovskites, we show how this information can be utilized to discover empirical rules for materials design.}, + keywords = {_tablet,CGCNN,GCN,GNN,library,solids,with-code}, + file = {/home/johannes/Nextcloud/Zotero/Xie_Grossman_2018_Crystal Graph Convolutional Neural Networks for an Accurate and Interpretable.pdf} +} + +@unpublished{xieUltrafastInterpretableMachinelearning2021, + title = {Ultra-Fast Interpretable Machine-Learning Potentials}, + author = {Xie, Stephen R. and Rupp, Matthias and Hennig, Richard G.}, + date = {2021-10-01}, + eprint = {2110.00624}, + eprinttype = {arxiv}, + primaryclass = {cond-mat, physics:physics}, + url = {http://arxiv.org/abs/2110.00624}, + urldate = {2022-05-09}, + abstract = {All-atom dynamics simulations are an indispensable quantitative tool in physics, chemistry, and materials science, but large systems and long simulation times remain challenging due to the trade-off between computational efficiency and predictive accuracy. To address this challenge, we combine effective two- and three-body potentials in a cubic B-spline basis with regularized linear regression to obtain machine-learning potentials that are physically interpretable, sufficiently accurate for applications, as fast as the fastest traditional empirical potentials, and two to four orders of magnitude faster than state-of-the-art machine-learning potentials. For data from empirical potentials, we demonstrate exact retrieval of the potential. For data from density functional theory, the predicted energies, forces, and derived properties, including phonon spectra, elastic constants, and melting points, closely match those of the reference method. The introduced potentials might contribute towards accurate all-atom dynamics simulations of large atomistic systems over long time scales.}, + archiveprefix = {arXiv}, + keywords = {descriptors,UFP}, + file = {/home/johannes/Nextcloud/Zotero/Xie et al_2021_Ultra-fast interpretable machine-learning potentials.pdf;/home/johannes/Zotero/storage/8585X9YA/2110.html} +} + +@article{xuSurveyMultiOutputLearning2020, + title = {Survey on {{Multi-Output Learning}}}, + author = {Xu, Donna and Shi, Yaxin and Tsang, Ivor W. and Ong, Yew-Soon and Gong, Chen and Shen, Xiaobo}, + date = {2020-07}, + journaltitle = {IEEE Transactions on Neural Networks and Learning Systems}, + volume = {31}, + number = {7}, + pages = {2409--2429}, + issn = {2162-2388}, + doi = {10.1109/TNNLS.2019.2945133}, + abstract = {The aim of multi-output learning is to simultaneously predict multiple outputs given an input. It is an important learning problem for decision-making since making decisions in the real world often involves multiple complex factors and criteria. In recent times, an increasing number of research studies have focused on ways to predict multiple outputs at once. Such efforts have transpired in different forms according to the particular multi-output learning problem under study. Classic cases of multi-output learning include multi-label learning, multi-dimensional learning, multi-target regression, and others. From our survey of the topic, we were struck by a lack in studies that generalize the different forms of multi-output learning into a common framework. This article fills that gap with a comprehensive review and analysis of the multi-output learning paradigm. In particular, we characterize the four Vs of multi-output learning, i.e., volume, velocity, variety, and veracity, and the ways in which the four Vs both benefit and bring challenges to multi-output learning by taking inspiration from big data. We analyze the life cycle of output labeling, present the main mathematical definitions of multi-output learning, and examine the field's key challenges and corresponding solutions as found in the literature. Several model evaluation metrics and popular data repositories are also discussed. Last but not least, we highlight some emerging challenges with multi-output learning from the perspective of the four Vs as potential research directions worthy of further studies.}, + eventtitle = {{{IEEE Transactions}} on {{Neural Networks}} and {{Learning Systems}}}, + keywords = {ML,multi-output learning,multi-target learning,output label representation,structured output prediction,Supervised learning}, + file = {/home/johannes/Nextcloud/Zotero/Xu et al_2020_Survey on Multi-Output Learning.pdf;/home/johannes/Zotero/storage/9TWMMATA/8892612.html} +} + +@article{yamadaPredictingMaterialsProperties2019, + title = {Predicting {{Materials Properties}} with {{Little Data Using Shotgun Transfer Learning}}}, + author = {Yamada, Hironao and Liu, Chang and Wu, Stephen and Koyama, Yukinori and Ju, Shenghong and Shiomi, Junichiro and Morikawa, Junko and Yoshida, Ryo}, + date = {2019-10-23}, + journaltitle = {ACS Central Science}, + shortjournal = {ACS Cent. Sci.}, + volume = {5}, + number = {10}, + pages = {1717--1730}, + publisher = {{American Chemical Society}}, + issn = {2374-7943}, + doi = {10.1021/acscentsci.9b00804}, + url = {https://doi.org/10.1021/acscentsci.9b00804}, + urldate = {2021-05-15}, + abstract = {There is a growing demand for the use of machine learning (ML) to derive fast-to-evaluate surrogate models of materials properties. In recent years, a broad array of materials property databases have emerged as part of a digital transformation of materials science. However, recent technological advances in ML are not fully exploited because of the insufficient volume and diversity of materials data. An ML framework called “transfer learning†has considerable potential to overcome the problem of limited amounts of materials data. Transfer learning relies on the concept that various property types, such as physical, chemical, electronic, thermodynamic, and mechanical properties, are physically interrelated. For a given target property to be predicted from a limited supply of training data, models of related proxy properties are pretrained using sufficient data; these models capture common features relevant to the target task. Repurposing of such machine-acquired features on the target task yields outstanding prediction performance even with exceedingly small data sets, as if highly experienced human experts can make rational inferences even for considerably less experienced tasks. In this study, to facilitate widespread use of transfer learning, we develop a pretrained model library called XenonPy.MDL. In this first release, the library comprises more than 140\,000 pretrained models for various properties of small molecules, polymers, and inorganic crystalline materials. Along with these pretrained models, we describe some outstanding successes of transfer learning in different scenarios such as building models with only dozens of materials data, increasing the ability of extrapolative prediction through a strategic model transfer, and so on. Remarkably, transfer learning has autonomously identified rather nontrivial transferability across different properties transcending the different disciplines of materials science; for example, our analysis has revealed underlying bridges between small molecules and polymers and between organic and inorganic chemistry.}, + keywords = {compositional descriptors,database-based descriptors,descriptors,library,ML,models,notebooks,OFM descriptor,pretrained models,python,pytorch,RDF descriptor,small data,transfer learning,visualization,with-code,XenonPy}, + file = {/home/johannes/Nextcloud/Zotero/Yamada et al_2019_Predicting Materials Properties with Little Data Using Shotgun Transfer Learning.pdf;/home/johannes/Zotero/storage/4F8PQPMD/acscentsci.html} +} + +@article{yangMachinelearningAcceleratedGeometry2021, + title = {Machine-Learning Accelerated Geometry Optimization in Molecular Simulation}, + author = {Yang, Yilin and Jiménez-Negrón, Omar A. and Kitchin, John R.}, + date = {2021-06-21}, + journaltitle = {The Journal of Chemical Physics}, + shortjournal = {J. Chem. Phys.}, + volume = {154}, + number = {23}, + pages = {234704}, + publisher = {{American Institute of Physics}}, + issn = {0021-9606}, + doi = {10.1063/5.0049665}, + url = {https://aip.scitation.org/doi/10.1063/5.0049665}, + urldate = {2021-06-24}, + abstract = {Geometry optimization is an important part of both computational materials and surface science because it is the path to finding ground state atomic structures and reaction pathways. These properties are used in the estimation of thermodynamic and kinetic properties of molecular and crystal structures. This process is slow at the quantum level of theory because it involves an iterative calculation of forces using quantum chemical codes such as density functional theory (DFT), which are computationally expensive and which limit the speed of the optimization algorithms. It would be highly advantageous to accelerate this process because then one could do either the same amount of work in less time or more work in the same time. In this work, we provide a neural network (NN) ensemble based active learning method to accelerate the local geometry optimization for multiple configurations simultaneously. We illustrate the acceleration on several case studies including bare metal surfaces, surfaces with adsorbates, and nudged elastic band for two reactions. In all cases, the accelerated method requires fewer DFT calculations than the standard method. In addition, we provide an Atomic Simulation Environment (ASE)-optimizer Python package to make the usage of the NN ensemble active learning for geometry optimization easier.}, + keywords = {ACSF,active learning,BPNN,DFT,GPR,ML,SingleNN,structure relaxation,surrogate model,to_read}, + file = {/home/johannes/Nextcloud/Zotero/false;/home/johannes/Zotero/storage/2L5JFJN8/5.html} +} + +@online{zachglickDoesItFeel2021, + type = {Tweet}, + title = {Does It Feel like Everyone in \#compchem Is Doing Machine Learning Now? {{I}} Thought so after \#{{ACSSpring2021}}, and Decided to Look at How Frequently Certain Phrases Appeared in {{COMP}} Division Abstracts over the Years: {{https://t.co/F4awnzVebs}}}, + author = {{Zach Glick}}, + date = {2021-04-21T13:23Z}, + url = {https://twitter.com/ZachLGlick/status/1384860348730298375}, + urldate = {2021-05-13}, + langid = {english}, + organization = {{@ZachLGlick}}, + keywords = {ML}, + file = {/home/johannes/Zotero/storage/WCHKALVA/1384860348730298375.html} +} + +@article{zahariaAcceleratingMachineLearning2018, + title = {Accelerating the {{Machine Learning Lifecycle}} with {{MLflow}}}, + author = {Zaharia, M. and Chen, Andrew and Davidson, A. and Ghodsi, A. and Hong, S. and Konwinski, A. and Murching, Siddharth and Nykodym, Tomas and Ogilvie, Paul and Parkhe, Mani and Xie, Fen and Zumar, Corey}, + date = {2018}, + journaltitle = {IEEE Data Eng. Bull.}, + abstract = {MLflow, an open source platform recently launched to streamline the machine learning lifecycle, covers three key challenges: experimentation, reproducibility, and model deployment, using generic APIs that work with any ML library, algorithm and programming language. Machine learning development creates multiple new challenges that are not present in a traditional software development lifecycle. These include keeping track of the myriad inputs to an ML application (e.g., data versions, code and tuning parameters), reproducing results, and production deployment. In this paper, we summarize these challenges from our experience with Databricks customers, and describe MLflow, an open source platform we recently launched to streamline the machine learning lifecycle. MLflow covers three key challenges: experimentation, reproducibility, and model deployment, using generic APIs that work with any ML library, algorithm and programming language. The project has a rapidly growing open source community, with over 50 contributors since its launch in June 2018.}, + file = {/home/johannes/Nextcloud/Zotero/Zaharia et al_2018_Accelerating the Machine Learning Lifecycle with MLflow.pdf} +} + +@article{zaverkinFastSampleEfficientInteratomic2021, + title = {Fast and {{Sample-Efficient Interatomic Neural Network Potentials}} for {{Molecules}} and {{Materials Based}} on {{Gaussian Moments}}}, + author = {Zaverkin, Viktor and Holzmüller, David and Steinwart, Ingo and Kästner, Johannes}, + date = {2021-10-12}, + journaltitle = {Journal of Chemical Theory and Computation}, + shortjournal = {J. Chem. Theory Comput.}, + volume = {17}, + number = {10}, + pages = {6658--6670}, + publisher = {{American Chemical Society}}, + issn = {1549-9618}, + doi = {10.1021/acs.jctc.1c00527}, + url = {https://doi.org/10.1021/acs.jctc.1c00527}, + urldate = {2022-01-02}, + abstract = {Artificial neural networks (NNs) are one of the most frequently used machine learning approaches to construct interatomic potentials and enable efficient large-scale atomistic simulations with almost ab initio accuracy. However, the simultaneous training of NNs on energies and forces, which are a prerequisite for, e.g., molecular dynamics simulations, can be demanding. In this work, we present an improved NN architecture based on the previous GM-NN model [Zaverkin V.; Kästner, J. J. Chem. Theory Comput. 2020, 16, 5410−5421], which shows an improved prediction accuracy and considerably reduced training times. Moreover, we extend the applicability of Gaussian moment-based interatomic potentials to periodic systems and demonstrate the overall excellent transferability and robustness of the respective models. The fast training by the improved methodology is a prerequisite for training-heavy workflows such as active learning or learning-on-the-fly.}, + file = {/home/johannes/Nextcloud/Zotero/Zaverkin et al_2021_Fast and Sample-Efficient Interatomic Neural Network Potentials for Molecules.pdf} +} + +@article{zeledonStructuralInformationFiltered2020, + title = {The Structural Information Filtered Features ({{SIFF}}) Potential: {{Maximizing}} Information Stored in Machine-Learning Descriptors for Materials Prediction}, + shorttitle = {The Structural Information Filtered Features ({{SIFF}}) Potential}, + author = {Zeledon, Jorge Arturo Hernandez and Romero, Aldo H. and Ren, Pengju and Wen, Xiaodong and Li, Yongwang and Lewis, James P.}, + date = {2020-06-03}, + journaltitle = {Journal of Applied Physics}, + shortjournal = {Journal of Applied Physics}, + volume = {127}, + number = {21}, + pages = {215108}, + publisher = {{American Institute of Physics}}, + issn = {0021-8979}, + doi = {10.1063/5.0002252}, + url = {https://aip.scitation.org/doi/10.1063/5.0002252}, + urldate = {2021-05-15}, + abstract = {Machine learning inspired potentials continue to improve the ability for predicting structures of materials. However, many challenges still exist, particularly when calculating structures of disordered systems. These challenges are primarily due to the rapidly increasing dimensionality of the feature-vector space which in most machine-learning algorithms is dependent on the size of the structure. In this article, we present a feature-engineered approach that establishes a set of principles for representing potentials of physical structures (crystals, molecules, and clusters) in a feature space rather than a physically motivated space. Our goal in this work is to define guiding principles that optimize information storage of the physical parameters within the feature representations. In this manner, we focus on keeping the dimensionality of the feature space independent of the number of atoms in the structure. Our Structural Information Filtered Features (SIFF) potential represents structures by utilizing a feature vector of low-correlated descriptors, which correspondingly maximizes information within the descriptor. We present results of our SIFF potential on datasets composed of disordered (carbon and carbon–oxygen) clusters, molecules with C7O2H2 stoichiometry in the GDB9-14B dataset, and crystal structures of the form (AlxGayInz)2O3 as proposed in the NOMAD Kaggle competition. Our potential's performance is at least comparable, sometimes significantly more accurate, and often more efficient than other well-known machine-learning potentials for structure prediction. However, primarily, we offer a different perspective on how researchers should consider opportunities in maximizing information storage for features.}, + keywords = {Behler,Behler-Parrinello potential,descriptors,feature engineering,ML,MLP,SIFF}, + file = {/home/johannes/Nextcloud/Zotero/Zeledon et al_2020_The structural information filtered features (SIFF) potential.pdf;/home/johannes/Zotero/storage/DQJJR84B/5.html} +} + +@report{zellerCorrelatedElectronsModels2012, + title = {Correlated Electrons: From Models to Materials}, + author = {Zeller, Rudolf}, + date = {2012}, + series = {Lecture {{Notes}} of the {{Autumn School}} on {{Correlated Electrons}}}, + number = {PreJuSER-136393}, + institution = {{Forschungszentrum Jülich GmbH Zenralbibliothek, Verlag}}, + url = {https://juser.fz-juelich.de/record/136393/}, + urldate = {2022-06-28}, + abstract = {Density-functional theory (DFT) is considered the Standard Model of solid-state physics. The state-of-the-art approximations to DFT, the local-density approximation (LDA) or its simple extensions, fail, however, even qualitatively, for strongly-correlated systems. When correlations are strong, electrons become entangled and novel properties emerge. Mott-transitions, Kondo- and heavy-fermion behavior, non-conventional superconductivity and orbital-order are just some examples of this emergent behavior. The realistic description of emergent properties is one of the grand-challenges of modern condensed-matter physics. To understand this physics beyond the Standard Model, nonperturbative many-body techniques are essential. Still, DFT-based methods are needed to devise materials-specific Hamiltonians for strong correlations. Mastering these novel techniques requires a vast background, ranging from DFT to model building and many-body physics. The aim of this school is to introduce advanced graduate students and up to the modern methods for modeling emergent properties of correlated electrons and to explore the relation of electron correlations with quantum entanglement and concepts from quantum information. A school of this size and scope requires support and help from many sources. We are very grateful for all the financial and practical support we have received. The Institute for Advanced Simulation and the German Research School for Simulation Sciences at the Forschungszentrum Jülich provided the funding and were vital for the organization of the school and the production of this book. The DFG Forschergruppe FOR1346 offered travel grants for students and the Institute for Complex Adaptive Matter (ICAM) travel support for international speakers and participants. The nature of a school makes it desirable to have the lecture-notes available already during the lectures. In this way the participants get the chance to work through the lectures thoroughly while they are given. We are therefore extremely grateful to the lecturers that, despite a tight schedule, provided their manuscripts in time for the production of this book. We are confident that the lecture notes collected here will not only serve the participants of the school but will also be useful for other students entering the exciting field of strongly correlated materials. We thank Mrs. H. Lexis of the Forschungszentrum Jülich Verlag and Mr. D. Laufenberg of the Graphische Betriebe for providing their expert support in producing the present volume on a tight schedule and for making even seemingly impossible requests possible. We heartily thank our students and postdocs that helped in proofreading the manuscripts, often on short notice: Carmine Autieri, Fabio Baruffa, Michael Baumgärtel, Monica Bugeanu, Andreas Flesch, Evgeny Gorelov, Amin Kiani Sheikhabadi, Joaquin Miranda, German Ulm, and Guoren Zhang. Finally, our special thanks go to Dipl.-Ing. R. Hölzle for his invaluable advice on all questions concerning the organization of such a school and to Mrs. L. Snyders and Mrs. E. George for expertly handling all practical issues. Pavarini, Eva; Koch, Erik; Anders, Frithjof; Jarrell, Mark (Eds. )}, + editorb = {Pavarini, Eva and Anders, Frithjof and Koch, Erik and Jarrell, Mark}, + editorbtype = {redactor}, + isbn = {9783893367962}, + langid = {english}, + keywords = {_tablet,KKR,PGI-1/IAS-1}, + file = {/home/johannes/Nextcloud/Zotero/Zeller_2012_Correlated electrons.pdf;/home/johannes/Zotero/storage/BKBRXSWN/136393.html} +} + +@article{zeniCompactAtomicDescriptors2021, + title = {Compact Atomic Descriptors Enable Accurate Predictions via Linear Models}, + author = {Zeni, Claudio and Rossi, Kevin and Glielmo, Aldo and de Gironcoli, Stefano}, + options = {useprefix=true}, + date = {2021-06-14}, + journaltitle = {The Journal of Chemical Physics}, + shortjournal = {J. Chem. Phys.}, + volume = {154}, + number = {22}, + pages = {224112}, + publisher = {{American Institute of Physics}}, + issn = {0021-9606}, + doi = {10.1063/5.0052961}, + url = {https://aip.scitation.org/doi/10.1063/5.0052961}, + urldate = {2022-05-11}, + abstract = {We probe the accuracy of linear ridge regression employing a three-body local density representation derived from the atomic cluster expansion. We benchmark the accuracy of this framework in the prediction of formation energies and atomic forces in molecules and solids. We find that such a simple regression framework performs on par with state-of-the-art machine learning methods which are, in most cases, more complex and more computationally demanding. Subsequently, we look for ways to sparsify the descriptor and further improve the computational efficiency of the method. To this aim, we use both principal component analysis and least absolute shrinkage operator regression for energy fitting on six single-element datasets. Both methods highlight the possibility of constructing a descriptor that is four times smaller than the original with a similar or even improved accuracy. Furthermore, we find that the reduced descriptors share a sizable fraction of their features across the six independent datasets, hinting at the possibility of designing material-agnostic, optimally compressed, and accurate descriptors.}, + keywords = {_tablet,ACE,descriptor dimred,descriptors,dimensionality reduction}, + file = {/home/johannes/Nextcloud/Zotero/Zeni et al_2021_Compact atomic descriptors enable accurate predictions via linear models.pdf} +} + +@article{zepeda-nunezDeepDensityCircumventing2021, + title = {Deep {{Density}}: {{Circumventing}} the {{Kohn-Sham}} Equations via Symmetry Preserving Neural Networks}, + shorttitle = {Deep {{Density}}}, + author = {Zepeda-Núñez, Leonardo and Chen, Yixiao and Zhang, Jiefu and Jia, Weile and Zhang, Linfeng and Lin, Lin}, + date = {2021-10-15}, + journaltitle = {Journal of Computational Physics}, + shortjournal = {Journal of Computational Physics}, + volume = {443}, + pages = {110523}, + issn = {0021-9991}, + doi = {10.1016/j.jcp.2021.110523}, + url = {https://www.sciencedirect.com/science/article/pii/S0021999121004186}, + urldate = {2022-07-08}, + abstract = {The recently developed Deep Potential [Phys. Rev. Lett. 120 (2018) 143001 [27]] is a powerful method to represent general inter-atomic potentials using deep neural networks. The success of Deep Potential rests on the proper treatment of locality and symmetry properties of each component of the network. In this paper, we leverage its network structure to effectively represent the mapping from the atomic configuration to the electron density in Kohn-Sham density function theory (KS-DFT). By directly targeting at the self-consistent electron density, we demonstrate that the adapted network architecture, called the Deep Density, can effectively represent the self-consistent electron density as the linear combination of contributions from many local clusters. The network is constructed to satisfy the translation, rotation, and permutation symmetries, and is designed to be transferable to different system sizes. We demonstrate that using a relatively small number of training snapshots, with each snapshot containing a modest amount of data-points, Deep Density achieves excellent performance for one-dimensional insulating and metallic systems, as well as systems with mixed insulating and metallic characters. We also demonstrate its performance for real three-dimensional systems, including small organic molecules, as well as extended systems such as water (up to 512 molecules) and aluminum (up to 256 atoms).}, + langid = {english}, + keywords = {_tablet,DFT,ML,ML-DFT,ML-ESM,prediction of electron density}, + file = {/home/johannes/Nextcloud/Zotero/Zepeda-Núñez et al_2021_Deep Density.pdf;/home/johannes/Zotero/storage/TJJ4NCEI/S0021999121004186.html} +} + +@article{zhangDeepPotentialMolecular2018, + title = {Deep {{Potential Molecular Dynamics}}: {{A Scalable Model}} with the {{Accuracy}} of {{Quantum Mechanics}}}, + shorttitle = {Deep {{Potential Molecular Dynamics}}}, + author = {Zhang, Linfeng}, + date = {2018}, + journaltitle = {Physical Review Letters}, + shortjournal = {Phys. Rev. Lett.}, + volume = {120}, + number = {14}, + doi = {10.1103/PhysRevLett.120.143001}, + file = {/home/johannes/Nextcloud/Zotero/Zhang_2018_Deep Potential Molecular Dynamics.pdf} +} + +@article{zhangMachineLearningMathbbZ2017, + title = {Machine Learning \$\{\textbackslash mathbb\{\vphantom{\}\}}{{Z}}\vphantom\{\}\vphantom\{\}\_\{2\}\$ Quantum Spin Liquids with Quasiparticle Statistics}, + author = {Zhang, Yi and Melko, Roger G. and Kim, Eun-Ah}, + date = {2017-12-13}, + journaltitle = {Physical Review B}, + shortjournal = {Phys. Rev. B}, + volume = {96}, + number = {24}, + pages = {245119}, + publisher = {{American Physical Society}}, + doi = {10.1103/PhysRevB.96.245119}, + url = {https://link.aps.org/doi/10.1103/PhysRevB.96.245119}, + urldate = {2021-05-21}, + abstract = {After decades of progress and effort, obtaining a phase diagram for a strongly correlated topological system still remains a challenge. Although in principle one could turn to Wilson loops and long-range entanglement, evaluating these nonlocal observables at many points in phase space can be prohibitively costly. With growing excitement over topological quantum computation comes the need for an efficient approach for obtaining topological phase diagrams. Here we turn to machine learning using quantum loop topography (QLT), a notion we have recently introduced. Specifically, we propose a construction of QLT that is sensitive to quasiparticle statistics. We then use mutual statistics between the spinons and visons to detect a Z2 quantum spin liquid in a multiparameter phase space. We successfully obtain the quantum phase boundary between the topological and trivial phases using a simple feed-forward neural network. Furthermore, we demonstrate advantages of our approach for the evaluation of phase diagrams relating to speed and storage. Such statistics-based machine learning of topological phases opens new efficient routes to studying topological phase diagrams in strongly correlated systems.}, + keywords = {ANN,ML,quantum spin liquid,topological phase,topological phase transition}, + file = {/home/johannes/Nextcloud/Zotero/Zhang et al_2017_Machine learning $ -mathbb Z _ 2 $ quantum spin liquids with quasiparticle.pdf;/home/johannes/Zotero/storage/39L63FS4/PhysRevB.96.html} +} + +@article{zhangMachineLearningTopological2018, + title = {Machine {{Learning Topological Invariants}} with {{Neural Networks}}}, + author = {Zhang, Pengfei and Shen, Huitao and Zhai, Hui}, + date = {2018-02-06}, + journaltitle = {Physical Review Letters}, + shortjournal = {Phys. Rev. Lett.}, + volume = {120}, + number = {6}, + pages = {066401}, + publisher = {{American Physical Society}}, + doi = {10.1103/PhysRevLett.120.066401}, + url = {https://link.aps.org/doi/10.1103/PhysRevLett.120.066401}, + urldate = {2021-05-21}, + abstract = {In this Letter we supervisedly train neural networks to distinguish different topological phases in the context of topological band insulators. After training with Hamiltonians of one-dimensional insulators with chiral symmetry, the neural network can predict their topological winding numbers with nearly 100\% accuracy, even for Hamiltonians with larger winding numbers that are not included in the training data. These results show a remarkable success that the neural network can capture the global and nonlinear topological features of quantum phases from local inputs. By opening up the neural network, we confirm that the network does learn the discrete version of the winding number formula. We also make a couple of remarks regarding the role of the symmetry and the opposite effect of regularization techniques when applying machine learning to physical systems.}, + keywords = {ANN,ML,regularization,symmetry,topological phase}, + file = {/home/johannes/Nextcloud/Zotero/Zhang et al_2018_Machine Learning Topological Invariants with Neural Networks.pdf;/home/johannes/Zotero/storage/XCPMLTVF/PhysRevLett.120.html} +} + +@article{zhangQuantumLoopTopography2017, + title = {Quantum {{Loop Topography}} for {{Machine Learning}}}, + author = {Zhang, Yi and Kim, Eun-Ah}, + date = {2017-05-22}, + journaltitle = {Physical Review Letters}, + shortjournal = {Phys. Rev. Lett.}, + volume = {118}, + number = {21}, + pages = {216401}, + publisher = {{American Physical Society}}, + doi = {10.1103/PhysRevLett.118.216401}, + url = {https://link.aps.org/doi/10.1103/PhysRevLett.118.216401}, + urldate = {2021-05-21}, + abstract = {Despite rapidly growing interest in harnessing machine learning in the study of quantum many-body systems, training neural networks to identify quantum phases is a nontrivial challenge. The key challenge is in efficiently extracting essential information from the many-body Hamiltonian or wave function and turning the information into an image that can be fed into a neural network. When targeting topological phases, this task becomes particularly challenging as topological phases are defined in terms of nonlocal properties. Here, we introduce quantum loop topography (QLT): a procedure of constructing a multidimensional image from the “sample†Hamiltonian or wave function by evaluating two-point operators that form loops at independent Monte Carlo steps. The loop configuration is guided by the characteristic response for defining the phase, which is Hall conductivity for the cases at hand. Feeding QLT to a fully connected neural network with a single hidden layer, we demonstrate that the architecture can be effectively trained to distinguish the Chern insulator and the fractional Chern insulator from trivial insulators with high fidelity. In addition to establishing the first case of obtaining a phase diagram with a topological quantum phase transition with machine learning, the perspective of bridging traditional condensed matter theory with machine learning will be broadly valuable.}, + keywords = {ANN,classification,ML,QLT,QMC,topological phase transition}, + file = {/home/johannes/Nextcloud/Zotero/Zhang_Kim_2017_Quantum Loop Topography for Machine Learning.pdf;/home/johannes/Zotero/storage/M9VFL53W/PhysRevLett.118.html} +} + +@article{zhangStrategyApplyMachine2018, + title = {A Strategy to Apply Machine Learning to Small Datasets in Materials Science}, + author = {Zhang, Ying and Ling, Chen}, + date = {2018-05-14}, + journaltitle = {npj Computational Materials}, + volume = {4}, + number = {1}, + pages = {1--8}, + publisher = {{Nature Publishing Group}}, + issn = {2057-3960}, + doi = {10.1038/s41524-018-0081-z}, + url = {https://www.nature.com/articles/s41524-018-0081-z}, + urldate = {2021-05-13}, + abstract = {There is growing interest in applying machine learning techniques in the research of materials science. However, although it is recognized that materials datasets are typically smaller and sometimes more diverse compared to other fields, the influence of availability of materials data on training machine learning models has not yet been studied, which prevents the possibility to establish accurate predictive rules using small materials datasets. Here we analyzed the fundamental interplay between the availability of materials data and the predictive capability of machine learning models. Instead of affecting the model precision directly, the effect of data size is mediated by the degree of freedom (DoF) of model, resulting in the phenomenon of association between precision and DoF. The appearance of precision–DoF association signals the issue of underfitting and is characterized by large bias of prediction, which consequently restricts the accurate prediction in unknown domains. We proposed to incorporate the crude estimation of property in the feature space to establish ML models using small sized materials data, which increases the accuracy of prediction without the cost of higher DoF. In three case studies of predicting the band gap of binary semiconductors, lattice thermal conductivity, and elastic properties of zeolites, the integration of crude estimation effectively boosted the predictive capability of machine learning models to state-of-art levels, demonstrating the generality of the proposed strategy to construct accurate machine learning models using small materials dataset.}, + issue = {1}, + langid = {english}, + keywords = {KRR,ML,models,rec-by-ruess,small data}, + file = {/home/johannes/Nextcloud/Zotero/Zhang_Ling_2018_A strategy to apply machine learning to small datasets in materials science.pdf;/home/johannes/Zotero/storage/PEGZREYC/s41524-018-0081-z.html} +} + +@article{zhaoQuantumOscillationsIrondoped2019, + title = {Quantum Oscillations in Iron-Doped Single Crystals of the Topological Insulator \$\textbackslash mathrm\{\vphantom\}{{S}}\vphantom\{\}\{\textbackslash mathrm\{b\}\}\_\{2\}\textbackslash mathrm\{\vphantom\}{{T}}\vphantom\{\}\{\textbackslash mathrm\{e\}\}\_\{3\}\$}, + author = {Zhao, Weiyao and Cortie, David and Chen, Lei and Li, Zhi and Yue, Zengji and Wang, Xiaolin}, + date = {2019-04-23}, + journaltitle = {Physical Review B}, + shortjournal = {Phys. Rev. B}, + volume = {99}, + number = {16}, + pages = {165133}, + publisher = {{American Physical Society}}, + doi = {10.1103/PhysRevB.99.165133}, + url = {https://link.aps.org/doi/10.1103/PhysRevB.99.165133}, + urldate = {2022-04-28}, + abstract = {We investigated the magnetotransport properties of Fe-doped topological insulator Sb1.96Fe0.04Te3 single crystals. With doping, the band structure changes significantly and multiple Fermi pockets become evident in the Shubnikov–de Haas oscillations, in contrast to the single frequency detected for pure Sb2Te3. Using complementary density functional theory calculations, we identify an additional bulk hole pocket introduced at the Γ point which originates from the chemical distortion associated with the Fe dopant. Experimentally, both doped and undoped samples are hole-carrier dominated; however, Fe doping also reduces the carrier density and mobility. The angle dependent quantum oscillations of Sb1.96Fe0.04Te3 were analyzed to characterize the complex Fermi surface and isolate the dimensionality of each SdH feature. Among those components, we found two oscillations frequencies, which related to two Fermi pockets are highly angle dependent. Moreover, the fermiology changes via Fe doping and may also provide a different Berry phase, as demonstrated by the Landau fan diagram, thus indicating a rich complexity in the underlying electronic structure.}, + file = {/home/johannes/Nextcloud/Zotero/Zhao et al_2019_Quantum oscillations in iron-doped single crystals of the topological insulator.pdf;/home/johannes/Zotero/storage/GGTED6FM/Zhao et al. - 2019 - Quantum oscillations in iron-doped single crystals.pdf;/home/johannes/Zotero/storage/8D5JL2DQ/PhysRevB.99.html} +} + +@thesis{zimmermannInitioDescriptionTransverse2014, + title = {Ab Initio Description of Transverse Transport Due to Impurity Scattering in Transition-Metals}, + author = {Zimmermann, Bernd}, + date = {2014}, + number = {FZJ-2014-05437}, + institution = {{Forschungszentrum Jülich GmbH Zentralbibliothek, Verlag}}, + url = {http://hdl.handle.net/2128/8063}, + urldate = {2022-08-12}, + abstract = {This thesis attempts to shed light on various spin-orbit driven transport phenomenain materials, as a crucial for the further development of the field of spintronics. Inparticular, we address the skew-scattering mechanism in dilute alloys, which gives rise to the anomalous and spin Hall effect, as well as spin-relaxation processes. We create the tools to access these quantities from \$\textbackslash textit\{ab initio\}\$ calculations in the framework of the full-potential all-electron Korringa-Kohn-Rostoker Green-function method, by (a) developing and implementing a new tetrahedron method for the calculation of complicated, multi-sheeted Fermi surfaces even of complex transition-metal compounds, and (b) developing an efficiently parallelized and thus highly scalable computer program (up to thousands of processors) for the precise calculation of scattering properties. In a first application of the new tetrahedron method, we calculate the Elliott-Yafet spin-mixing parameter on the Fermi surfaces of 5\$\textbackslash textit\{d\}\$ and 6\$\textbackslash textit\{sp\}\$ metals, and discover a yet unexplored dependence on the electron's spin-polarization direction. As we show, this anisotropy can reach gigantic values in uniaxial hcp crystals due to the emergenceof large spin-ip hot-areas or hot-loops on the Fermi surface, supported by the low symmetry of the hcp crystal. A simple model is able to reveal an interesting interplay between the orbital character of the states at special points, lines or areas in the Brillouin zone and the matrix-elements of the spin-flip part of the spin-orbit coupling operator. We further calculate the skew-scattering contribution to the anomalous Hall effect(AHE) in dilute alloys based on a ferromagnetic host for the first time. A systematic study of 3\$\textbackslash textit\{d\}\$ impurities in bcc Fe, as well as the non-magnetic hosts Pd, Pt and Au, allows us to identify trends across the periodic table. In all our calculations, we also observe a strong correlation between the spin Hall effect and anomalous Hall effect in these materials, which is of interest for the creation and detection of strongly spin-polarized currents. A Fermi-surface analysis of the contributions to the AHE reveals a non-trivial, peaked behavior at small hot-spots around spin-orbit lifted degeneracies. We then proceed to the more complicated \$\textbackslash textit\{L\}\$1\$\_\{0\}\$-ordered alloy FePt and address different kinds of disorder. We showcase the power of our method by treating the very complicated compounds Fe\$\_\{x\}\$Mn\$\_\{1-x\}\$Si and MnSi\$\_\{1-x\}\$Ge\$\_\{x\}\$, based on the non-Fermi liquid manganese silicide (MnSi). Finally, we also calculate the pure spin Hall effect for 4\$\textbackslash textit\{d\}\$/5\$\textbackslash textit\{sp\}\$ and 5\$\textbackslash textit\{d\}\$/6\$\textbackslash textit\{sp\}\$ impurities in fcc Ir and hcp Re hosts. For the latter, we discover a strong dependence on the electron's spin-polarization direction. Zimmermann, Bernd}, + isbn = {9783893369850}, + langid = {english}, + keywords = {_tablet,Boltzmann theory,Hall AHE,Hall effect,Hall SHE,juKKR,KKR,PGI-1/IAS-1,thesis,transport properties}, + file = {/home/johannes/Nextcloud/Zotero/Zimmermann_2014_Ab initio description of transverse transport due to impurity scattering in.pdf;/home/johannes/Zotero/storage/QL7I6VYG/171881.html} +} + +@article{zungerUnderstandingDopingQuantum2021, + title = {Understanding {{Doping}} of {{Quantum Materials}}}, + author = {Zunger, Alex and Malyi, Oleksandr I.}, + date = {2021-03-10}, + journaltitle = {Chemical Reviews}, + shortjournal = {Chem. Rev.}, + volume = {121}, + number = {5}, + pages = {3031--3060}, + publisher = {{American Chemical Society}}, + issn = {0009-2665}, + doi = {10.1021/acs.chemrev.0c00608}, + url = {https://doi.org/10.1021/acs.chemrev.0c00608}, + urldate = {2021-08-23}, + abstract = {Doping mobile carriers into ordinary semiconductors such as Si, GaAs, and ZnO was the enabling step in the electronic and optoelectronic revolutions. The recent emergence of a class of “quantum materialsâ€, where uniquely quantum interactions between the components produce specific behaviors such as topological insulation, unusual magnetism, superconductivity, spin–orbit-induced and magnetically induced spin splitting, polaron formation, and transparency of electrical conductors, pointed attention to a range of doping-related phenomena associated with chemical classes that differ from the traditional semiconductors. These include wide-gap oxides, compounds containing open-shell d electrons, and compounds made of heavy elements yet having significant band gaps. The atomistic electronic structure theory of doping that has been developed over the past two decades in the subfield of semiconductor physics has recently been extended and applied to quantum materials. The present review focuses on explaining the main concepts needed for a basic understanding of the doping phenomenology and indeed peculiarities in quantum materials from the perspective of condensed matter theory, with the hope of forging bridges to the chemists that have enabled the synthesis of some of the most interesting compounds in this field.}, + file = {/home/johannes/Nextcloud/Zotero/Zunger_Malyi_2021_Understanding Doping of Quantum Materials.pdf} +} + +@preamble{ "\ifdefined\DeclarePrefChars\DeclarePrefChars{'’-}\else\fi " }