Q2. What is the effect of the periodic filters on the properties of diamond and graphite?
Note that while the single-atom filters are circular due to the rotational invariance, the periodic filters become rotationally equivariant with respect to the orientation of the lattice, which still keeps the property prediction rotationally invariant.
Q3. What have the authors contributed in "Schnet – a deep learning architecture for molecules and materials" ?
SchNet this paper is a variant of DTNNs that can learn a representation from first principles that adapts to the task and scale at hand from property prediction across chemical compound space to force fields in the configurational space of single molecules.
Q4. What future works have the authors mentioned in the paper "Schnet – a deep learning architecture for molecules and materials" ?
This gives rise to the possibility to encode known quantumchemical constraints and symmetries within the model without losing the flexibility of a neural network. These encouraging results will guide future work such as studies of larger molecules and periodic systems as well as further developments toward interpretable deep learning architectures to assist chemistry research. This is crucial in order to be able to accurately represent, e. g., the full potential-energy surface and, in particular, its anharmonic behavior. The authors have presented the deep learning architecture SchNet which can be applied to a variety of applications ranging from the prediction of chemical properties for diverse datasets of molecules and materials to highly accurate predictions of potential energy surfaces and energy-conserving force fields.
Q5. What could be used to understand the formation and distribution of defects?
2. In solids, such local chemical potentials could be used to understand the formation and distribution of defects, such as vacancies and interstitials.
Q6. What is the function of the filter-generating network?
The filter-generating network determines how interactions between atoms are modeled and can be used to constrainthe model and include chemical knowledge.
Q7. What is the cost of a training step?
Since there are a maximum number of atoms being located within a given cutoff, the computational cost of a training step scales linearly with the system size if the authors precompute the indices of nearby atoms.
Q8. How many epochs does SchNet require to converge?
the model requires much less epochs to converge, e.g., using 110k training examples reduces the required number of epochs from 2400 with T = 2 to less than 750 epochs with T = 6.
Q9. Why do the authors need to use the PBCs to describe the interactions of atoms?
Due to the linearity of the convolution, the authors are, therefore, able to apply the PBCs directly to the filter to accurately describe the atom interactions while keeping invariance to the choice of the unit cell.
Q10. How many examples do you use for the training set?
The authors use 20k C20 reference calculations as the training set, 4.5k examples for early stopping, and report the test error on the remaining data.
Q11. How many examples are used to predict bulk crystal formation energies?
The authors employ SchNet to predict formation energies for bulk crystals using 69 640 structures and reference calculations from the Materials Project (MP) repository.
Q12. What is the effect of normalizing the filter response xl+1i?
The authors find that the training is more stable when normalizing the filter response xl+1i by the number of atoms within the cutoff range.
Q13. How does SchNet perform on the small training set?
SchNet performs better while using the combined loss with energies and forces on 1000 reference calculations than training on energies of 50 000 examples.
Q14. What is the effect of the filter-generating network on the properties of diamond and graphite?
While the authors have followed a data-driven approach where the authors only incorporate basic invariances in the filters, careful design of the filter-generating network provides the possibility to incorporate further chemical knowledge in the network.
Q15. What is the distribution of the errors of all predicted properties?
The distributions of the errors of all predicted properties are shown in Appendix A. Extending SchNet with interpretable, property-specific output layers, e.g., for the dipole moment,57 is subject to future work.
Q16. How many training examples are used to test the SchNet model?
the data are randomly split into 60 000 training examples, a validation set of 4500 examples and the remaining data as test set.