Thursday, November 27, 2025

From Random Determinants to the Ground State

Hao Zhang and Matthew Otten (2025)
Highlighted by Jan Jensen


The paper introduces a method they call TrimCI that very efficiently finds a relatively small set of determinants that accurately describes strongly correlated systems. (Well, it actually works for any system, but the main advantage is for strongly correlated systems). 

Unlike most new correlation methods, this one is actually simple enough to describe in a few sentences. TrimCI starts by constructing a set of orthogonal (non-optimised!) MOs (e.g. by diagonalising the AO overlap matrix). From these MOs you construct a small number of random determinants (e.g.100), construct the wavefunction (i.e. construct the Hamiltonian matrix and diagonalise, as per usual). Then you compute all the Hamiltonian elements between this wavefunction ($H_{ij}$) and the remaining determinants and add determinants with sufficiently large |$H_{ij}$| to the wavefunction. Finally, there is the trimming step "which removes negligible basis states by first diagonalising randomised blocks of the core and then performing a global diagonalising step on the surviving set." And repeat.


The authors find that this approach converges much quicker than other similar methods, using many fewer determinants. Another big advantage is that the method does not require a single-determinant ground state as a starting point and is thus not sensitive to how much such a single-determinant deviates from the actual wavefunction.

So, what's the catch here? In order to be practically useful, we need to compute energy differences with mHa accuracy, and I did not see any TrimCI results for chemical systems where the energy had converged to that kind of accuracy. It's possible that error cancellation can help here, but that needs to be investigated. The authors do look at extrapolation, which looks promising, but needs to be systematically investigated. Yet another option is to use the (compact) TrimCI wavefunction as an ansatz for dynamic-correlation methods.

It's also not clear what AO basis set it used for some of these calculations (including the one shown above). I suspect small basis sets are used and even FCI energies with very small basis sets are of limited practical use. Are the TrimCI calculations on large systems still practical with more realistic basis sets?

Nevertheless, this seems like a very promising step in the right direction.


This work is licensed under a Creative Commons Attribution 4.0 International License.



Friday, October 31, 2025

Electron flow matching for generative reaction mechanism prediction

Joonyoung F. Joung, Mun Hong Fong, Nicholas Casetti, Jordan P. Liles,  Ne S. Dassanayake & Connor W. Coley (2025)
Highlighted by Jan Jensen



While the title says reaction mechanism prediction, it's really reaction mechanism-based reaction outcome prediction. The approach uses glow matching (a generalization of diffusion-based  approaches) to predict changes to the bond-electron (BE) matrix (basically connectivity matrix with the lone pair electron count on the diagonal), thus ensuring mass and charge conservation because changes in the BE matrix are constrained to sum to 0. The method is trained 1.4 million elementary reaction steps derived primarily from the USPTO dataset.  

Recursive predictions yield a complete reaction mechanism step by step, starting from the reactants. (I assume the products are defined as the state where no more changes are predicted.) The method is probabilistic so several different reaction outcomes are possible if the process is repeated, and ranked according to frequency. Another option is to use DFT calculation to rank the different mechanisms.
 
Like any ML method its applicability is tied to the training set. For example, of 22,000 reactions from patents reported in 2024 that were not assigned  a specific reaction class in the Pistachio dataset, the approach successfully recovered products in only 351 cases. However, the authors show that a new reaction class can be added with as few as 32 examples.

Thursday, September 25, 2025

Fundamental Study of Density Functional Theory Applied to Triplet State Reactivity: Introduction of the TRIP50 Dataset

William B. Hughes, Mihai V. Popescu, Robert S. Paton (2025)
Highlighted by Jan Jensen


While this paper presents an interesting and useful benchmark and dataset of barrier heights involving organic molecules in the triplet state, I am highlighting this paper for a different reason. 

While compiling the data set the authors "observed a  common tendency for triplet SCF calculations to converge non-Aufbau solutions, resulting in catastrophic predictions in both thermochemistry and activation energy  barriers and leading to errors as high as 26.4 kcal/mol." They go on to note that "Since such  errors  cannot be predicted a priori, manual inspection of spin densities for triplet-state calculations can be helpful to ensure the lowest triplet state has been converged with KS-DFT."

I remember the days when the SCF would routinely fail to converge even for simple singlet ground state molecules. So when they did converge and you got odd results, one of the first things you checked was the orbitals. But those days are long gone and I don't think it would occur to me now. I'd be much more likely to ascribe it to some deficiency in the functional. 

I now wonder how many of such wrong conclusions are scattered throughout the literature, especially for molecules with "funky" electronic structure, such as transition metal complexes.  Manual inspection of the MOs is not going to be a practical option for many of these studies, and SCF stability checks did not identify all problems! 

However, most QM packages have several options for the MO guess and it might be a good idea to use more than one of them and check whether they all converge to the same SCF solution. It'll be just like the old days.


This work is licensed under a Creative Commons Attribution 4.0 International License.



Friday, August 29, 2025

UMA: A Family of Universal Models for Atoms

Brandon M. Wood, Misko Dzamba1, Xiang Fu, Meng Gao, Muhammed Shuaibi, Luis Barroso-Luque, Kareem Abdelmaqsoud, Vahe Gharakhanyan, John R. Kitchin, Daniel S. Levine, Kyle Michel, Anuroop Sriram, Taco Cohen, Abhishek Das, Ammar Rizvi, Sushree Jagriti Sahoo, Zachary W. Ulissi, C. Lawrence Zitnick (2025)
Highlighted by Jan Jensen

I use xTB extensively in my research and I am often asked why don't switch to machine learning potentials (MLPs) instead. My answer has always been that they have too many limitations: limited atom types, no charged molecules, can't handle reactions, efficiency on CPUs, solvent effects, etc. I know these can be overcome by making bespoke MLPs, then it is not really a simple replacement for xTB, but a whole new workflow. 

However, the new UMA MLP from Meta seems to address all but one of my concerns (more on that below). UMA is trained to reproduce DFT energies and gradients calculated for nearly half a billion 3D structures, spanning molecules, surfaces, reactions, etc, containing atoms from virtually all of the periodic table. It is also possible to specify the charge and multiplicity and the cost seems to be comparable to xTB when running on CPUs, when interfaced with ORCA. So this is all very encouraging.

Two main questions remain. One is the accuracy, and by that I mean how faithfully it reproduces ωB97M-V/def2-TZVPD results (in the case of molecules) for molecules outside it's training set. AFAIK nothing is published yet, but encouraging results are being shared online.

The other main question is how to include implicit solvent effects. In cases where it is OK to optimise i the gas phase, one option is to compute the solvation energy with some other method and add it to the gas phase UMA results. Even if you do that at the DFT level, UMA has still saved you a lot if time. However, if the problem requires optimisation in solvent, then you have to use a faster method like xTB to compute the solvent effects on the gradient in order to get any time-savings. Depending on how well xTB does on the system of interest, this could "contaminate" the UMA results. Alternatively, a purely ML approach would basically amount to redoing UMA for molecules with continuum solvation included. Explicit solvation is fine in principle, but impractical for routine applications.

Anyway, before this is resolved there could be some fairly routine applications that still cannot be address satisfactorily with MLPs.



This work is licensed under a Creative Commons Attribution 4.0 International License.

Thursday, July 31, 2025

Computing solvation free energies of small molecules with first principles accuracy

 J. Harry Moore, Daniel  J. Cole, and Gábor Csányi (2025)
Highlighted by Jan Jensen


Local CCSD(T) has made it possible to reach near chemical accuracy for many real life applications. However, most real life applications happen in solution,  where the only realistic option is still continuum solvation methods, which, in general, do offer chemical accuracy (especially for charged systems). 

In principle, this can be fixed explicit solvation at the CCSD(T) level but of course the need for sampling makes this practically impossible at present. However, this paper by Moore and co-workers is a step in that direction. 

The idea is that ML potentials now approach chemical accuracy and are fast enough for proper sampling. The authors developed a MLP that is compatible with alchemical transformation (needed for sampling efficiency) and showed that experimental logP values (the difference in solvation energy between water and octanol) of drug-like compounds can be predicted within 0.65 kcal/mol (0.45 log units), i.e chemical accuracy.

Furthermore, the calculations took "only" about 4-7 days per molecule (octanol simulations converge slower than water) on a single node, containing either 8 NVIDIA A100, or 8 NVIDIA L40S, GPUs. While this is too slow for routine applications it is fast enough to create benchmark sets. This is great news since experimental solvation energies only are measured for relatively small molecules.

However, there are some caveats. The main one is that only neutral molecules where tested, since the MLP only was trained on neutral compounds, and it is not clear whether the same accuracy can be obtained for charged systems. 



This work is licensed under a Creative Commons Attribution 4.0 International License.

Friday, June 27, 2025

g-xTB: A General-Purpose Extended Tight-Binding Electronic Structure Method For the Elements H to Lr (Z=1–103)

Thomas Froitzheim, Marcel Müller, Andreas Hansen, and Stefan Grimme (2025)
Highlighted by Jan Jensen

This highlight is coming to you live from Oslo where Grimme presented the release of (a preliminary version of) g-xTB at WATOC yesterday. Grimme has been talking about this method for a few years now and many in the community (not least me) have been waiting for this moment with some anticipation.

To cut a long story (and paper!) short g-xTB offers mid-level DFT accuracy at semi-empirical cost (30-50% slower than GFN2-xTVB) including reaction energies and barrier heights. That's quite a statement! Structures of TM-complexes also seems to be improved.

This comes about a month after the release of the general ML-potential UMA and the paper offers some tantalising preliminary comparisons. For example, MAEs for reaction energies and forward barrier heights for the BH9 data shown above are 1.6 and 1.9 kcal/mol, respectively! However, UMA also appears to have some problems with some large extended systems, some non-covalent interactions, and some TM complexes. 

Given that UMA and g-xTB are roughly the same costs (on CPUs) it will be interesting to see how these methods will co-evolve over the coming years. 

Note that you can apply both methods through ORCA 



This work is licensed under a Creative Commons Attribution 4.0 International License.



Friday, May 30, 2025

Repurposing quantum chemical descriptor datasets for on-the-fly generation of informative reaction representations: application to hydrogen atom transfer reactions

Javier E. Alfonso-Ramos, Rebecca M. Neeser, Thijs Stuyver (2024)
Highlighted by Jan Jensen



If you have very little data, the single most useful thing you can do is find good descriptors. Sigman, Doyle, and others have shown this very nicely for reactivity predictions of transition metal containing catalysts, but there's less systematic work for other types of reactions. 

In this paper, Stuyver and co-workers suggest a descriptor set for barriers of hydrogen atom transfer (HAT) reactions that are based valence bond (VB) theory. In practise this translates to computing the bond dissociation energies (BDEs) without relaxing the geometry, and combining them with the BD free energies (BDFE, where ΔBDFE corresponds to ΔGrp). In addition, atomic Mulliken charges, spin densities, and buried volume are also added. All descriptors are predicted by surrogate models to avoid QM-based calculations.

Using these descriptors they get significantly better barrier predictions compared to fingerprint or graph convolution representation, even using simple models such as linear regression. Even the simple Bell-Evans-Polanyi model (a linear model based solely on ΔGrp) outperforms the models using fingerprints and graph convolution, with an R2 of 0.71 compared to 0.65 for graph convolution. For, comparison the R2s for the VB-based descriptors are 0.80-0.85, depending on the ML-model.

I wonder what other approximate chemical methods contain inspirations for new descriptors?


This work is licensed under a Creative Commons Attribution 4.0 International License.