Happy to see our work on SLCO2A1 with @smlea.bsky.social, Nakanishi and Newstead labs out now. Important insight into how prostaglandin and many drugs are transported. Hats off to @weitse-hsu.bsky.social for computational work!
@oxfordbiochemistry.bsky.social
www.nature.com/articles/s41...
Posts by Wei-Tse Hsu
Now out in JACS! 🎉 : "Computing Solvation Free Energies of Small Molecules with Experimental Accuracy"! It's been a pleasure to collaborate on this with Harry Moore (@jhmchem.bsky.social) & Gábor Csányi pubs.acs.org/doi/10.1021/...
New Preprint!! We show that binding entropy can be quantitatively predicted from crystallographic ensemble models, accounting for both protein conformational entropy and solvent entropy! www.biorxiv.org/content/10.6...
🚀 Bottom line:
With careful filtering, co-folding predictions can indeed teach ML about binding affinity.
👉 Read the full JCIM paper: pubs.acs.org/doi/full/10....
Work with Aniket Magarkar
@boehringerglobal.bsky.social and @philbiggin.bsky.social @ox.ac.uk
(6/6)
🔎 SI highlights:
- AEV-PLIG beats Boltz-2 in 4 target classes in the FEP benchmark (loses 1, ties 6); both are competitive with FEP+ in some cases.
- ipLDDT & ligand pLDDT are also effective filters; pTM, PAE, PDE are not
- Boltz confidence seems to generalize better than its structure module
(5/6)
❓ Are co-folding predictions good enough to train scoring functions?
👉 Yes — with careful filtering. We see no performance difference b/w models trained on:
- experimental structures
- corresponding co-folding predictions
This holds across AEV-PLIG, EHIGN, and RF-Score.
(4/6)
❓ When can we trust a co-folding prediction?
👉 From reproducing HiQBind with Boltz-1x, a few simple heuristics are recommended high-quality cofolding augmentation:
1️⃣ single-chain systems
2️⃣ Boltz confidence > 0.9
3️⃣ train–test similarity > 60%
(3/6)
❓ How much can data augmentation actually improve scoring?
👉 Short answer: only if the added data are high-quality. Adding BindingNet v1 clearly improved performance, but v2 did not—despite being 10x larger—due to its substantially lower quality.
Quality beats quantity.
(2/6)
📢 Can AI-Predicted Complexes Teach Machine Learning to Compute Drug Binding Affinity?
In our recent JCIM work, we tested whether co-folding models can be used for data augmentation for training ML-based scoring functions (SFs).
We asked 3 simple but critical questions. 👇
(1/6)