Advertisement ยท 728 ร— 90

Posts by Aditi Krishnapriyan

Preview
Towards Fast, Specialized Machine Learning Force Fields: Distilling... The foundation model (FM) paradigm is transforming Machine Learning Force Fields (MLFFs), leveraging general-purpose representations and scalable training to perform a variety of computational...

7/ This was a very fun project with Ishan Amin and Sanjeev Raja, and will appear at #ICLR2025! Paper and code below:

Paper: openreview.net/forum?id=1du...

Code: github.com/ASK-Berkeley...

1 year ago 1 0 0 0
Post image

6/ The distilled MLFFs are much faster to run than the original large-scale MLFF: not everyone has the GPU resources to use big models and many scientists only care about studying specific systems (w/ the correct physics!). This is a way to get the best of all worlds!

1 year ago 1 0 1 0
Post image

5/ We can also balance training at scale efficiently (often w/ minimal constraints) with distilling the correct physics into the small MLFF at test time: e.g., taking energy gradients to get conservative forces, and ensuring energy conservation for molecular dynamics.

1 year ago 0 0 1 0
Post image Post image Post image

4/ Smaller, specialized MLFFs distilled from the large-scale model are more accurate than training from scratch on the same subset of data: the representations from the large-scale model help boost performance, while the smaller models are much faster to run

1 year ago 0 0 1 0
Post image

3/ We formulate our distillation procedure as the smaller MLFF is trained to match Hessians of the energy predictions of the large-scale model (using subsampling methods to improve efficiency). This works better than distillation methods to try to match features.

1 year ago 1 0 1 0

2/ Model distillation involves transferring the general-purpose representations learned by a large-scale model into smaller, faster models: in our case, specialized to specific regions of chemical space. We can use these faster MLFFs for a variety of downstream tasks.

1 year ago 1 0 1 0
Post image

1/ Machine learning force fields are hot right now ๐Ÿ”ฅ: models are getting bigger + being trained on more data. But how do we balance size, speed, and specificity? We introduce a method for doing model distillation on large-scale MLFFs into fast, specialized MLFFs! More details below:

#ICLR2025

1 year ago 11 2 1 1

๐Ÿ˜ƒ

1 year ago 2 0 0 0
Advertisement

Would also appreciate being added, thanks!

1 year ago 2 0 1 0