Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Further training and/or finetuning #6

Open
rohanvarm opened this issue Jan 16, 2025 · 1 comment
Open

Further training and/or finetuning #6

rohanvarm opened this issue Jan 16, 2025 · 1 comment

Comments

@rohanvarm
Copy link

Hi, do you have the code used to do the training/ complete finetuning of the model?

@awf
Copy link

awf commented Jan 22, 2025

Yes, Minimol was trained using our Graphium fork:
https://github.com/graphcore-research/graphium-smg

and there is a config file in Minimol that should work with Graphium for training:
https://github.com/graphcore-research/minimol/blob/main/minimol/ckpts/minimol_v1/config.yaml

But: It was trained on an IPU cluster some time ago, so there is a high likelihood that if trained on a GPU cluster today, results might differ (certainly we have seen differences where small implementation details mean we do not reach the same loss on GPU). We don't currently maintain a working training pipeline, so it is not clear what support we could offer in making it run, but at least you can inspect the code.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants