Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

More details about training process #15

Open
Li-dacheng opened this issue Nov 30, 2023 · 3 comments
Open

More details about training process #15

Li-dacheng opened this issue Nov 30, 2023 · 3 comments

Comments

@Li-dacheng
Copy link

Can you provide more details about your training, such as the source of the training data? Is it necessary for the crosslink data to represent a 25A distance between two proteins in a complex? Also, could you share the training script?

Thank you for your work; it's truly a significant breakthrough.

@lhatsk
Copy link
Collaborator

lhatsk commented Dec 4, 2023

We trained on proteins from the DIPS data set. You can find more details in the supplement: https://www.biorxiv.org/content/10.1101/2023.06.07.544059v2.supplementary-material

We trained on SDA data (~25 A) because most of our real data stems from SDA. We also tested it with real DSSO data (~30 A) which worked in many cases. There is also a network trained on photoAA crosslinking data (10 A). We haven't released the distogram networks yet which would work with arbitrary cutoffs.

For the training script, use the original Uni-Fold script: https://github.com/dptech-corp/Uni-Fold/blob/main/train_multimer.sh

@u-lupo
Copy link

u-lupo commented Oct 15, 2024

Hi @lhatsk. Sorry for writing on a very old thread, but just to be sure: did you rather use the fine-tuning Uni-Fold script here https://github.com/dptech-corp/Uni-Fold/blob/main/finetune_multimer.sh (the train_multimer one you shared seems to be for training from scratch). And, did you use the same parameters as in this script?

@lhatsk
Copy link
Collaborator

lhatsk commented Oct 16, 2024

We only fine-tuned the network in the paper due to limited resources. I used the finetune_multimer.sh script you linked. Same parameters just removed the warmup steps.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

3 participants