-
Notifications
You must be signed in to change notification settings - Fork 16
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
More details about training process #15
Comments
We trained on proteins from the DIPS data set. You can find more details in the supplement: https://www.biorxiv.org/content/10.1101/2023.06.07.544059v2.supplementary-material We trained on SDA data (~25 A) because most of our real data stems from SDA. We also tested it with real DSSO data (~30 A) which worked in many cases. There is also a network trained on photoAA crosslinking data (10 A). We haven't released the distogram networks yet which would work with arbitrary cutoffs. For the training script, use the original Uni-Fold script: https://github.com/dptech-corp/Uni-Fold/blob/main/train_multimer.sh |
Hi @lhatsk. Sorry for writing on a very old thread, but just to be sure: did you rather use the fine-tuning Uni-Fold script here https://github.com/dptech-corp/Uni-Fold/blob/main/finetune_multimer.sh (the |
We only fine-tuned the network in the paper due to limited resources. I used the finetune_multimer.sh script you linked. Same parameters just removed the warmup steps. |
Can you provide more details about your training, such as the source of the training data? Is it necessary for the crosslink data to represent a 25A distance between two proteins in a complex? Also, could you share the training script?
Thank you for your work; it's truly a significant breakthrough.
The text was updated successfully, but these errors were encountered: