-
Notifications
You must be signed in to change notification settings - Fork 29
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Tricks for training? #6
Comments
Hello, did you try to run our provided scripts or to reproduce the results by yourself? |
I run the code provided in your repository. Only for the KS equation(regular), I can't get a good result, the L2 error reported in the paper is 3.49e − 04, but I got an error of 4.155e-02. |
Apologies for the delayed response. We recently tested our code for that example and successfully reproduced the results on an NVIDIA A6000. However, we did notice a performance drop when using other GPUs. This may be attributed to the example being sensitive to initial values, causing small errors to accumulate and eventually lead to significant discrepancies. |
Can you share your code? I'm using Windows system and have been trying to reproduce the causality PINN without the modified MLP and Taylor-mode automatic differentiation. However, I've noticed that the results are not satisfactory. |
I reproduced the results using the code provided in the repository and didn't modify anything. I did notice that modified MLP and Taylor-mode automatic differentiation can improve accuracy and efficiency. |
Can we exchange contact information for further discussion? If you have QQ, my QQ number is: 1194453382. |
Thank you for sharing your idea and source code. I would like to know whether any other tricks were adopted in the training process for the results I got are not as good as the ones reported in the paper.
The text was updated successfully, but these errors were encountered: