Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

【Question】Question about initial finetune loss #6201

Open
SolenoidWGT opened this issue Feb 19, 2025 · 2 comments
Open

【Question】Question about initial finetune loss #6201

SolenoidWGT opened this issue Feb 19, 2025 · 2 comments

Comments

@SolenoidWGT
Copy link

Hello, recently I read a blog about colossial supporting lora finetune deepseek-v3, it is a very great work for opensource community. But I have a question about the picture in the blog.

Image

My question is why the initial loss in this picture is so high, whether it is from scratch testing or due to other reasons?

@ver217
Copy link
Member

ver217 commented Feb 20, 2025

As lora weight is initialied from random

@447428054
Copy link

Hello, recently I read a blog about colossial supporting lora finetune deepseek-v3, it is a very great work for opensource community. But I have a question about the picture in the blog.

Image

My question is why the initial loss in this picture is so high, whether it is from scratch testing or due to other reasons?

R1 SFT Bug,loss should start from 1

https://zhuanlan.zhihu.com/p/26682456562

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

3 participants