Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

dataset size effect on fine tuning #6

Open
shadow1999k opened this issue Dec 23, 2024 · 0 comments
Open

dataset size effect on fine tuning #6

shadow1999k opened this issue Dec 23, 2024 · 0 comments

Comments

@shadow1999k
Copy link

Hi,
Thanks for sharing the implementations. I had a question about whether using my own credit related dataset size would affect the quality of fine-tuning the Llama model. Imagine that it is a table formatted dataset with lots of feature (about 2000 features) but total available records (with binary target laebles 0 as non-defaulter and 1 as defaulter) about 200,000 samples.

Are there any suggestions or limitations related to such data that I might consider when deciding whether fine-tuning might be a good option for my project?

Thanks

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

1 participant