Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

perplexity 微调之后变大了?需要继续微调? #416

Open
cqray1990 opened this issue Aug 31, 2024 · 1 comment
Open

perplexity 微调之后变大了?需要继续微调? #416

cqray1990 opened this issue Aug 31, 2024 · 1 comment
Labels
bug Something isn't working

Comments

@cqray1990
Copy link

微调之前估计效果:
***** eval metrics *****
eval_loss = 2.5866
eval_runtime = 0:02:10.44
eval_samples = 1000
eval_samples_per_second = 7.666
eval_steps_per_second = 0.958
perplexity = 13.2845

sft 微调之后的估计效果,perplexity 怎么还变大了

***** eval metrics *****
epoch = 2.0
eval_loss = 2.8984
eval_runtime = 0:00:00.33
eval_samples = 10
eval_samples_per_second = 29.506
eval_steps_per_second = 2.951
perplexity = 18.1445

@cqray1990 cqray1990 added the bug Something isn't working label Aug 31, 2024
@shibing624
Copy link
Owner

建议:
1.训练集质量不高,提升数据质量;
2.lr 调低
3.batch size调大

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
bug Something isn't working
Projects
None yet
Development

No branches or pull requests

2 participants