fix rmsnorm init weight bug.
#59
by
Shan1990
- opened
Using torch.ones to init rmsnorm weight. And torch.empty gets random weight tensor, which maybe out of float value limits.
@zRzRzRzRzRzRzR pls review this pr, thx.
check now
zRzRzRzRzRzRzR
changed pull request status to
merged