fix rmsnorm init weight bug.
#59
by
Shan1990 - opened
Using torch.ones to init rmsnorm weight. And torch.empty gets random weight tensor, which maybe out of float value limits.
@zRzRzRzRzRzRzR pls review this pr, thx.
check now
ZHANGYUXUAN-zR changed pull request status to
merged