Skip to content

Comments

flip order of weight+bias application in LayerNormANE#2

Open
carsonswope wants to merge 1 commit intoapple:mainfrom
carsonswope:main
Open

flip order of weight+bias application in LayerNormANE#2
carsonswope wants to merge 1 commit intoapple:mainfrom
carsonswope:main

Conversation

@carsonswope
Copy link

Hi, I'm attempting to duplicate the pytorch LayerNorm functionality, and the formula that pytorch uses is clearly (out * weight) + bias, which does not match the code in LayerNormANE.

So I changed it for my use case, and thought I'd open a PR in case this is in fact a bug.

However.. looking at 4b37184, it looks like there is some history and/or legacy reasons for the order being this way, so feel free to reject if I'm missing something :)

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

1 participant