[HW3] Using nn.MultiheadAttention
and nn.LayerNorm
#140
-
Could we use |
Beta Was this translation helpful? Give feedback.
Answered by
seominjoon
Nov 17, 2021
Replies: 1 comment
-
You can either use PyTorch's or copy from Annotated Transformer, whichever is easier for you. |
Beta Was this translation helpful? Give feedback.
0 replies
Answer selected by
yuneg11
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
You can either use PyTorch's or copy from Annotated Transformer, whichever is easier for you.