-
Notifications
You must be signed in to change notification settings - Fork 3k
Open
Description
Line 42 in 37baab7
| self.c_proj = nn.Linear(config.n_embd, config.n_embd) |
Why do we need an additional linear transformation after the MHA and before the MLP when the dimensions are the same?
(I understand that this is how the initial transformer implementation was written, but I took this operation to be for dimension consistency between sequential attention operations. It seems superfluous here since the first linear in the MLP can already take linear combinations of the attention outputs.)
Metadata
Metadata
Assignees
Labels
No labels