-
Notifications
You must be signed in to change notification settings - Fork 306
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Is it possible to add cross attention aka encoder outputs to SRU++ ? #164
Comments
Hi @hadaev8 At the moment, we haven't implemented a SRU++ "decoder" in which there are both self attention and cross attention. There are two options you could choose:
Note we are assuming all input & hidden dimensions are
|
@taoleicn Where can I find the transform_module definition? |
@hadaev8 yes and no. Yes in the sense that within each SRU++ layer, the layer will attend to both self outputs and the memory inputs. Re: transform_module how SRUpp set transform_module as the attention sub-module: forward method of SRUppCell: |
@taolei87 |
@hadaev8 i'm not sure i follow. can you elaborate more on your question? |
@taolei87 |
It is a 3-dimensional tensor (memory_seq_len, batch_size, hidden_size). See an illustration below:
I updated the pseudo code in the previous reply for a correction. |
@taolei87 Spotted this thing: |
No description provided.
The text was updated successfully, but these errors were encountered: