Expand description
Attention module
Structs§
- Generate a padding attention mask.
- Cache for the Multi Head Attention layer.
- Multihead attention forward pass input argument.
- Multihead attention outputs.
- The multihead attention module as describe in the paper Attention Is All You Need.
- Configuration to create a Multi Head Attention layer using the init function.
- The record type for the module.
- The record item type for the module.
Functions§
- Generate an autoregressive attention mask.
- Generation padding attention mask.