Encoder-Decoder Architecture
Cross-Attention Mechanism
Process in the decoder that allows it to focus on specific parts of the encoder's output, weighting the importance of each input token for generating the current output token.
← 뒤로