Cross-Attention

Attention mechanism that connects two different sequences (e.g., source and target languages in translation), unlike self-attention which works within one sequence.