In Transformer models, what is the main purpose of positional encoding?
To specify the absolute position of each token in a sequence
To preserve the sequential order of tokens during self-attention operations
Baroque art features strong contrasts, while Rococo art prefers more subtle transitions
Baroque art is generally larger in scale than Rococo art

Machine Learning Exercises are loading ...