Rotary Positional Embeddings

Intermediate

Encodes positional information via rotation in embedding space.

AdvertisementAd space — term-top

Why It Matters

Rotary positional embeddings are significant for enhancing the performance of AI models in understanding language. By allowing models to better capture the relationships between words based on their positions, they improve tasks such as translation, summarization, and sentiment analysis. This innovation contributes to the ongoing advancement of natural language processing technologies.

Rotary positional embeddings are a method for encoding positional information in transformer models, designed to enhance the model's ability to capture relative positions of tokens in a sequence. Unlike traditional absolute positional encodings, which assign fixed positions to tokens, rotary embeddings utilize a rotation mechanism in the embedding space. The positional encoding for a token at position p is given by: PE(p) = [cos(p / 10000^(2i/d)), sin(p / 10000^(2i/d))] for each dimension i of the embedding. This allows the model to maintain a sense of relative positioning, which is particularly beneficial for tasks requiring an understanding of the order and distance between tokens. The incorporation of rotary embeddings has been shown to improve performance in various NLP tasks by enabling better generalization and context awareness.

Keywords

Domains

Related Terms

Welcome to AI Glossary

The free, self-building AI dictionary. Help us keep it free—click an ad once in a while!

Search

Type any question or keyword into the search bar at the top.

Browse

Tap a letter in the A–Z bar to browse terms alphabetically, or filter by domain, industry, or difficulty level.

3D WordGraph

Fly around the interactive 3D graph to explore how AI concepts connect. Click any word to read its full definition.