Results for "masked"
Masked Language Model
Intermediate
Predicts masked tokens in a sequence, enabling bidirectional context; often used for embeddings rather than generation.
Self-Supervised Learning
Intermediate
Learning from data by constructing “pseudo-labels” (e.g., next-token prediction, masked modeling) without manual annotation.