Results for "on-device training"
Converting text into discrete units (tokens) for modeling; subword tokenizers balance vocabulary size and coverage.
An RNN variant using gates to mitigate vanishing gradients and capture longer context.
The set of tokens a model can represent; impacts efficiency, multilinguality, and handling of rare strings.
A model that assigns probabilities to sequences of tokens; often trained by next-token prediction.
Generates sequences one token at a time, conditioning on past tokens.
Predicts masked tokens in a sequence, enabling bidirectional context; often used for embeddings rather than generation.
Reinforcement learning from human feedback: uses preference data to train a reward model and optimize the policy.
Model-generated content that is fluent but unsupported by evidence or incorrect; mitigated by grounding and verification.
Systematic differences in model outcomes across groups; arises from data, labels, and deployment context.
Rules and controls around generation (filters, validators, structured outputs) to reduce unsafe or invalid behavior.
Human or automated process of assigning targets; quality, consistency, and guidelines matter heavily.
Measure of consistency across labelers; low agreement indicates ambiguous tasks or poor guidelines.
Standardized documentation describing intended use, performance, limitations, data, and ethical considerations.
Practices for operationalizing ML: versioning, CI/CD, monitoring, retraining, and reliable production management.
Converts logits to probabilities by exponentiation and normalization; common in classification and LMs.
System for running consistent evaluations across tasks, versions, prompts, and model settings.
System design where humans validate or guide model outputs, especially for high-stakes decisions.
Hidden behavior activated by specific triggers, causing targeted mispredictions or undesired outputs.
Models that process or generate multiple modalities, enabling vision-language tasks, speech, video understanding, etc.
AI focused on interpreting images/video: classification, detection, segmentation, tracking, and 3D understanding.
A model is PAC-learnable if it can, with high probability, learn an approximately correct hypothesis from finite samples.
Assigning labels per pixel (semantic) or per instance (instance segmentation) to map object boundaries.
Reduction in uncertainty achieved by observing a variable; used in decision trees and active learning.
Converting audio speech into text, often using encoder-decoder or transducer architectures.
Measures divergence between true and predicted probability distributions.
A theoretical framework analyzing what classes of functions can be learned, how efficiently, and with what guarantees.
Variability introduced by minibatch sampling during SGD.
A narrow minimum often associated with poorer generalization.
Early architecture using learned gates for skip connections.
Built-in assumptions guiding learning efficiency and generalization.