Results for "statistical learning"
Fine-tuning on (prompt, response) pairs to align a model with instruction-following behaviors.
A high-capacity language model trained on massive corpora, exhibiting broad generalization and emergent behaviors.
Ensuring model behavior matches human goals, norms, and constraints, including reducing harmful or deceptive outputs.
Automated detection/prevention of disallowed outputs (toxicity, self-harm, illegal instruction, etc.).
Techniques to understand model decisions (global or local), important in high-stakes and regulated settings.
Studying internal mechanisms or input influence on outputs (e.g., saliency maps, SHAP, attention analysis).
Local surrogate explanation method approximating model behavior near a specific input.
Protecting data during network transfer and while stored; essential for ML pipelines handling sensitive data.
Processes and controls for data quality, access, lineage, retention, and compliance across the AI lifecycle.
Central system to store model versions, metadata, approvals, and deployment state.
Time from request to response; critical for real-time inference and UX.
How many requests or tokens can be processed per unit time; affects scalability and cost.
Hardware resources used for training/inference; constrained by memory bandwidth, FLOPs, and parallelism.
Reducing numeric precision of weights/activations to speed inference and reduce memory with acceptable accuracy loss.
A dataset + metric suite for comparing models; can be gamed or misaligned with real-world goals.
Removing weights or neurons to shrink models and improve efficiency; can be structured or unstructured.
System for running consistent evaluations across tasks, versions, prompts, and model settings.
Stress-testing models for failures, vulnerabilities, policy violations, and harmful behaviors before release.
Methods to protect model/data during inference (e.g., trusted execution environments) from operators/attackers.
Models that process or generate multiple modalities, enabling vision-language tasks, speech, video understanding, etc.
Identifying and localizing objects in images, often with confidence scores and bounding rectangles.
Assigning labels per pixel (semantic) or per instance (instance segmentation) to map object boundaries.
Error due to sensitivity to fluctuations in the training dataset.
AI subfield dealing with understanding and generating human language, including syntax, semantics, and pragmatics.
Measures divergence between true and predicted probability distributions.
Converting audio speech into text, often using encoder-decoder or transducer architectures.
Optimization problems where any local minimum is global.
Generating speech audio from text, with control over prosody, speaker identity, and style.
A point where gradient is zero but is neither a max nor min; common in deep nets.
The shape of the loss function over parameter space.