Results for "data mismatch"
Predicting case success probabilities.
Requirement to reveal AI usage in legal decisions.
Identifying suspicious transactions.
AI applied to scientific problems.
AI proposing scientific hypotheses.
A conceptual framework describing error as the sum of systematic error (bias) and sensitivity to data (variance).
Training one model on multiple tasks simultaneously to improve generalization through shared structure.
Configuration choices not learned directly (or not typically learned) that govern training or architecture.
A scalar measure optimized during training, typically expected loss over data, sometimes with regularization terms.
A robust evaluation technique that trains/evaluates across multiple splits to estimate performance variability.
Halting training when validation performance stops improving to reduce overfitting.
A gradient method using random minibatches for efficient training on large datasets.
Randomly zeroing activations during training to reduce co-adaptation and overfitting.
Nonlinear functions enabling networks to approximate complex mappings; ReLU variants dominate modern DL.
Mechanism that computes context-aware mixtures of representations; scales well and captures long-range dependencies.
Attention where queries/keys/values come from the same sequence, enabling token-to-token interactions.
The text (and possibly other modalities) given to an LLM to condition its output behavior.
Crafting prompts to elicit desired behavior, often using role, structure, constraints, and examples.
Achieving task performance by providing a small number of examples inside the prompt without weight updates.
Architecture that retrieves relevant documents (e.g., from a vector DB) and conditions generation on them to reduce hallucinations.
Fine-tuning on (prompt, response) pairs to align a model with instruction-following behaviors.
A preference-based training method optimizing policies directly from pairwise comparisons without explicit RL loops.
Model trained to predict human preferences (or utility) for candidate outputs; used in RLHF-style pipelines.
Local surrogate explanation method approximating model behavior near a specific input.
Ordering training samples from easier to harder to improve convergence or generalization.
A hidden variable influences both cause and effect, biasing naive estimates of causal impact.
Standardized documentation describing intended use, performance, limitations, data, and ethical considerations.
Automated testing and deployment processes for models and data workflows, extending DevOps to ML artifacts.
A broader capability to infer internal system state from telemetry, crucial for AI services and agents.
Time from request to response; critical for real-time inference and UX.