Results for "deep learning"
Deep Learning
IntermediateA branch of ML using multi-layer neural networks to learn hierarchical representations, often excelling in vision, speech, and language.
Deep Learning is a type of machine learning that uses structures called neural networks, which are inspired by how the human brain works. Imagine a series of layers where each layer learns to recognize different features of an image, like edges, shapes, and eventually, whole objects. This is how ...
AI applied to scientific problems.
Awareness and regulation of internal processes.
A mismatch between training and deployment data distributions that can degrade model performance.
The relationship between inputs and outputs changes over time, requiring monitoring and model updates.
A parameterized mapping from inputs to outputs; includes architecture + learned parameters.
Configuration choices not learned directly (or not typically learned) that govern training or architecture.
A high-priority instruction layer setting overarching behavior constraints for a chat model.
A preference-based training method optimizing policies directly from pairwise comparisons without explicit RL loops.
Model trained to predict human preferences (or utility) for candidate outputs; used in RLHF-style pipelines.
Systematic differences in model outcomes across groups; arises from data, labels, and deployment context.
Rules and controls around generation (filters, validators, structured outputs) to reduce unsafe or invalid behavior.
Feature attribution method grounded in cooperative game theory for explaining predictions in tabular settings.
Measure of consistency across labelers; low agreement indicates ambiguous tasks or poor guidelines.
When some classes are rare, requiring reweighting, resampling, or specialized metrics.
Policies and practices for approving, monitoring, auditing, and documenting models in production.
Standardized documentation describing intended use, performance, limitations, data, and ethical considerations.
Structured dataset documentation covering collection, composition, recommended uses, biases, and maintenance.
Techniques that fine-tune small additional components rather than all weights to reduce compute and storage.
Observing model inputs/outputs, latency, cost, and quality over time to catch regressions and drift.
PEFT method injecting trainable low-rank matrices into layers, enabling efficient fine-tuning.
Inputs crafted to cause model errors or unsafe behavior, often imperceptible in vision or subtle in text.
Maliciously inserting or altering training data to implant backdoors or degrade performance.
Attacks that infer whether specific records were in training data, or reconstruct sensitive training examples.
Mechanisms for retaining context across turns/sessions: scratchpads, vector memories, structured stores.
Chooses which experts process each token.
Set of all actions available to the agent.
Formal framework for sequential decision-making under uncertainty.
Fundamental recursive relationship defining optimal value functions.
Expected cumulative reward from a state or state-action pair.
Inferring sensitive features of training data.