Results for "step breakdown"
Diffusion model trained to remove noise step by step.
Choosing step size along gradient direction.
Stepwise reasoning patterns that can improve multi-step tasks; often handled implicitly or summarized for safety/privacy.
Adjusting learning rate over training to improve convergence.
Generative model that learns to reverse a gradual noise process.
Controls amount of noise added at each diffusion step.
Diffusion performed in latent space for efficiency.
Monte Carlo method for state estimation.
Iterative method that updates parameters in the direction of negative gradient to minimize loss.
Popular optimizer combining momentum and per-parameter adaptive step sizes via first/second moment estimates.
Converting text into discrete units (tokens) for modeling; subword tokenizers balance vocabulary size and coverage.
Architecture that retrieves relevant documents (e.g., from a vector DB) and conditions generation on them to reduce hallucinations.
Fine-tuning on (prompt, response) pairs to align a model with instruction-following behaviors.
Search algorithm for generation that keeps top-k partial sequences; can improve likelihood but reduce diversity.
Samples from the k highest-probability tokens to limit unlikely outputs.
Coordinating tools, models, and steps (retrieval, calls, validation) to deliver reliable end-to-end behavior.
Matrix of second derivatives describing local curvature of loss.
Prevents attention to future tokens during training/inference.
Balancing learning new behaviors vs exploiting known rewards.
Models time evolution via hidden states.
Optimization under uncertainty.
Breaking tasks into sub-steps.
Temporary reasoning space (often hidden).
US approval process for medical AI devices.
Training a smaller “student” model to mimic a larger “teacher,” often improving efficiency while retaining performance.