Results for "function calls"
Methods for breaking goals into steps; can be classical (A*, STRIPS) or LLM-driven with tool calls.
Agent calls external tools dynamically.
Constraining model outputs into a schema used to call external APIs/tools safely and deterministically.
Enables external computation or lookup.
Coordinating tools, models, and steps (retrieval, calls, validation) to deliver reliable end-to-end behavior.
Expected return of taking action in a state.
A scalar measure optimized during training, typically expected loss over data, sometimes with regularization terms.
Probability of data given parameters.
Converts logits to probabilities by exponentiation and normalization; common in classification and LMs.
Expected cumulative reward from a state or state-action pair.
Lowest possible loss.
A function measuring prediction error (and sometimes calibration), guiding gradient-based optimization.
A parameterized function composed of interconnected units organized in layers with nonlinear activations.
Direction of steepest ascent of a function.
Nonlinear functions enabling networks to approximate complex mappings; ReLU variants dominate modern DL.
Neural networks can approximate any continuous function under certain conditions.
Inferring reward function from observed behavior.
Learning a function from input-output pairs (labeled data), optimizing performance on predicting outputs for unseen inputs.
A parameterized mapping from inputs to outputs; includes architecture + learned parameters.
The learned numeric values of a model adjusted during training to minimize a loss function.
Iterative method that updates parameters in the direction of negative gradient to minimize loss.
The shape of the loss function over parameter space.
Matrix of second derivatives describing local curvature of loss.
Combines value estimation (critic) with policy learning (actor).
Learns the score (∇ log p(x)) for generative sampling.
Matrix of first-order derivatives for vector-valued functions.
Matrix of curvature information.
Describes likelihoods of random variable outcomes.
Visualization of optimization landscape.
Choosing step size along gradient direction.