Results for "feedforward"
Architecture based on self-attention and feedforward layers; foundation of modern LLMs and many multimodal models.
Gradients grow too large, causing divergence; mitigated by clipping, normalization, careful init.
Networks with recurrent connections for sequences; largely supplanted by Transformers for many tasks.
Neural networks can approximate any continuous function under certain conditions.
Control without feedback after execution begins.