Concept graph
Glossary
Short definitions with deeper context and cross-links to sibling terms.
Safety
attention
Attention weights how much each token should attend to others when producing representations.
Neural Networks
Attention Mechanism
A technique that allows models to focus on specific parts of the input data when making predictions.
Inference
flash attention
flash attention is a core generative-AI concept used across modeling, product, and governance discussions.
Machine Learning
graph-attention-network
A neural network architecture that employs attention mechanisms to process graph-structured data.
Training
multi-head attention
multi-head attention is a core generative-AI concept used across modeling, product, and governance discussions.
Machine Learning
scalable-dot-product-attention
An efficient variant of attention mechanism designed for large datasets.
Data
scaled dot-product attention
scaled dot-product attention is a core generative-AI concept used across modeling, product, and governance discussions.
Product
transformer
A transformer is an architecture built from attention layers; most frontier LLMs are decoder-only or encoder–decoder transformers.