Results for "attention weights"

AdvertisementAd space — search-top

18 results

Attention Head Intermediate

A single attention mechanism within multi-head attention.

AI Economics & Strategy
Self-Attention Intermediate

Attention where queries/keys/values come from the same sequence, enabling token-to-token interactions.

Transformers & LLMs
Sparse Attention Intermediate

Attention mechanisms that reduce quadratic complexity.

AI Economics & Strategy
Graph Attention Network Intermediate

GNN using attention to weight neighbor contributions dynamically.

Model Architectures
Cross-Attention Intermediate

Attention between different modalities.

Computer Vision
Weight Initialization Intermediate

Methods to set starting weights to preserve signal/gradient scales across layers.

Foundations & Theory
Fine-Tuning Intermediate

Updating a pretrained model’s weights on task-specific data to improve performance or adapt style/behavior.

Large Language Models
Parameter-Efficient Fine-Tuning Intermediate

Techniques that fine-tune small additional components rather than all weights to reduce compute and storage.

Foundations & Theory
Quantization Intermediate

Reducing numeric precision of weights/activations to speed inference and reduce memory with acceptable accuracy loss.

Foundations & Theory
Pruning Intermediate

Removing weights or neurons to shrink models and improve efficiency; can be structured or unstructured.

Foundations & Theory
Open-Weight Model Intermediate

Models whose weights are publicly available.

AI Economics & Strategy
Transformer Intermediate

Architecture based on self-attention and feedforward layers; foundation of modern LLMs and many multimodal models.

Transformers & LLMs
Positional Encoding Intermediate

Injects sequence order into Transformers, since attention alone is permutation-invariant.

Foundations & Theory
Interpretability Intermediate

Studying internal mechanisms or input influence on outputs (e.g., saliency maps, SHAP, attention analysis).

Foundations & Theory
Causal Mask Intermediate

Prevents attention to future tokens during training/inference.

AI Economics & Strategy
Key-Value Cache Intermediate

Stores past attention states to speed up autoregressive decoding.

AI Economics & Strategy
Attention Intermediate

Mechanism that computes context-aware mixtures of representations; scales well and captures long-range dependencies.

Transformers & LLMs
Multi-Head Attention Intermediate

Allows model to attend to information from different subspaces simultaneously.

AI Economics & Strategy

Welcome to AI Glossary

The free, self-building AI dictionary. Help us keep it free—click an ad once in a while!

Search

Type any question or keyword into the search bar at the top.

Browse

Tap a letter in the A–Z bar to browse terms alphabetically, or filter by domain, industry, or difficulty level.

3D WordGraph

Fly around the interactive 3D graph to explore how AI concepts connect. Click any word to read its full definition.