Results for "training loss"

AdvertisementAd space — search-top

54 results

Dropout Intermediate

Randomly zeroing activations during training to reduce co-adaptation and overfitting.

Foundations & Theory
Next-Token Prediction Intermediate

Training objective where the model predicts the next token given previous tokens (causal modeling).

Foundations & Theory
DPO Intermediate

A preference-based training method optimizing policies directly from pairwise comparisons without explicit RL loops.

Optimization
Curriculum Learning Intermediate

Ordering training samples from easier to harder to improve convergence or generalization.

Foundations & Theory
Data Augmentation Intermediate

Expanding training data via transformations (flips, noise, paraphrases) to improve robustness.

Foundations & Theory
Federated Learning Intermediate

Training across many devices/silos without centralizing raw data; aggregates updates, not data.

Foundations & Theory
Reproducibility Intermediate

Ability to replicate results given same code/data; harder in distributed training and nondeterministic ops.

Foundations & Theory
Compute Intermediate

Hardware resources used for training/inference; constrained by memory bandwidth, FLOPs, and parallelism.

Foundations & Theory
Distillation Intermediate

Training a smaller “student” model to mimic a larger “teacher,” often improving efficiency while retaining performance.

Foundations & Theory
Data Poisoning Intermediate

Maliciously inserting or altering training data to implant backdoors or degrade performance.

Foundations & Theory
Automation Bias Intermediate

Tendency to trust automated suggestions even when incorrect; mitigated by UI design, training, and checks.

Foundations & Theory
Variance Term Intermediate

Error due to sensitivity to fluctuations in the training dataset.

AI Economics & Strategy
Learning Rate Schedule Intermediate

Adjusting learning rate over training to improve convergence.

AI Economics & Strategy
Warmup Intermediate

Gradually increasing learning rate at training start to avoid divergence.

AI Economics & Strategy
Causal Mask Intermediate

Prevents attention to future tokens during training/inference.

AI Economics & Strategy
Gradient Leakage Intermediate

Recovering training data from gradients.

AI Economics & Strategy
Model Inversion Intermediate

Inferring sensitive features of training data.

AI Economics & Strategy
Generative Model Advanced

Models that learn to generate samples resembling training data.

Diffusion & Generative Models
Saddle Plateau Intermediate

Flat high-dimensional regions slowing training.

Foundations & Theory
Deceptive Alignment Advanced

Model behaves well during training but not deployment.

AI Safety & Alignment
Exposure Bias Intermediate

Differences between training and inference conditions.

Model Failure Modes
Simulation Advanced

Artificial environment for training/testing agents.

Simulation & Sim-to-Real
Dataset Shift Intermediate

Differences between training and deployed patient populations.

AI in Healthcare
Hybrid Training Advanced

Combining simulation and real-world data.

Simulation & Sim-to-Real

Welcome to AI Glossary

The free, self-building AI dictionary. Help us keep it free—click an ad once in a while!

Search

Type any question or keyword into the search bar at the top.

Browse

Tap a letter in the A–Z bar to browse terms alphabetically, or filter by domain, industry, or difficulty level.

3D WordGraph

Fly around the interactive 3D graph to explore how AI concepts connect. Click any word to read its full definition.