Glossary

A small, evolving reference for terms that show up repeatedly across the blog. Inline glossary terms open a compact definition card, while each term here has its own entry page for longer notes and future expansion.

Glossary

Attention

A neural mechanism that lets a model weigh which parts of the input matter most when processing or predicting.

Architecture LLMs

Glossary

Autoencoder

A model that learns a compressed representation of its input and then tries to reconstruct the original input from it.

Architecture Representation

Glossary

Collaborative Filtering

A recommendation approach that predicts preferences from patterns in many users' interactions rather than item content alone.

Recommenders Personalization

Glossary

Corpus

A structured collection of text or other examples that you analyze, search, or use to train and evaluate a model.

Data NLP

Glossary

Cosine Similarity

A measure of how aligned two vectors are, commonly used to compare embeddings by direction rather than raw magnitude.

Similarity Retrieval

Glossary

Embedding

A dense numeric representation that places similar items close together in vector space.

Representation NLP

Glossary

Fine-Tuning

The process of continuing training on a pre-trained model so it becomes better suited to a narrower task or domain.

Training LLMs

Glossary

Inference

The stage where a trained model is used to make predictions or generate outputs on new inputs.

Deployment Models

Glossary

Large Language Model

A very large language model, usually transformer-based, trained on broad text data and adapted for many downstream tasks.

LLMs Language

Glossary

Learning Rate

A training hyperparameter that controls how large each parameter update is during optimization.

Training Optimization

Glossary

Overfitting

A failure mode where a model matches the training data too closely and performs worse on new, unseen examples.

Training Generalization

Glossary

Pre-Training

The initial large-scale training phase that teaches a model general patterns before narrower task-specific adaptation.

Training LLMs

Glossary

Prompt

The input text or instructions given to a generative model to shape the response it produces.

LLMs Generative AI

Glossary

Regularization

Techniques that limit model complexity or penalize certain behaviors so the model generalizes better to new data.

Training Generalization

Glossary

Token

The basic unit a language model reads and predicts, which may be a word, character, or subword fragment.

LLMs Language

Glossary

Transfer Learning

Reusing knowledge learned on one task or dataset to help solve a different but related task more efficiently.

Training Reuse

Glossary

Transformer

A neural network architecture built around attention mechanisms that became the foundation for many modern language models.

Architecture LLMs