Vedwix glossary

The working
studio glossary.

Real definitions, real examples, and how Vedwix uses each term in client work. 110 terms across 6 categories.

Terms110
Categories6
AI (51)
RAG

Retrieval-Augmented Generation: an LLM technique where the model retrieves relevant documents before generating a response.

Fine-tuning

The process of training a base LLM further on your own data to specialize its outputs.

Embedding

A vector representation of text, image, or other data used for similarity search.

Vector Database

A database optimized for storing and querying high-dimensional vectors (embeddings).

Hybrid Search

A search technique combining keyword (BM25) and semantic (vector) retrieval.

Reranker

A second-stage model that reorders retrieved results by relevance to the query.

Eval Harness

A test suite for AI features that measures quality, regressions, and edge cases.

AI Agent

An LLM-powered system that autonomously chooses tools and takes multi-step actions.

Tool Use

An LLM's ability to call external functions (search, calculator, database, etc.) during a response.

Function Calling

A structured way for LLMs to invoke developer-defined functions with typed arguments.

Structured Output

LLM responses constrained to a JSON schema or specific format.

LLM-as-Judge

Using one LLM to evaluate the outputs of another LLM (or itself) against criteria.

Prompt Engineering

The practice of crafting LLM inputs to produce better, more reliable outputs.

Chain of Thought

A prompting technique where the model is asked to reason step by step before answering.

Prompt Caching

API-level caching of prompt prefixes to reduce cost and latency on repeated calls.

Context Window

The maximum number of tokens an LLM can process in a single call.

Token

The unit of text an LLM processes — typically 3-4 characters or about 0.75 of a word.

Temperature

A parameter (0-2) controlling how random or deterministic an LLM's output is.

Top-p / Nucleus Sampling

A sampling parameter that limits LLM output to the smallest set of tokens whose probabilities sum to p.

LoRA

Low-Rank Adaptation: a lightweight fine-tuning method that trains small adapter layers on top of a frozen base model.

Supervised Fine-Tuning (SFT)

Fine-tuning a model on labeled input-output pairs.

RLHF

Reinforcement Learning from Human Feedback: training a model based on human preference rankings of outputs.

DPO (Direct Preference Optimization)

A simpler alternative to RLHF that trains directly on preference pairs without a reward model.

QLoRA

Quantized LoRA: combines LoRA with 4-bit quantization to fine-tune large models on consumer GPUs.

Quantization

Reducing the numerical precision of model weights to make inference cheaper and faster.

Small Language Model (SLM)

A compact LLM (typically 1-15B parameters) optimized for specific tasks or constrained environments.

Mixture of Experts (MoE)

An architecture where each forward pass routes tokens through a subset of "expert" sub-networks.

Transformer

The neural network architecture underlying virtually every modern LLM.

Attention Mechanism

The transformer component that lets each token in a sequence attend to other tokens.

Pretraining

The initial training of a foundation model on massive amounts of unlabeled text.

Foundation Model

A large pretrained model that can be adapted to many downstream tasks.

Frontier Model

The most capable AI models — currently models like Claude Opus, GPT-4, and Gemini Ultra.

Reasoning Model

An LLM trained or post-trained to perform multi-step reasoning, often using extended hidden thinking tokens.

Multimodal Model

An LLM that can process more than text — images, audio, video, or structured inputs.

AI Observability

Logging, tracing, and monitoring of LLM calls in production.

Red-Teaming

Adversarial testing of an AI system to find harmful, biased, or wrong outputs.

Jailbreak

A prompt that bypasses an LLM's safety training to make it produce restricted content.

Prompt Injection

An attack where malicious instructions in user input or retrieved data hijack the LLM's behavior.

Hallucination

When an LLM generates plausible-sounding but factually incorrect information.

Benchmarks

Standardized evaluation suites for comparing AI models on common tasks.

MCP (Model Context Protocol)

An open protocol for connecting LLMs to external tools, data sources, and contexts.

A2A (Agent-to-Agent)

Protocols and patterns for AI agents to discover, communicate, and coordinate with each other.

Multi-Agent System

A system where multiple AI agents collaborate or compete on a task.

Inference

The process of running an already-trained model to produce predictions or generations.

KV Cache

A runtime cache of attention key/value tensors that speeds up sequential token generation.

Speculative Decoding

An inference technique using a smaller "draft" model to propose tokens that a larger model verifies.

GGUF

A quantized model file format used for efficient CPU and GPU inference, popularized by llama.cpp.

Edge AI

Running AI models on devices (phones, browsers, IoT) rather than in the cloud.

Context Rot

A degradation in LLM quality as context length grows, even within the model's stated window.

Guardrails

Runtime checks that validate or filter LLM inputs and outputs against policies.

Total Cost of Ownership (TCO)

The full cost of running an AI feature in production — inference, eval, observability, ops, and people.

SEO (31)
Programmatic SEO

Generating large numbers of SEO-optimized pages from templates and structured data.

Long-Tail Keywords

Specific, lower-volume search queries that collectively make up most search demand.

Keyword Research

Identifying search queries to target based on volume, difficulty, intent, and business fit.

Search Intent

The underlying goal behind a user's search query — informational, navigational, commercial, transactional.

E-E-A-T

Experience, Expertise, Authoritativeness, Trustworthiness — Google's framework for content quality.

YMYL

Your Money or Your Life — content where accuracy can affect health, finances, or wellbeing.

Helpful Content System

A Google ranking system that demotes pages written primarily for SEO rather than humans.

Thin Content

Pages with little or no unique value — typically targeted in pSEO penalties.

SERP

Search Engine Results Page — what users see after a search.

AI Overviews

Google's AI-generated summaries that appear at the top of SERPs for many queries.

llms.txt

A proposed standard file telling LLM crawlers how to find and use a site's content.

AI Search Optimization

Optimizing content for AI-powered search interfaces (ChatGPT, Perplexity, Claude, AI Overviews).

Featured Snippet

A boxed answer at the top of a SERP, pulled from one ranking page.

Schema Markup

Structured data added to a page to help search engines understand its content.

JSON-LD

A JSON-based format for embedding structured data (like schema markup) in web pages.

Core Web Vitals

Google's page experience metrics: LCP, INP (formerly FID), and CLS.

XML Sitemap

A file listing all important URLs on a site, submitted to search engines.

Crawl Budget

The number of URLs a search engine will crawl on a site within a given timeframe.

Indexation

The process of search engines storing a page in their database for retrieval in search results.

Canonical Tag

A link tag indicating the preferred URL when multiple URLs serve the same content.

Meta Title (Title Tag)

The HTML title element — the primary signal for what a page is about and the headline shown in SERPs.

Meta Description

A 150-160 character summary of a page that often appears under the title in SERPs.

On-Page SEO

Optimizations applied directly to a page — content, headings, links, schema, meta tags.

Internal Linking

Links between pages on the same site — a major signal of topical authority and crawl depth.

Backlink

A link from another site pointing to yours — a major off-page ranking signal.

Domain Authority

A Moz/Ahrefs-style aggregate score (0-100) estimating a domain's ability to rank.

Site Architecture

The hierarchical structure of a site — how URLs and internal links are organized.

URL Structure

The pattern of URL paths — affects SEO, UX, and shareability.

Pillar Page

A comprehensive top-level page on a topic, supporting and linked to by many narrower cluster pages.

Topic Cluster

A pillar page plus its supporting cluster pages — a coordinated set on one topic.

Pillar / Cluster Strategy

A content architecture pairing one comprehensive pillar page with many supporting cluster pages on a topic.

Building with these concepts?

Vedwix ships AI, web, brand, and SEO work. Brief us in three sentences.

Start a project