Any new topic comes with new language, and AI is no different. I've put together an AI glossary to help you on your learning path, so that you can more easily make sense of the news articles, industry updates and product descriptions you may come across. If there's a term you've been seeing and I haven't included here, contact me and I'll add it to the list!
Click on whatever term you're interested in, and the definition will appear below the term.
API (Application Programming Interface)
A way for software to communicate with AI services like
ChatGPT. Developers use APIs to integrate AI into websites or apps.
Examples:
- Zapier using OpenAI to summarize emails
- Plugins adding AI to WordPress
Learn more: https://platform.openai.com/docs
Artificial Intelligence (AI)
AI refers to systems that perform tasks typically requiring
human intelligence, such as recognizing language, making decisions, or
generating content.
Examples:
- ChatGPT writing a blog post
- Image recognition in photo apps
- Virtual assistants like Siri
Learn more: https://library.uhd.edu/c.php?g=1361971&p=10058923
Bias (in AI)
Systematic favoritism or discrimination in AI output due to
biased training data or flawed modeling.
Examples:
- AI showing gender bias in hiring simulations
- Stereotypes in image generation
Learn more:
https://www.brookings.edu/articles/algorithmic-bias-detection-and-mitigation-best-practices-and-policies-to-reduce-consumer-harms/
Chain‑of‑Thought Prompting
A prompting method where the AI outlines reasoning steps
before the final answer.
Examples:
- Arithmetic problems with step-by-step solutions
Learn more: https://en.wikipedia.org/wiki/Chain-of-thought_prompting
Chatbot
A software tool that mimics human conversation using AI,
typically through a messaging interface.
Examples:
- Customer service bots
- AI companions
- ChatGPT
Learn more: https://en.wikipedia.org/wiki/Chatbot
Closed Model
Proprietary AI models where training data and code are not
public.
Examples:
- ChatGPT
- Claude
- Gemini
Learn more: https://www.openai.com/gpt-4
Context Window
The range of tokens the model can remember in one session.
Older input may be forgotten once the token limit is reached.
Examples:
- Long documents need summarization to fit within the window
Learn more: https://help.openai.com/en/articles/6825453-chatgpt-data-usage-faq
Deep Learning
A type of neural network with many layers, enabling complex
pattern recognition from raw data.
Examples:
- Advanced image/video recognition
- Self-driving car systems
Learn more: https://www.ft.com/ai-glossary
Embedding
Converting text into numerical vectors that preserve meaning
for use in tasks like search, clustering, or RAG.
Examples:
- Searching a database using vector similarity
- Semantic search engines
Learn more: https://platform.openai.com/docs/guides/embeddings
Explainability
The ability to understand and interpret how an AI made a
decision. Crucial for trust and transparency.
Examples:
- Visualizing neural network decisions
- Highlighting evidence for an answer
Learn more: https://en.wikipedia.org/wiki/Explainable_artificial_intelligence
Few-Shot Learning
Providing a few examples within the prompt to help the model
understand the task before responding.
Examples:
- Providing 2–3 labeled examples of question/answer pairs
Learn more: https://en.wikipedia.org/wiki/Few-shot_learning
Fine‑Tuning
Adjusting a pre-trained model on task-specific data to
improve performance.
Examples:
- Customizing GPT for financial reports
Learn more: https://en.wikipedia.org/wiki/Fine-tuning_(machine_learning)
Foundation Model
A massive pre-trained model like GPT or BERT that serves as
the base for multiple downstream tasks.
Examples:
- GPT family
- LLaMA
- Stable Diffusion
Learn more: https://en.wikipedia.org/wiki/Foundation_model
Generative AI
AI that creates new content—text, images, code—based on
training data.
Examples:
- ChatGPT writing stories
- DALL·E generating images
Learn more: https://www.zendesk.com/blog/generative-ai-glossary/
Guardrails
Safety features embedded in AI systems to prevent harmful,
biased, or dangerous output.
Examples:
- Blocking hate speech
- Preventing misinformation
Learn more:
https://www.microsoft.com/en-us/security/blog/2023/06/15/seven-guardrails-to-help-developers-build-safer-generative-ai/
Hallucination (in AI)
When an AI generates plausible-sounding but factually
incorrect or nonsensical content. This can occur because the model predicts
text based on patterns rather than verifying facts.
Examples:
- Chatbot fabricating quotes or data
- AI writing fictional citations in research assistance
Learn more:
https://en.wikipedia.org/wiki/Hallucination_(artificial_intelligence)
Inference
The
process of using a trained model to generate predictions or outputs, like
answering a question or writing text.
Examples:
- ChatGPT replying to your prompt
- Image classifier labeling a photo
Learn more:
https://machinelearningmastery.com/what-is-inference-in-machine-learning/
Large Language Model (LLM)
A deep learning model trained on vast text corpora to
understand and generate human-like language.
Examples:
- ChatGPT
- Google Gemini
- Claude
Learn more: https://en.wikipedia.org/wiki/Large_language_model
Machine Learning (ML)
A subset of AI where algorithms learn patterns from data to
make predictions or decisions without explicit programming.
Examples:
- Spam filters
- Movie recommendations
- Fraud detection systems
Learn more: https://pryon.com/glossary/
Multimodal Model
An AI model that processes more than one type of data—like
text and images together.
Examples:
- GPT-4 with vision
- Gemini summarizing charts
Learn more: https://openai.com/gpt-4
Natural Language Generation (NLG)
Systems that automatically produce human-readable text from
data.
Examples:
- Weather reports
- Code comments
- Image captions
Learn more: https://en.wikipedia.org/wiki/Natural-language_generation
Natural Language Processing (NLP)
AI subfield focused on analyzing, interpreting, and
generating human language.
Examples:
- Chatbots
- Translation apps
- Sentiment analysis
Learn more: https://library.mit.edu/ai-glossary
Neural Network
ML model inspired by the human brain, composed of
interconnected nodes (“neurons”) that process data through layers.
Examples:
- Facial recognition
- Voice assistants
- LLMs like GPT
Learn more: https://library.uhd.edu/c.php?g=1361971&p=10058923
Open Source Model
AI models made publicly available, often with access to
weights and code, enabling customization and transparency.
Examples:
- Mistral
- LLaMA 2
- Stable Diffusion
Learn more: https://huggingface.co/models
Parameter
A learned weight within a neural network controlling model
behavior.
Examples:
- GPT-4 has hundreds of billions of parameters
Learn more: https://library.uhd.edu/c.php?g=1361971&p=10058923
Prompt
Input text/instruction given to an LLM to guide its output.
Examples:
- “Write a 150‑word summary…”
- “Generate a recipe for vegan lasagna”
Learn more: https://www.axios.com/2023/06/22/artificial-intelligence-glossary
Prompt Engineering
Crafting prompts strategically to get accurate and useful AI
output.
Examples:
- Using “Chain-of-thought”
- Few-shot prompting
Learn more: https://en.wikipedia.org/wiki/Prompt_engineering
Retrieval-Augmented Generation (RAG)
Combines LLMs with external data sources, allowing the AI to
pull in fresh, factual information at query time.
Examples:
- Chatbot searching a private knowledge base for accurate answers
Learn more: https://en.wikipedia.org/wiki/Retrieval-augmented_generation
Text Completion
The AI predicts and continues text based on the input
prompt.
Examples:
- Typing 'How are you' and the model replies 'doing today?'
Learn more: https://platform.openai.com/docs/guides/gpt
Text-to-Image
AI that generates images from text descriptions using
diffusion models or GANs. These systems interpret descriptive input and create
realistic or artistic imagery that matches the description.
Examples:
- DALL·E creating a 'cat playing violin'
- Midjourney visualizing fantasy scenes
- Stable Diffusion generating landscapes
Learn more: https://en.wikipedia.org/wiki/Text-to-image_model
Token
A unit of text (word or subword) processed by models for
input and output.
Examples:
- “Hello” as “Hell” + “o”
Learn more: https://shelf.io/blog/ai-glossary/
Token Limit
The maximum number of tokens (text chunks) an LLM can
process at once—includes both prompt and response. Larger limits allow longer
conversations or documents.
Examples:
- GPT-4 Turbo has 128k token limit
Learn more: https://openai.com/pricing
Training Data
The text and other data used to train the AI. It forms the
foundation of what the model knows and how it responds.
Examples:
- Books, websites, public datasets
Learn more: https://openai.com/enterprise-privacy
Transformer
A neural network architecture using self-attention
mechanisms, revolutionizing language tasks.
Examples:
- GPT
- BERT
- LLaMA
Learn more: https://docs.nebuly.com/
Vector Database
A database optimized to store and retrieve embeddings for
similarity search.
Examples:
- Pinecone
- Weaviate
- FAISS
Learn more: https://www.pinecone.io/learn/vector-database/
Zero-Shot Learning
A model’s ability to perform tasks it wasn't explicitly
trained on, using general understanding from training data.
Examples:
- Summarizing a blog post without training on blog formats
- Translating new language pairs
Learn more: https://en.wikipedia.org/wiki/Zero-shot_learning
Shopping Cart |
|
Questions?
Projects?
Ideas?