top of page

Comprehensive Glossary of AI Terms and Definitions & List of Generative AI Products

Updated: 4 days ago

Updated: March 6, 2025


1. Glossary

A

Adversarial AI: Techniques used to manipulate AI models, often by introducing deceptive inputs to exploit vulnerabilities.


Agentic AI: Highly autonomous AI capable of goal-setting, long-term planning, and decision-making with minimal human intervention.


AGI (Artificial General Intelligence): AI systems with human-like ability to learn and perform any intellectual task.


AI Agent: A system that perceives its environment, makes decisions, and takes actions to achieve specific goals.


AI Ethics: The study and application of moral principles in AI development, ensuring responsible use of AI technologies.


Algorithm: A step-by-step procedure or formula for solving a problem or completing a task.


Ambient Intelligence: Electronic environments that are sensitive and responsive to human presence.


Artificial Intelligence (AI): The simulation of human intelligence by machines, including learning, reasoning, and self-correction.


Attention Mechanism: A technique allowing neural networks to focus on specific parts of input data.


AutoML (Automated Machine Learning): AI-driven tools that automate the process of training and optimizing machine learning models.


Autoencoder: Neural networks that learn efficient data representations by attempting to reconstruct their inputs.


Autonomous Agents: AI-driven systems that interact with environments in real-time, learning and adapting without direct human input (e.g., AI trading bots, robotics, digital assistants).


Augmented Intelligence: AI systems designed to enhance human capabilities rather than replace them.


B

Backpropagation: Algorithm for training neural networks by calculating gradients and adjusting weights.


Bias (AI Ethics): The presence of systematic errors in AI models that lead to unfair, skewed, or discriminatory outcomes. Bias in AI can arise from biased training data, flawed algorithms, or human prejudices embedded in decision-making processes.


Bias (Statistical): Systematic error that causes models to favor certain outcomes.


Bias in AI: Systematic errors in AI models caused by biased training data or flawed algorithms, leading to unfair outcomes.


Big Data: Extremely large datasets that may be analyzed to reveal patterns and trends.


Black Box Problem: The inability to explain or interpret how AI systems arrive at their decisions.


BERT (Bidirectional Encoder Representations from Transformers): Language model designed to understand context by considering words bidirectionally.


C

ChatGPT: Conversational AI system developed by OpenAI based on the GPT architecture.


Classification: Task of assigning inputs to specific categories or classes.


Clustering: Grouping similar data points together based on patterns or similarities.


CNN (Convolutional Neural Network): Neural network architecture specialized for image processing.


Computer Vision: AI technology that allows machines to interpret and analyze visual information from the world, such as images and videos.


Contrastive Learning: A self-supervised learning technique that helps models differentiate between similar and dissimilar examples to improve feature representation.


Corpus: A collection of texts used for training language models.


D

Data Mining: Process of discovering patterns and extracting valuable information from large datasets.


Data Science: Interdisciplinary field that uses scientific methods to extract knowledge from data.


Decision Tree: A tree-like model of decisions and their possible consequences.


Deep Learning: Machine learning approach using neural networks with multiple layers.


Deep Learning Iteration: The repeated training of neural networks through multiple layers to improve the model's ability to recognize complex patterns.


Diffusion Models: Generative models that create data by gradually denoising a random signal.


Dimensionality Reduction: Techniques to reduce the number of variables under consideration.


E

Edge AI: AI computation performed on local devices rather than relying on cloud-based processing, reducing latency and improving privacy.


Embeddings: Dense vector representations of data that capture semantic relationships.


Embodied AI: AI systems integrated with physical bodies that can interact with their environment.


Ensemble Learning: Combining multiple models to improve performance and robustness.


Ethical AI: The practice of designing AI systems that are fair, unbiased, and aligned with human values and ethical principles.


Explainable AI (XAI): AI systems designed to make their functioning transparent and interpretable.


F

Federated Learning: A decentralized ML approach where models are trained across multiple devices while keeping data local for privacy.


Feed-Forward Layers: Neural network layers that process token embeddings by applying transformations to refine meaning and improve understanding.


Feature: Individual measurable property or characteristic used as input for machine learning.


Feature Engineering: Process of selecting, transforming, or creating features to improve model performance.


Few-Shot Learning: A technique where AI models learn new tasks with only a few training examples.


Fine-tuning: Process of further training a pre-trained model on a specific dataset or task.


Foundation Models: Large-scale pre-trained models (e.g., GPT, BERT, DALL·E) that serve as a base for fine-tuning on specialized tasks.


G

GANs (Generative Adversarial Networks): Architecture using two neural networks competing against each other.


GPT (Generative Pre-trained Transformer): Large language model architecture using transformers.


Generative AI: AI systems that create new content, such as text, images, music, or videos, often using models like GANs and transformers.


Gradient Descent: A fundamental optimization algorithm used in training deep learning models by adjusting model weights to minimize error.


Graph Neural Networks: Neural networks that operate on graph-structured data.


H

Hallucination (AI): When AI systems generate false or misleading information that appears plausible but is incorrect.


Hidden Layer: Layers in neural networks between input and output layers.


Human-in-the-Loop (HITL) AI: AI systems that incorporate human oversight and feedback to improve accuracy and decision-making.


Hyperparameter: Parameter whose value is set before the learning process begins.


I

Image Generation: Creating new images using AI, often with text prompts.


Image Recognition: Technology that identifies objects, people, or scenes in digital images.


Inference: Process of using a trained model to make predictions.


Information Retrieval: Systems for obtaining relevant information from large collections.


Intent Recognition: Identifying the purpose or goal behind user inputs or queries.


K

K-means Clustering: Algorithm that partitions data into k distinct clusters.


Knowledge Distillation: Process of transferring knowledge from a large model to a smaller one.


Knowledge Graph: Network representing relationships between entities.


L

Large Language Model (LLM): A deep learning model trained on vast text datasets to generate human-like text based on statistical probabilities.


Latent Space: Compressed representation of data captured by models like VAEs.


Learning Rate: Hyperparameter controlling how much to adjust model weights during training.


LSTM (Long Short-Term Memory): Recurrent neural network architecture designed to remember long-term dependencies.


M

Model Context Protocol (MCP): MCP is aimed at enabling AI systems to better understand and adapt to the user’s intent and maintain continuity across conversations or interactions. It is particularly useful in applications like conversational AI, where maintaining context is crucial for providing coherent and relevant responses.


Machine Learning (ML): A subset of AI that enables computers to learn patterns from data and make predictions or decisions without explicit programming.


Metadata: Data that provides information about other data.


Metaverse: Collective virtual shared spaces, often incorporating AI technologies.


Multimodal AI: AI models that process and understand multiple types of data (e.g., text, images, and audio) simultaneously.


N

Natural Language Processing (NLP): A field of AI that enables machines to understand, interpret, and generate human language.


Natural Language Understanding (NLU): Subset of NLP focused on comprehension of language.


Neural Network: A computing system inspired by the human brain, consisting of interconnected layers of nodes (neurons) that process information.


Neuro-Symbolic AI: AI that combines neural networks (learning-based AI) with symbolic reasoning (rule-based AI) to improve explainability and logic-driven decision-making.


Normalization: Process of rescaling input data to improve model performance.


O

One-Hot Encoding: Representation of categorical variables as binary vectors.


Ontology: Formal naming and definition of categories, properties, and relations.


Overfitting: A modeling error in ML where the algorithm performs well on training data but poorly on new data due to excessive complexity.


P

Perceptron: Simplest type of artificial neural network.


Precision: Proportion of positive identifications that were actually correct.


Prediction & Sampling: The technique of generating outputs by estimating the most likely next elements based on probability distributions learned during training.


Prompt Engineering: Designing and refining inputs to guide AI systems toward desired outputs.


Pruning: Technique to reduce model size by removing unnecessary connections.


Q

Q-learning: Reinforcement learning algorithm to learn quality of actions.


Quantization: Technique to reduce model size by using fewer bits to represent weights.


Quantum AI: The application of quantum computing principles to enhance AI performance, particularly in solving complex optimization problems.


R

RAG (Retrieval-Augmented Generation): Technique combining information retrieval with text generation.

Recall: Proportion of actual positives that were correctly identified.


Recommendation System: System that suggests items or content to users based on their preferences.

Regression: Technique for predicting continuous values.


Reinforcement Learning (RL): A type of ML where an agent learns by interacting with an environment and receiving rewards or penalties based on its actions.


RNN (Recurrent Neural Network): Neural network architecture designed for sequential data.


Robotics Process Automation (RPA): The use of AI-driven bots to automate repetitive business processes and workflows.


S

Self-Attention Mechanism: A component of transformer models that determines how different parts of input data relate to each other, allowing for context-aware predictions.


Self-Improving AI: AI systems that can autonomously refine their models and adapt without explicit retraining, improving their performance over time.


Self-Supervised Learning: A form of ML where the model generates its own training labels from raw data, reducing the need for human-labeled datasets.


Semi-Supervised Learning: A hybrid ML approach that uses a small amount of labeled data combined with a large amount of unlabeled data.


Semantic Search: Search method based on meaning rather than literal matches.


Sentiment Analysis: Process of determining emotional tone behind text.


Stable Diffusion: Type of diffusion model for generating high-quality images.


Supervised Learning: A type of ML where the model is trained on labeled data, meaning the input data is paired with the correct output.


Support Vector Machine (SVM): Algorithm that finds the hyperplane that best divides a dataset.


Swarm Intelligence: A decentralized AI approach where multiple agents collaborate to solve complex problems, inspired by natural systems like ant colonies but also applied in robotics, traffic optimization, and financial modeling.


Synthetic Data: Artificially generated data used to train AI models when real-world data is scarce or sensitive.


T

TensorFlow: Open-source machine learning framework developed by Google.


Text-to-Image: AI systems that generate images based on text descriptions.


Tokenization & Embeddings: The process of breaking text into smaller units (tokens) and converting them into numerical representations in a multi-dimensional space.


Training Data: The dataset used to train an AI model, helping it learn and recognize patterns before making predictions.


Transfer Learning: The practice of applying knowledge learned from one task or domain to improve performance on a different but related task.


Transformer Model: A deep learning architecture, such as GPT and BERT, used in NLP for tasks like translation, summarization, and text generation.


Turing Test: Test of a machine's ability to exhibit intelligent behavior indistinguishable from a human.


U

Underfitting: When an ML model is too simple and fails to capture patterns in the training data, resulting in poor performance.


Unsupervised Learning: A type of ML where the model learns patterns and structures from unlabeled data without explicit instructions.


User Intent: The purpose or goal behind a user's query or action.


V

Validation Set: Subset of data used to tune hyperparameters.


VAE (Variational Autoencoder): Generative model that learns latent representations.


Vector Database: Database designed to store and query vector embeddings.


Vision Transformer (ViT): Transformer model adapted for computer vision tasks.


W

Weight (Neural Network): Parameter determining the strength of connection between neurons.


Word Embedding: Representation of words as vectors capturing semantic relationships.


Z

Zero-Shot Learning: An AI model's ability to make predictions about unseen data without direct prior training on that specific category.


 

2. Easily Confused Terms in AI


1. Bias (Statistical) vs. Bias (AI Ethics) vs. Bias in AI

These three related terms address different aspects of bias. Statistical bias relates to systematic errors in model predictions, AI Ethics bias refers to the broader moral implications, and Bias in AI combines both aspects.


2. Agentic AI vs. AI Agent vs. Autonomous Agents

The differences are meaningful but subtle:

  • An AI agent refers to any AI system that perceives its environment, makes decisions, and takes actions to achieve a goal. AI agents can range from simple rule-based systems to complex, autonomous learning models.

  • Agentic AI is a more advanced concept, referring to AI systems that demonstrate high levels of autonomy, adaptability, and goal-directed behavior. These systems are not just executing predefined tasks but can self-improve, plan, and act with minimal human intervention.

  • Autonomous Agents emphasizes real-time interaction and adaptation


Key Differences:

Feature

AI Agent

Agentic AI

Definition

Any AI system that perceives and acts to achieve a goal.

AI that exhibits high autonomy, adaptability, and long-term decision-making.

Autonomy Level

Varies (can be rule-based or learning-based).

High (capable of self-directed learning and decision-making).

Complexity

Can be simple (e.g., rule-based bots).

More advanced, often involving multi-step planning.

Human Oversight

Often requires supervision or pre-programmed logic.

Less human intervention; can generate and execute its own tasks.

Examples

Chatbots, recommendation systems, autonomous drones.

AI researchers, autonomous AI-driven businesses, advanced AI assistants.


4. Self-Supervised Learning vs. Unsupervised Learning

These learning paradigms sound similar but operate differently. Self-supervised learning generates its own labels from the data structure itself, while unsupervised learning identifies patterns without any labels. The "self" prefix suggests independence in both cases, making them easy to confuse.


5. XAI (Explainable AI) vs. Ethical AI

Both terms relate to responsible AI development but address different aspects. XAI focuses on making AI decisions interpretable and transparent, while Ethical AI encompasses broader moral considerations including fairness and societal impact.


6. Foundation Models vs. Large Language Models

Foundation Models is the broader category that includes LLMs but also encompasses other modalities like vision. People often use these terms interchangeably, especially when discussing text-based systems, which can lead to confusion about their scope.


7. Transfer Learning vs. Few-Shot Learning vs. Zero-Shot Learning

These terms represent a progression of related capabilities in AI systems, but the distinctions can be subtle. They all involve applying knowledge across contexts, but with varying amounts of example data.


8. Reinforcement Learning vs. Self-Improving AI

Both involve systems that get better over time, but through different mechanisms. Reinforcement learning uses explicit reward signals, while self-improving AI can encompass broader approaches to autonomous improvement.


9. Embeddings vs. Word Embeddings vs. Tokenization & Embeddings

These related terms describe different aspects of representing language in vector space. The overlap in terminology can make it difficult to understand where one concept ends and another begins.


10. Generative AI vs. Text-to-Image vs. Image Generation

Generative AI is the broader category, while the others are specific applications. The hierarchical relationship might not be clear from the current definitions.


11. Computer Vision vs. Image Recognition

Image recognition is a subset of computer vision, but the terms are sometimes used interchangeably, which can cause confusion about their scope.


12. Edge AI vs. IoT and AI

These terms have significant overlap since many IoT implementations leverage edge computing for AI processing. The connection between these concepts might not be evident from separate definitions.

To address these potential confusions, the glossary could benefit from:

  1. Cross-references between related terms

  2. Hierarchical organization that shows parent-child relationships between concepts

  3. Comparison tables for closely related term sets (like you provided for AI Agent vs. Agentic AI)

  4. Visual concept maps showing how different terms relate to each other



3. Comprehensive List of Generative AI Products


1. Text Generation and Conversational AI

OpenAI Products:

  • ChatGPT: Conversational AI assistant based on the GPT models, with web browsing and image recognition capabilities

  • GPT-4o: Latest multimodal language model with capabilities for text, image, and audio processing

  • GPT-4: Advanced large language model with enhanced reasoning capabilities

  • GPT-3.5 Turbo: Efficient version of GPT-3.5 optimized for chat applications

  • DALL-E 3: Text-to-image generation model integrated with ChatGPT

  • Sora: AI video generation system (announced but limited access)


Anthropic Products:

  • Claude 3.7 Sonnet: Current state-of-the-art conversational AI assistant

  • Claude 3.5 Sonnet: Conversational assistant optimized for reasoning

  • Claude 3.5 Haiku: Faster, more efficient version optimized for everyday tasks

  • Claude 3 Opus: Flagship model designed for complex reasoning and tasks

  • Claude Code: An agentic command line tool (in research preview)


Google Products:

  • Gemini: Family of multimodal AI models (formerly Bard)

  • Gemini Ultra: Google's most advanced AI model

  • Gemini Pro: Balanced model for various tasks

  • Gemini Nano: Lightweight model for on-device applications

  • PaLM 2: Large language model powering many Google products

  • MedPaLM: Specialized medical knowledge model

  • Med-PaLM 2: Enhanced medical AI model


Meta AI Products:

  • Llama 3: Open-source large language model

  • Llama 2: Previous version of Meta's open-source LLM

  • Llama 3.1: Enhanced version with improved capabilities

  • Meta AI Assistant: Consumer-facing chatbot available on WhatsApp, Instagram, etc.

  • Code Llama: Specialized model for programming tasks


Microsoft Products:

  • Copilot: Integrated AI assistant in Microsoft products (formerly Bing Chat)

  • Microsoft Copilot Studio: Tool for building custom copilots

  • Azure OpenAI Service: Cloud platform offering various AI models

  • Orca: Research model focused on reasoning


Other Text Generation Products:

  • Cohere Command: Enterprise-focused LLM

  • Mistral AI: Family of open-source language models

  • Mixtral 8x7B: Mixture-of-experts model from Mistral AI

  • Perplexity AI: AI search engine with generative responses

  • Character.AI: Platform for creating conversational AI characters

  • Pi: Personal AI assistant from Inflection AI

  • Claude.ai: Consumer-facing application for Anthropic's Claude

  • Khanmigo: Educational AI tutor from Khan Academy

  • Poe: Platform offering access to multiple AI models

  • Anthropic API: API access to Claude models

  • HuggingChat: Open-source chatbot by Hugging Face

  • LLaMA Chat: Meta's interface for Llama models


2. Open source

DeepSeek Products

Large Language Models:

  • DeepSeek LLM: Their flagship large language model series, with various sizes including 7B, 67B parameters

  • DeepSeek Coder: Specialized coding models designed specifically for programming tasks

  • DeepSeek-V2: Their second-generation language model with improved capabilities

  • DeepSeek-Math: Specialized model for mathematical reasoning


3. Image Generation

Text-to-Image Models:
  • Midjourney: AI image generation via Discord

  • DALL-E 3: OpenAI's advanced text-to-image model

  • DALL-E 2: Earlier version of OpenAI's image generator

  • Stable Diffusion: Open-source image generation model

  • Stable Diffusion XL: Enhanced version with improved quality

  • Stable Diffusion 3: Latest version with advanced capabilities

  • Adobe Firefly: Creative image generation integrated with Adobe products

  • Canva Text to Image: Integrated image generation in Canva

  • DreamStudio: Interface for Stability AI's models

  • Leonardo.AI: AI image generator for creative professionals

  • Ideogram: Text-to-image with strong typography handling

  • Imagen: Google's text-to-image model

  • Bing Image Creator: Microsoft's image generation tool

  • Craiyon (formerly DALL-E mini): Lightweight image generator


4. Image Editing/Enhancement:

  • Adobe Generative Fill: AI-powered content generation in Photoshop

  • Luminar Neo: AI photo editor with generative capabilities

  • Runway Gen-2: Image and video generation and editing

  • Lensa AI: Portrait enhancement and stylization

  • Topaz Photo AI: AI-powered photo enhancement


5. Video Generation:

  • Runway Gen-2: AI video generation and editing platform

  • Synthesia: AI video creation with virtual avatars

  • Pika Labs: Text-to-video generation

  • HeyGen: AI video generation platform

  • Sora: OpenAI's text-to-video model (limited access)

  • D-ID: AI-powered talking avatars from images

  • Descript: AI video and podcast editing tools

  • ElevenLabs: Voice cloning and video dubbing


6. Audio and Music Generation

  • ElevenLabs Voice Lab: AI voice generation and cloning

  • MUBERT: AI music generation

  • Suno: AI music creation from text prompts

  • Udio: AI music generation platform

  • SoundDraw: AI music composition tool

  • AudioCraft: Meta's audio generation model

  • MusicLM: Google's music generation system

  • Voicemod: Real-time voice modification

  • Play.ht: AI voice generation

  • Resemble.ai: Voice cloning technology

  • VALL-E: Microsoft's text-to-speech model


7. Code and Development

  • GitHub Copilot: AI pair programmer from GitHub/Microsoft

  • Replit Ghostwriter: AI coding assistant

  • Amazon CodeWhisperer: AWS coding assistant

  • Tabnine: AI code completion tool

  • Codeium: AI coding assistant

  • Cursor: AI-first code editor

  • DeepMind AlphaCode: AI system for competitive programming

  • Anthropic Claude Code: Code-focused AI tool


8. Specialized Industry Tools

Design and Creative:
  • Figma AI: AI features in the Figma design platform

  • Framer AI: Website building with AI

  • Jasper: AI content creation platform

  • Runway: Creative tools for video editing and generation


Business and Productivity:
  • Notion AI: Integrated AI in Notion workspace

  • Microsoft 365 Copilot: AI integrated across Microsoft Office

  • Google Workspace AI: AI features in Google's productivity suite

  • Anthropic Claude for Slack: Claude integration in Slack


Research and Education:
  • Elicit: AI research assistant

  • Consensus: AI-powered scientific search engine

  • ChatPDF: Conversational interface for PDFs

  • Khan Academy Khanmigo: Educational AI tutor


Customer Service and Sales:
  • HubSpot ChatSpot: CRM-integrated AI assistant

  • Intercom Fin: Customer service AI

  • Salesforce Einstein: AI for CRM and customer insights

  • Drift: Conversational marketing platform with AI


9. Multi-Modal Systems

  • GPT-4o: OpenAI's omni-modal system (text, image, audio)

  • Gemini: Google's multimodal model

  • Claude 3 models: Multimodal capabilities across text and images

  • Anthropic API: Multimodal capabilities through API

  • Fuyu-8B: Adept's open-source multimodal model

Primary_Logologonew[1].png
Green triangle with building corner_edit
bottom of page