AI Glossary

Every AI term you'll encounter, explained in plain language. No jargon required to understand the jargon.

Core AI Concepts

Software that can perform tasks that normally require human thinking, such as recognizing images, understanding language, making predictions, or generating text. The term covers everything from the autocomplete on your phone to systems that can write essays and hold conversations. AI doesn’t “think” the way you do. It finds patterns in enormous amounts of data and uses those patterns to produce useful outputs.

The method behind most modern AI. Instead of programming a computer with explicit rules ("if the email contains ‘Nigerian prince,’ mark it spam"), you feed it thousands of examples and let it figure out the rules on its own. The computer "learns" by adjusting itself until its predictions get more accurate. Think of it like teaching a child to recognize dogs: you don’t explain fur density and snout geometry, you just show them hundreds of dogs until they get it.

A more powerful form of machine learning that uses layered structures called neural networks. "Deep" refers to the many layers of processing the data passes through, not the depth of understanding. Deep learning is what made the current AI boom possible. It is the technique behind image recognition, voice assistants, and large language models.

A software architecture loosely inspired by how neurons in the human brain connect. It consists of layers of small processing units ("nodes") that each make simple decisions. When millions of these nodes work together across many layers, the network can recognize incredibly complex patterns. The "neural" part is a metaphor. There are no actual brain cells involved.

A set of step-by-step instructions a computer follows to solve a problem or complete a task. A recipe is an algorithm. Sorting your email inbox by date is an algorithm. In AI, algorithms are the mathematical procedures that guide how a system learns from data and makes decisions. When people say "the algorithm" on social media, they usually mean the recommendation system that decides what content you see.

The end product of training an AI system. When a machine learning system finishes learning from data, the patterns it discovered are saved as a "model." Think of the model as the finished student: the training is school, the data is the textbooks, and the model is the graduate who can now apply what it learned. When someone says "GPT-4 is a model," they mean it’s a trained system ready to do work.

The internal settings a model adjusts during training to get better at its task. A model with 70 billion parameters has 70 billion tiny dials it can tune. More parameters generally means the model can capture more nuance and complexity, but also requires more computing power and energy to run. When you see headlines about models having "hundreds of billions of parameters," this is what they’re counting.

The process of teaching an AI model by feeding it data and letting it adjust its parameters until it produces good results. Training a large model can take weeks or months using thousands of specialized computers running simultaneously. It’s enormously expensive, often millions of dollars for a single training run on the largest models.

What happens after training, when a model is actually being used. Every time you ask ChatGPT a question or Claude writes you an email, that’s inference. Training is learning; inference is doing. Training happens once (or periodically); inference happens every time someone uses the model.

AI systems that create new content (text, images, music, video, or code) rather than just analyzing or categorizing existing content. Large language models like Claude and ChatGPT are generative AI. So are image generators like Midjourney and DALL-E. The "generative" part means these systems produce something new each time, even if what they produce is built from patterns they learned during training.

Language Models & How They Work

An AI model trained on massive amounts of text (books, websites, articles, code, conversations) that can understand and generate human language. "Large" refers to both the enormous amount of training data and the billions of parameters inside the model. Claude, GPT-4, Gemini, and Llama are all LLMs. They don’t understand language the way you do; they are extremely sophisticated pattern-matching systems that predict what words should come next, given what’s already been said.

The broader field of teaching computers to work with human language: reading it, understanding it, translating it, summarizing it. NLP existed long before the current AI wave (think: spell-check, spam filters, Google Translate). Modern LLMs are the most advanced NLP systems ever built, but NLP as a field includes much simpler tools too.

The basic unit an LLM works with. A token isn’t always a whole word. It can be a word, part of a word, or even a punctuation mark. "Unbreakable" might be split into "un," "break," and "able," which is three tokens. Models process, generate, and charge by tokens. When a service says a model has a "200,000-token context window," it means the model can handle roughly 150,000 words of text at once.

The amount of text a model can "see" and work with at one time. Think of it as the model’s short-term memory. A small context window means the model can only consider a few pages of text; a large one means it can hold an entire book. If your conversation with an AI gets long enough to exceed the context window, the model starts "forgetting" the earliest parts of the conversation.

The text you type to communicate with an AI model: your question, instruction, or request. "Write me a cover letter for a marketing job" is a prompt. The quality of your prompt directly affects the quality of the response. This isn’t a flaw; it’s how the technology works. Learning to write clear, specific prompts is genuinely useful and doesn’t require technical skills.

Hidden instructions given to an AI model before you ever start talking to it. These are set by the company or developer and shape how the model behaves: its personality, safety rules, and what it will and won’t do. You typically don’t see the system prompt, but it’s the reason Claude sounds different from ChatGPT, even though both are LLMs.

When an AI model confidently states something that is factually wrong or entirely made up. The model isn’t "lying." It doesn’t have beliefs or intentions. It’s generating text that follows the patterns it learned, and sometimes those patterns produce plausible-sounding nonsense. This is one of the most important limitations to understand: AI can be fluent and wrong at the same time.

A setting that controls how creative or predictable a model’s responses are. Low temperature means the model picks the most likely next word every time (predictable, consistent). High temperature means it’s more willing to make unexpected choices (creative, varied, but also more likely to go off the rails). It’s like the difference between a jazz musician improvising and one reading sheet music.

Taking a pre-trained model and training it further on a smaller, more specialized dataset to make it better at a specific task. A general-purpose LLM might be fine-tuned on legal documents to create a legal AI assistant, or on medical texts to create a clinical support tool. Fine-tuning is much cheaper and faster than training a model from scratch.

A technique where human reviewers rate the model’s responses, and the model uses those ratings to improve. Think of it as a teacher grading essays and the student learning what "good" looks like. RLHF is a major reason modern AI assistants are helpful and conversational rather than producing raw, unfiltered text. It’s also how safety behaviors are instilled.

A technique that lets an AI model look up real information from a database or document collection before responding, rather than relying solely on what it memorized during training. Think of it as the difference between answering a test from memory versus being allowed to check your notes. RAG helps reduce hallucinations and keeps responses grounded in actual sources.

A way of representing words, sentences, or documents as lists of numbers so that similar meanings end up close together mathematically. "Happy" and "joyful" would have similar embeddings; "happy" and "carburetor" would not. Embeddings are how AI systems understand that concepts are related even when the exact words are different. They power search, recommendations, and clustering.

An AI system that can work with more than one type of input, for example text and images, or text and audio. A multimodal model might be able to look at a photo of your fridge and suggest recipes, or listen to a voice memo and summarize it. Most cutting-edge models are moving in this direction.

A program designed to have conversations with people through text. The term predates modern AI by decades. Early chatbots followed simple scripts. Today’s AI chatbots, powered by LLMs, can hold nuanced, context-aware conversations. Claude, ChatGPT, and Gemini are all chatbots, though their capabilities extend far beyond just "chatting."

An AI system that can take actions on its own, not just answer questions, but actually do things like browse the web, write and run code, manage files, or interact with other software. An agent can break a task into steps, execute those steps, evaluate the results, and adjust its approach. This is a rapidly evolving area and represents a shift from AI as a "tool you ask questions" to AI as a "coworker that does tasks."

The broader concept of AI systems that operate with some degree of autonomy, planning, executing, and adapting without needing a human to guide every step. Agentic AI is the umbrella term for agents, autonomous workflows, and self-directed AI systems. It’s one of the most actively discussed areas in AI development right now.

The AI Industry & Key Players

The company behind ChatGPT and the GPT series of models. Founded as a nonprofit in 2015, it shifted to a "capped profit" structure and has become one of the most prominent AI companies in the world. Its products are among the most widely used consumer AI tools.

The company behind Claude (the AI you may be reading this on). Founded in 2021 by former OpenAI researchers, Anthropic focuses on AI safety: building AI systems that are helpful, harmless, and honest. Their research emphasizes understanding how AI models work internally and making them more reliable and trustworthy.

Google’s AI research lab, formed by merging Google Brain and DeepMind. They build the Gemini family of models and integrate AI across Google’s products (Search, Gmail, Docs, Android, and more). DeepMind is also known for foundational AI research breakthroughs like AlphaFold, which predicted the structure of nearly every known protein.

Meta’s (formerly Facebook) AI research division. They develop the Llama family of models, which are notable for being open-source, meaning anyone can download and use them. Meta’s approach of releasing models publicly has significantly shaped the AI ecosystem and made advanced AI accessible to smaller companies and researchers.

AI models whose underlying code and parameters are publicly released so anyone can inspect, use, modify, and build on them. Llama (Meta) and Mistral are prominent examples. Open-source models give smaller companies and independent developers access to powerful AI without building their own from scratch. The debate over whether to open-source AI models, balancing accessibility against misuse, is one of the biggest ongoing conversations in the field.

The most capable AI models currently in existence, the cutting edge. When people say "frontier models," they typically mean the latest releases from companies like Anthropic, OpenAI, and Google. These models define the upper boundary of what AI can do right now. The term implies both capability and responsibility, since frontier models are the first to encounter new challenges and risks.

A large AI model trained on broad data that can be adapted for many different tasks. Think of it as a generalist that can be specialized. GPT-4, Claude, and Gemini are foundation models. They weren’t trained to do one specific thing but can be applied to writing, coding, analysis, translation, and more. The "foundation" metaphor suggests these models are the base layer that other products and applications are built on top of.

A way for software programs to talk to each other. In the AI context, an API lets developers plug an AI model into their own apps, websites, or tools. When a startup says they’re "built on Claude" or "powered by GPT-4," they’re using those models through an API. You don’t interact with the API directly. The products you use do it behind the scenes.

A standardized test used to measure how well an AI model performs. Just like SAT scores let you compare students, benchmarks let you compare models. Common benchmarks test things like math ability, coding skill, reading comprehension, and reasoning. Benchmarks are useful but imperfect. A model can score well on a benchmark without being great at real-world tasks, similar to how being a good test-taker doesn’t always make you good at the job.

Data Centers, Computing & Energy

A building (or complex of buildings) packed with thousands of powerful computers. Data centers are where AI models are trained and run. They require enormous amounts of electricity, industrial cooling systems, and high-speed internet connections. When you send a message to Claude, your text travels to a data center, gets processed on specialized hardware, and the response travels back, usually in seconds.

Using someone else’s computers over the internet instead of running software on your own machine. When you use Google Docs, Netflix, or an AI assistant, the heavy lifting happens on servers in a data center somewhere, and you see the results on your screen. Amazon Web Services (AWS), Microsoft Azure, and Google Cloud are the three biggest cloud providers. Most AI companies rent computing power from these providers rather than building their own data centers.

A computer chip originally designed to render video game graphics. It turns out the same kind of math that draws 3D graphics (processing many calculations simultaneously) is exactly what AI training needs. GPUs became the engine of the AI revolution. NVIDIA is the dominant GPU manufacturer for AI workloads, and their chips are in such high demand that supply shortages have been a real constraint on AI development.

Google’s custom-designed chip built specifically for AI workloads. While GPUs were repurposed from gaming, TPUs were designed from the ground up for the kind of math AI models require. They’re used internally at Google to train and run Gemini and other models. Think of it as the difference between converting a truck into an ambulance versus building an ambulance from scratch.

A catch-all term for the computational power used to train and run AI models. "We need more compute" means "we need more processing power." Compute is one of the three fundamental ingredients of modern AI, alongside data and algorithms. Access to compute is a major competitive advantage. The companies with the most computing resources can train the biggest models.

Any specialized processor designed or optimized for AI tasks. GPUs, TPUs, and newer designs from companies like AMD, Intel, and startups are all AI chips. The global race to design faster, more efficient AI chips is reshaping the semiconductor industry. These chips are so strategically important that governments have imposed export restrictions on them.

Data centers generate enormous heat. All those thousands of processors running 24/7 need to be kept cool or they’ll overheat and fail. Traditional data centers use industrial air conditioning. Newer AI-focused data centers are exploring liquid cooling (running coolant directly over chips) and even building facilities near cold climates or bodies of water. Cooling is one of the biggest operational costs and engineering challenges in running AI at scale.

Training a single large AI model can consume as much electricity as a small town uses in a year. Running that model for millions of users every day consumes even more. The energy demands of AI have become a serious concern, driving conversations about sustainability, renewable energy commitments, and whether the benefits of AI justify its environmental footprint. Major AI companies have signed deals for nuclear, solar, and wind power specifically to run their data centers.

A contract where a company agrees to buy electricity directly from a power generator, often a renewable energy project, at a set price for a set period (usually 10–20 years). AI companies use PPAs to secure massive amounts of clean energy for their data centers. When you read that Microsoft or Google signed a deal with a nuclear plant, that’s a PPA.

The total greenhouse gas emissions associated with building, training, and running AI systems. This includes the electricity used by data centers, the manufacturing of chips and hardware, the cooling systems, and even the construction of the buildings. Estimates vary widely, but the AI industry’s energy consumption is growing faster than almost any other sector. Understanding this footprint is part of making informed decisions about AI use.

Processing data on or near the device itself rather than sending it to a distant data center. When your phone uses face recognition to unlock without contacting a server, that’s edge computing. It’s faster (no round trip to the cloud), can work offline, and can be more private (your data never leaves your device). Many AI applications are moving toward edge computing for these reasons.

The delay between when you send a request and when you get a response. When you ask Claude a question and there’s a pause before the answer appears, that’s latency. It’s affected by the distance to the data center, how busy the servers are, and how complex the request is. Lower latency means a snappier, more conversational experience.

Privacy, Security & Your Data

Your right to control how your personal information is collected, used, stored, and shared. In the AI context, this includes what happens to the conversations you have with AI assistants, whether your data is used to train future models, and whether companies can identify you from your usage patterns. Data privacy isn’t just a technical issue. It’s a legal, ethical, and personal one.

The information used to teach an AI model during training. For LLMs, training data typically includes text from books, websites, articles, public forums, code repositories, and more. A key question: was the training data collected ethically? Did the people who wrote that content consent to it being used this way? These questions are at the center of ongoing lawsuits and policy debates.

The process of gathering information, often about you. Every time you use an app, visit a website, or interact with an AI, data is potentially being collected: what you typed, how long you spent, what you clicked, where you are, what device you’re using. Not all data collection is harmful, but understanding that it’s happening is the first step to making informed choices about the tools you use.

A European Union law enacted in 2018 that sets strict rules for how companies collect, store, and use personal data. It gives EU citizens specific rights: the right to see what data a company has on you, the right to have your data deleted, and the right to object to certain uses of your data. Even if you’re not in the EU, GDPR matters. Many global companies apply its standards worldwide because it’s easier than maintaining different rules for different regions.

A California law that gives residents more control over their personal data, similar in spirit to GDPR but with some different specifics. It gives Californians the right to know what data is being collected about them, the right to delete it, and the right to opt out of its sale. Because so many tech companies are based in or do business in California, CCPA’s effects ripple far beyond the state.

The legal right to have your personal data deleted from a company’s systems. Established under GDPR and adopted in various forms elsewhere. In the AI context, this gets complicated: if your data was used to train a model, "deleting" that data doesn’t necessarily undo its influence on the model. This is an unresolved tension between traditional privacy rights and how AI systems actually work.

Removing or scrambling identifying information from data so that it can’t be traced back to a specific person. For example, replacing names with random IDs or removing zip codes. In theory, anonymized data protects privacy while still being useful for analysis. In practice, research has repeatedly shown that supposedly anonymized data can often be re-identified, especially when combined with other data sources.

Scrambling data so that only authorized people can read it. When your messaging app says it uses "end-to-end encryption," it means your messages are scrambled on your device and only unscrambled on the recipient’s device. Not even the company running the app can read them in between. In the AI context, encryption protects your data in transit (while it’s being sent to a server) and at rest (while it’s stored). It doesn’t protect your data while it’s being actively processed by the model, which is a current technical limitation.

Two approaches to data collection. Opt-in means a company can’t collect or use your data unless you actively agree. Opt-out means the company collects your data by default and you have to take action to stop it. Privacy advocates strongly prefer opt-in because it puts the choice in your hands from the start. Many AI services default to opt-out for training data, meaning your conversations may be used to improve the model unless you go into settings and turn that off.

The legal agreement you accept when you sign up for a product or service. It defines what the company can and can’t do with your data, what you’re allowed and not allowed to do with their product, and what happens if things go wrong. These documents are notoriously long and difficult to read. That said, the sections about data usage, data sharing, and what happens to your content are worth understanding, especially for AI tools where you might be sharing sensitive information.

How long a company keeps your data after you provide it. Some services delete your conversation history immediately; others keep it for 30 days, a year, or indefinitely. Data retention policies vary widely between AI providers and even between different tiers of the same product. Shorter retention generally means less risk to your privacy. This is worth checking in the settings of any AI tool you use regularly.

Any data that can be used to identify a specific person: your name, email address, phone number, Social Security number, home address, IP address, biometric data, and so on. AI companies have varying policies on how they handle PII that appears in conversations. Some automatically strip it; others log it. Understanding what counts as PII helps you make smarter decisions about what you share with AI tools.

A mathematical technique that lets companies analyze patterns in data without exposing any individual’s information. Imagine a survey about salary. Differential privacy adds carefully calculated noise to the data so you can learn the average salary without being able to figure out what any one person earns. Apple and Google have used this technique in their products. It’s one of the more promising approaches to balancing useful data analysis with genuine privacy protection.

A technique where an AI model is trained across many devices without the data ever leaving those devices. Instead of sending your data to a central server, the model comes to your data, learns a little bit, and only sends back what it learned (not your actual data). Your phone’s keyboard autocomplete often works this way. It’s a privacy-preserving approach, though it comes with its own set of technical challenges.

AI Safety, Ethics & Governance

The field focused on making sure AI systems do what we want them to do, and don’t do things we don’t want them to do. This ranges from immediate concerns (making sure a chatbot doesn’t give dangerous medical advice) to long-term concerns (ensuring very powerful future AI systems remain under human control). AI safety isn’t about being afraid of robots; it’s about building reliable, trustworthy technology.

The challenge of making sure an AI system’s goals and behaviors match human intentions and values. A model that’s "aligned" does what you actually want, not a technically correct but unhelpful literal interpretation of your request. Alignment is harder than it sounds because human values are complex, contextual, and sometimes contradictory. It’s one of the central technical and philosophical challenges in AI development.

When an AI system produces unfair or skewed results because the data it was trained on reflected existing human prejudices. If a hiring model was trained mostly on data from a male-dominated industry, it might learn to favor male candidates, not because anyone told it to, but because that’s the pattern in the data. AI bias isn’t a bug that can be fixed with a single patch; it requires ongoing attention to training data, evaluation methods, and real-world outcomes.

Limits and rules built into an AI system to prevent harmful, inappropriate, or off-topic outputs. When a chatbot refuses to help you make a weapon or tells you it can’t provide medical diagnoses, that’s a guardrail at work. Guardrails are implemented through a combination of system prompts, RLHF, and content filtering. The challenge is setting guardrails that prevent genuine harm without making the system so restricted it’s unhelpful.

The practice of deliberately trying to make an AI system fail, behave badly, or produce harmful outputs. Essentially, hiring people to attack your own product before bad actors do. Red teams try to bypass guardrails, extract dangerous information, or trick the model into producing biased or toxic content. It’s a critical part of safety testing before releasing a model to the public.

A broad term for the practice of developing and deploying AI in ways that are ethical, fair, transparent, and accountable. It encompasses bias mitigation, safety testing, privacy protection, environmental impact, and honest communication about what AI can and can’t do. Most major AI companies have published responsible AI principles; the degree to which they follow through varies.

The ability to understand how and why an AI model produced a specific output. Current large models are largely "black boxes." They produce results, but even the people who built them can’t always explain exactly why the model said what it said. Explainability research aims to open that black box, which is important for trust, debugging, accountability, and regulatory compliance.

Laws and rules governing how AI can be developed and used. The EU’s AI Act is the most comprehensive AI law to date. The U.S. has taken a more sector-specific approach with executive orders and agency guidelines. China has its own set of AI regulations. The regulatory landscape is evolving rapidly and varies significantly by country.

The European Union’s landmark regulation classifying AI systems by risk level, from minimal risk (spam filters) to unacceptable risk (social scoring by governments). Higher-risk AI systems face stricter requirements: transparency obligations, human oversight mandates, data quality standards, and more. The Act began taking effect in stages starting in 2024 and will influence AI regulation worldwide.

AI-generated synthetic media, usually a video or audio recording that makes it look or sound like a real person is saying or doing something they never actually said or did. Deepfakes range from harmless entertainment to serious threats: political disinformation, fraud, and non-consensual intimate imagery. Detection tools exist but lag behind generation capabilities. When you see a surprising video online, questioning whether it’s real is now a reasonable first instinct.

False or misleading information. Misinformation is spread unintentionally, when someone shares something they genuinely believe is true. Disinformation is spread deliberately to deceive. AI makes both worse: generative AI can produce convincing but false text, images, and video at scale. At the same time, AI tools can potentially help detect and combat false content. The net effect is an information environment where critical thinking matters more than ever.

The theoretical risk that a sufficiently advanced AI system could pose a fundamental threat to human civilization. This is the most speculative and debated area of AI safety. Some researchers consider it a serious concern that deserves proactive attention; others consider it a distraction from more immediate, concrete AI harms like bias, job displacement, and surveillance. Reasonable people disagree. What matters is that you’ve heard both sides.

AI in Everyday Life

The AI system behind "you might also like" on Netflix, Spotify, Amazon, YouTube, TikTok, and virtually every other platform you use. These algorithms analyze your behavior (what you watch, click, skip, linger on, purchase) and predict what you’ll engage with next. They’re extraordinarily effective at keeping you on a platform, which is both their value and the core concern about them.

AI assistants you can talk to directly: Claude, ChatGPT, Gemini, Copilot. These are the most visible AI products right now. You type a question or request, and the AI responds. They can write, summarize, brainstorm, explain, tutor, translate, and code. Their limitations: they can be confidently wrong (hallucinations), they don’t have real-time knowledge unless connected to search, and they don’t truly "understand" you. They respond to patterns.

Search engines that use generative AI to produce direct answers instead of (or alongside) a list of links. Google’s AI Overviews and Perplexity are examples. Instead of "here are ten web pages that might answer your question," AI-powered search says "here’s the answer, synthesized from multiple sources." The trade-off: it’s faster and more convenient, but harder to evaluate where the information came from and whether it’s accurate.

AI features built directly into the software you already use, like AI in Microsoft Word that helps you write, AI in Gmail that drafts replies, or AI in Photoshop that edits images from text descriptions. These tools aim to make you more productive within your existing workflow rather than requiring you to switch to a separate AI app.

The feature on your phone or in your email that suggests the next word or phrase as you type. This is one of the oldest and most familiar forms of AI in your daily life. Modern predictive text is powered by small language models running on your device. It’s the simplest version of what a full LLM does: predicting what comes next.

AI-powered tools like Siri, Alexa, and Google Assistant that you interact with by speaking. They use speech recognition to understand your words, natural language processing to interpret your intent, and text-to-speech to respond. The latest generation of voice assistants is being upgraded with LLM capabilities, making them significantly more conversational and capable than their earlier versions.

AI is being used to analyze medical images (like X-rays and MRIs), predict patient outcomes, accelerate drug discovery, power mental health chatbots, and assist with medical documentation. AI is not replacing doctors. It’s being used as a tool to support clinical decisions. The regulatory environment for medical AI is strict and evolving. If an AI tool gives you health information, it should always include a reminder to consult an actual healthcare professional.

AI tools used for tutoring, generating practice problems, providing feedback on writing, creating lesson plans, and personalizing learning. Students use AI for homework help (raising concerns about academic integrity). Teachers use AI to save time on administrative tasks. The impact on education is still unfolding, and every school and university is grappling with where to draw the line.

The Business & Economics of AI

Software you access online through a subscription instead of buying and installing on your computer. Most AI tools you interact with (Claude, ChatGPT, Midjourney) follow this model. You pay monthly or yearly for access. The AI runs on the company’s servers; you use it through a website or app.

Companies that provide AI capabilities through an API so other businesses can add AI to their own products without building models from scratch. Instead of training your own LLM (which would cost millions), you pay Anthropic, OpenAI, or Google per usage to access theirs. This is how most AI-powered apps work behind the scenes.

The observation that AI models generally get more capable when you increase three things: the amount of training data, the number of parameters, and the amount of compute used for training. Scaling laws have driven the "bigger is better" approach in AI development. Whether these laws will continue to hold, or whether we’ll hit diminishing returns, is one of the most important open questions in the field.

A hypothetical AI system that can perform any intellectual task a human can, with the same flexibility and generality. Current AI is "narrow." Claude is great at language tasks but can’t ride a bike or feel emotions. AGI would be a system with human-level versatility. Whether AGI is five years away, fifty years away, or fundamentally impossible depends on who you ask. It’s the north star (or the bogeyman, depending on your perspective) of AI research.

The concern that AI will eliminate certain jobs or significantly reduce the number of humans needed to do them. History shows that major technologies do displace specific jobs while creating new ones, but the transition can be painful and unevenly distributed. AI is most likely to augment many jobs (making workers more productive), fully automate some jobs (particularly routine, repetitive tasks), and create new jobs that don’t exist yet. The honest answer is that nobody knows exactly how this will play out.

Legal rights over creative works and inventions. AI has created a tangle of IP questions: Can AI-generated content be copyrighted? If an AI model was trained on copyrighted books and artwork, do the original creators deserve compensation? Multiple lawsuits are working through these questions right now. The legal framework hasn’t caught up with the technology, and the outcomes will shape the AI industry for decades.

Technical Infrastructure Terms

The foundational material (usually silicon) used to make computer chips. Every AI chip, every GPU, every processor in every device you own is built on semiconductors. The global semiconductor supply chain, concentrated heavily in Taiwan and South Korea, has become a major geopolitical concern. When you hear about "chip wars" or export controls on advanced chips, semiconductors are what’s at stake.

The company whose GPUs power the vast majority of AI training and inference worldwide. NVIDIA went from being primarily a gaming hardware company to one of the most valuable companies on Earth because their chips turned out to be ideal for AI workloads. Their dominance in the AI chip market gives them enormous influence over the pace and direction of AI development.

A group of computers linked together to work as a single system. AI training requires thousands of GPUs working in concert, organized into clusters. Building and managing these clusters is a major engineering challenge. When you read that a company has a "10,000-GPU cluster," it means they’ve connected 10,000 graphics processors into one massive system dedicated to AI work.

How much data can be transferred over a network connection in a given amount of time. High bandwidth means more data can flow faster. AI workloads require enormous bandwidth, both between the GPUs in a cluster (which need to constantly share data during training) and between users and data centers (to deliver fast responses). It’s like the width of a highway: more lanes means more traffic can flow.

The percentage of time a system is operational and available. A data center boasting "99.99% uptime" means it’s down for less than about 53 minutes per year. For AI services that millions of people depend on, even brief downtime is costly and disruptive. Maintaining high uptime requires redundant power supplies, backup systems, and around-the-clock monitoring.

Having backup systems in place so that if one component fails, another takes over immediately. Data centers build redundancy into everything: power supplies, cooling systems, network connections, and the servers themselves. It’s the same principle as having a spare tire in your car. You hope you never need it, but you can’t afford not to have it.

The amount of work a system can handle in a given time period. In AI, this often means how many requests (questions, prompts, queries) a model can process per second. High throughput means the system can serve more users simultaneously. Companies optimize throughput to keep costs down and response times fast, especially during peak usage.

Emerging & Advanced Concepts

Data generated by AI models rather than collected from the real world. If you need a million examples to train a model but only have a thousand real ones, you might use AI to generate the rest. Synthetic data can help with privacy (no real people’s data involved) and availability (you can create data for rare scenarios). The risk is that if synthetic data has flaws, the model trained on it inherits those flaws.

The idea that knowledge gained from one task can be applied to another. A model trained to recognize objects in photos can be adapted to recognize medical conditions in X-rays, because the low-level visual skills transfer. Transfer learning is one of the reasons foundation models are so valuable. The general capabilities they learn during pre-training provide a powerful starting point for many specific applications.

A model’s ability to perform a task it was never explicitly trained for. Zero-shot means the model can do it with no examples; you just describe what you want. Few-shot means you give it a handful of examples and it picks up the pattern. The fact that modern LLMs can do this at all is one of their most remarkable and surprising capabilities. It’s why you can ask Claude to do things it was never specifically taught to do, and it often succeeds.

A distinction that matters more than it sounds. Open weights means the trained model’s parameters are publicly available; you can download and run it. Open source traditionally means the full package: code, data, training process, and documentation are all available. Some models marketed as "open source" are really just open weights. The difference matters because without the full picture, you can use the model but can’t fully understand or reproduce how it was built.

An architecture where a model contains multiple specialized sub-networks ("experts"), and for any given input, only a few relevant experts activate. Think of a hospital where you see a specialist based on your symptoms rather than every doctor examining every patient. MoE models can be very large in total parameter count but efficient to run because only a fraction of the model is working at any time.

A technique for making AI models smaller and faster by reducing the precision of their internal numbers. Instead of using highly precise decimal numbers (like 3.14159265), quantized models use rougher approximations (like 3.14). The model gets slightly less accurate but dramatically more efficient, often small enough to run on a phone or laptop instead of a server farm. It’s one of the key techniques making AI more accessible on everyday devices.

AI that runs directly on your phone, laptop, or other personal device rather than on a remote server. Apple Intelligence, Google’s on-device features, and various phone-specific AI features all fall into this category. The advantages: it’s faster (no internet round trip), works offline, and keeps your data on your device. The limitation: your phone is far less powerful than a data center, so on-device models are necessarily smaller and less capable.

A virtual replica of a physical object, system, or process. Factories create digital twins of their production lines to test changes virtually before implementing them physically. Cities create digital twins to model traffic flow. AI is increasingly used to power these simulations, making them more accurate and responsive. You probably won’t interact with a digital twin directly, but they’re quietly shaping how infrastructure, manufacturing, and logistics are designed.

Jargon You’ll See in Headlines

A marketing term applied to everything from toothbrushes to tax software. Sometimes it means a product genuinely uses sophisticated AI. Sometimes it means a basic algorithm or a simple automation. Treat it the same way you’d treat "military-grade" or "clinically proven." It might be meaningful, it might be marketing. Ask what specifically the AI does before being impressed.

When a new technology fundamentally changes how an industry works, often displacing established players. AI is genuinely disruptive in some areas (customer service, content creation, software development) and overhyped in others. When someone says AI will "disrupt" an industry, they might be right, but the timeline and extent are almost always uncertain.

A hypothetical AI system that surpasses the best human intelligence at essentially everything: science, creativity, social skills, general wisdom. No such system exists or is close to existing. It’s a concept from AI safety research and science fiction that occasionally makes its way into headlines. File it under "worth knowing about but not worth losing sleep over right now."

Business jargon for a competitive advantage that’s hard for others to replicate. "Does OpenAI have a moat?" is asking whether competitors can catch up to their technology and market position. In AI, moats include proprietary data, computing resources, talent, brand recognition, and distribution. The AI industry is evolving so fast that today’s moat can become tomorrow’s puddle.

Making a system handle more work. Scaling an AI service means serving more users, processing more requests, or handling more complex tasks without the system slowing down or breaking. "This won’t scale" is tech-speak for "this works fine for 100 people but will fall apart for a million." Scaling laws (defined above) refer to a related but distinct concept about model capability.

The idea that making AI safe and aligned with human values has a cost. The model might become less capable, slower, or more expensive. Safety features take resources to develop and can constrain what the model does. The term acknowledges a real tension: safety and capability can sometimes pull in opposite directions. How to balance them is an ongoing debate.

This glossary is a living document. Language evolves, technology advances, and new concepts emerge constantly. If a term you encountered isn't here, that's an invitation, not a failing.

Last updated: March 2026