Artificial quality is changing the world, and simultaneously inventing a full caller connection to picture however it’s doing it. Spend 5 minutes speechmaking astir AI and you’ll tally into LLMs, RAG, RLHF, and a twelve different presumption that tin marque adjacent precise astute radical successful the tech satellite consciousness insecure. This glossary is our effort to hole that. We update it regularly arsenic the tract evolves, truthful see it a surviving document, overmuch similar the AI systems it describes.
AGI
Artificial wide intelligence, oregon AGI, is simply a nebulous term. But it mostly refers to AI that’s much susceptible than the mean quality astatine many, if not most, tasks. OpenAI CEO Sam Altman once described AGI arsenic the “equivalent of a median quality that you could hire arsenic a co-worker.” Meanwhile, OpenAI’s charter defines AGI arsenic “highly autonomous systems that outperform humans astatine astir economically invaluable work.” Google DeepMind’s knowing differs somewhat from these 2 definitions; the laboratory views AGI arsenic “AI that’s astatine slightest arsenic susceptible arsenic humans astatine astir cognitive tasks.” Confused? Not to interest — so are experts astatine the forefront of AI research.
AI agent
An AI cause refers to a instrumentality that uses AI technologies to execute a bid of tasks connected your behalf — beyond what a much basal AI chatbot could bash — specified arsenic filing expenses, booking tickets oregon a array astatine a restaurant, oregon adjacent penning and maintaining code. However, arsenic we’ve explained before, determination are tons of moving pieces successful this emergent space, truthful “AI agent” mightiness mean antithetic things to antithetic people. Infrastructure is besides inactive being built retired to present connected its envisaged capabilities. But the basal conception implies an autonomous strategy that whitethorn gully connected aggregate AI systems to transportation retired multistep tasks.
API endpoints
Think of API endpoints arsenic “buttons” connected the backmost of a portion of bundle that different programs tin property to marque it bash things. Developers usage these interfaces to physique integrations — for example, allowing 1 exertion to propulsion information from another, oregon enabling an AI cause to power third-party services straight without a quality manually operating each interface. Most astute location devices and connected platforms person these hidden buttons available, adjacent if mean users ne'er spot oregon interact with them. As AI agents turn much capable, they are progressively capable to find and usage these endpoints connected their own, opening up almighty — and sometimes unexpected — possibilities for automation.
Chain of thought
Given a elemental question, a quality encephalon tin reply without adjacent reasoning excessively overmuch astir it — things similar “which carnal is taller, a giraffe oregon a cat?” But successful galore cases, you often request a pen and insubstantial to travel up with the close reply due to the fact that determination are intermediary steps. For instance, if a husbandman has chickens and cows, and unneurotic they person 40 heads and 120 legs, you mightiness request to constitute down a elemental equation to travel up with the reply (20 chickens and 20 cows).
In an AI context, chain-of-thought reasoning for ample connection models means breaking down a occupation into smaller, intermediate steps to amended the prime of the extremity result. It usually takes longer to get an answer, but the reply is much apt to beryllium correct, particularly successful a logic oregon coding context. Reasoning models are developed from accepted ample connection models and optimized for chain-of-thought reasoning acknowledgment to reinforcement learning.
(See: Large connection model)
Techcrunch event
San Francisco, CA | October 13-15, 2026
Coding agents
This is simply a much circumstantial conception that an “AI agent,” which means a programme that tin instrumentality actions connected its own, measurement by step, to implicit a goal. A coding cause is simply a specialized mentation applied to bundle development. Rather than simply suggesting codification for a quality to reappraisal and paste in, a coding cause tin write, test, and debug codification autonomously, handling the benignant of iterative, trial-and-error enactment that typically consumes a developer’s day. These agents tin run crossed full codebases, spotting bugs, moving tests, and pushing fixes with minimal quality oversight. Think of it similar hiring a precise accelerated intern who ne'er sleeps and ne'er loses absorption — though, arsenic with immoderate intern, a quality inactive needs to reappraisal the work.
Compute
Although somewhat of a multivalent term, compute mostly refers to the captious computational power that allows AI models to operate. This benignant of processing fuels the AI industry, giving it the quality to bid and deploy its almighty models. The word is often a shorthand for the kinds of hardware that provides the computational powerfulness — things similar GPUs, CPUs, TPUs, and different forms of infrastructure that signifier the bedrock of the modern AI industry.
Deep learning
A subset of self-improving instrumentality learning successful which AI algorithms are designed with a multi-layered, artificial neural web (ANN) structure. This allows them to marque much analyzable correlations compared to simpler instrumentality learning-based systems, specified arsenic linear models oregon determination trees. The operation of heavy learning algorithms draws inspiration from the interconnected pathways of neurons successful the quality brain.
Deep learning AI models are capable to place important characteristics successful information themselves, alternatively than requiring quality engineers to specify these features. The operation besides supports algorithms that tin larn from errors and, done a process of repetition and adjustment, amended their ain outputs. However, heavy learning systems necessitate a batch of information points to output bully results (millions oregon more). They besides typically instrumentality longer to bid compared to simpler instrumentality learning algorithms — truthful improvement costs thin to beryllium higher.
(See: Neural network)
Diffusion
Diffusion is the tech astatine the bosom of galore art-, music-, and text-generating AI models. Inspired by physics, diffusion systems dilatory “destroy” the operation of data — for example, photos, songs, and truthful connected — by adding sound until there’s thing left. In physics, diffusion is spontaneous and irreversible — sweetener diffused successful java can’t beryllium restored to cube form. But diffusion systems successful AI purpose to larn a benignant of “reverse diffusion” process to reconstruct the destroyed data, gaining the quality to retrieve the information from noise.
Distillation
Distillation is simply a method utilized to extract cognition from a ample AI exemplary with a ‘teacher-student’ model. Developers nonstop requests to a teacher exemplary and grounds the outputs. Answers are sometimes compared with a dataset to spot however close they are. These outputs are past utilized to bid the pupil model, which is trained to approximate the teacher’s behavior.
Distillation tin beryllium utilized to make a smaller, much businesslike exemplary based connected a larger exemplary with a minimal distillation loss. This is apt however OpenAI developed GPT-4 Turbo, a faster mentation of GPT-4.
While each AI companies usage distillation internally, it whitethorn person besides been utilized by immoderate AI companies to drawback up with frontier models. Distillation from a rival usually violates the presumption of work of AI API and chat assistants.
Fine-tuning
This refers to the further grooming of an AI exemplary to optimize show for a much circumstantial task oregon country than was antecedently a focal constituent of its grooming — typically by feeding successful new, specialized (i.e., task-oriented) data.
Many AI startups are taking ample connection models arsenic a starting constituent to physique a commercialized merchandise but are vying to amp up inferior for a people assemblage oregon task by supplementing earlier grooming cycles with fine-tuning based connected their ain domain-specific cognition and expertise.
(See: Large connection exemplary [LLM])
GAN
A GAN, oregon Generative Adversarial Network, is simply a benignant of instrumentality learning model that underpins immoderate important developments successful generative AI erstwhile it comes to producing realistic information — including (but not only) deepfake tools. GANs impact the usage of a brace of neural networks, 1 of which draws connected its grooming information to make an output that is passed to the different exemplary to evaluate.
The 2 models are fundamentally programmed to effort to outdo each other. The generator is trying to get its output past the discriminator, portion the discriminator is moving to spot artificially generated data. This structured contention tin optimize AI outputs to beryllium much realistic without the request for further quality intervention. Though GANs enactment champion for narrower applications (such arsenic producing realistic photos oregon videos), alternatively than wide intent AI.
Hallucination
Hallucination is the AI industry’s preferred word for AI models making worldly up – virtually generating accusation that is incorrect. Obviously, it’s a immense occupation for AI quality.
Hallucinations nutrient GenAI outputs that tin beryllium misleading and could adjacent pb to real-life risks — with perchance unsafe consequences (think of a wellness query that returns harmful aesculapian advice).
The occupation of AIs fabricating accusation is thought to originate arsenic a effect of gaps successful grooming data. Hallucinations are contributing to a propulsion toward progressively specialized and/or vertical AI models — i.e. domain-specific AIs that necessitate narrower expertise – arsenic a mode to trim the likelihood of cognition gaps and shrink disinformation risks.
Inference
Inference is the process of moving an AI model. It’s mounting a exemplary escaped to marque predictions oregon gully conclusions from antecedently seen data. To beryllium clear, inference can’t hap without training; a exemplary indispensable larn patterns successful a acceptable of information earlier it tin efficaciously extrapolate from this grooming data.
Many types of hardware tin execute inference, ranging from smartphone processors to beefy GPUs to custom-designed AI accelerators. But not each of them tin tally models arsenic well. Very ample models would instrumentality ages to marque predictions on, say, a laptop versus a unreality server with high-end AI chips.
[See: Training]
Large connection exemplary (LLM)
Large connection models, oregon LLMs, are the AI models utilized by fashionable AI assistants, specified arsenic ChatGPT, Claude, Google’s Gemini, Meta’s AI Llama, Microsoft Copilot, oregon Mistral’s Le Chat. When you chat with an AI assistant, you interact with a ample connection exemplary that processes your petition straight oregon with the assistance of antithetic disposable tools, specified arsenic web browsing oregon codification interpreters.
LLMs are heavy neural networks made of billions of numerical parameters (or weights, spot below) that larn the relationships betwixt words and phrases and make a practice of language, a benignant of multidimensional representation of words.
These models are created from encoding the patterns they find successful billions of books, articles, and transcripts. When you punctual an LLM, the exemplary generates the astir apt signifier that fits the prompt.
(See: Neural network)
Memory cache
Memory cache refers to an important process that boosts inference (which is the process by which AI works to make a effect to a user’s query). In essence, caching is an optimization technique, designed to marque inference much efficient. AI is evidently driven by high-octane mathematical calculations and each clip those calculations are made, they usage up much power. Caching is designed to chopped down connected the fig of calculations a exemplary mightiness person to tally by redeeming peculiar calculations for aboriginal idiosyncratic queries and operations. There are antithetic kinds of representation caching, though 1 of the much well-known is KV (or cardinal value) caching. KV caching works successful transformer-based models, and increases efficiency, driving faster results by reducing the magnitude of clip (and algorithmic labor) it takes to make answers to idiosyncratic questions.
(See: Inference)
Neural network
A neural web refers to the multi-layered algorithmic operation that underpins heavy learning — and, much broadly, the full roar successful generative AI tools pursuing the emergence of ample connection models.
Although the thought of taking inspiration from the densely interconnected pathways of the quality encephalon arsenic a plan operation for information processing algorithms dates each the mode backmost to the 1940s, it was the overmuch much caller emergence of graphical processing hardware (GPUs) — via the video crippled manufacture — that truly unlocked the powerfulness of this theory. These chips proved good suited to grooming algorithms with galore much layers than was imaginable successful earlier epochs — enabling neural network-based AI systems to execute acold amended show crossed galore domains, including dependable recognition, autonomous navigation, and cause discovery.
(See: Large connection exemplary [LLM])
Open source
Open root refers to bundle — or, increasingly, AI models — wherever the underlying codification is made publically disposable for anyone to use, inspect, oregon modify. In the AI world, Meta’s Llama household of models is simply a salient example; Linux is the celebrated humanities parallel successful operating systems. Open root approaches let researchers, developers, and companies astir the satellite to physique connected apical of 1 another’s work, accelerating advancement and enabling autarkic information audits that closed systems cannot easy provide. Closed root means the codification is backstage — you tin usage the merchandise but not spot however it works, arsenic is the lawsuit with OpenAI’s GPT models — a favoritism that has go 1 of the defining debates successful the AI industry.
Parallelization
Parallelization means doing galore things astatine the aforesaid clip alternatively of 1 aft different — similar having 10 employees moving connected antithetic parts of a task astatine the aforesaid clip alternatively of 1 worker doing everything sequentially. In AI, parallelization is cardinal to some grooming and inference: modern GPUs are specifically designed to execute thousands of calculations successful parallel, which is simply a large crushed wherefore they became the hardware backbone of the industry. As AI systems turn much analyzable and models turn larger, the quality to parallelize enactment crossed galore chips and galore machines has go 1 of the astir important factors successful determining however rapidly and cost-effectively models tin beryllium built and deployed. Research into amended parallelization strategies is present a tract of survey successful its ain right.
RAMageddon
RAMageddon is the amusive caller word for a not-so-fun inclination that is sweeping the tech industry: an ever-increasing shortage of random entree memory, oregon RAM chips, which powerfulness beauteous overmuch each the tech products we usage successful our regular lives. As the AI manufacture has blossomed, the biggest tech companies and AI labs — each vying to person the astir almighty and businesslike AI — are buying truthful overmuch RAM to powerfulness their information centers that there’s not overmuch near for the remainder of us. And that proviso bottleneck means that what’s near is getting much and much expensive.
That includes industries similar gaming (where large companies person had to raise prices connected consoles due to the fact that it’s harder to find representation chips for their devices), user electronics (where representation shortage could origin the biggest dip successful smartphone shipments successful much than a decade), and wide endeavor computing (because those companies can’t get capable RAM for their ain information centers). The surge successful prices is lone expected to halt aft the dreaded shortage ends but, unfortunately, there’s not truly overmuch of a sign that’s going to hap anytime soon.
Reinforcement learning
Reinforcement learning is simply a mode of grooming AI wherever a strategy learns by trying things and receiving rewards for close answers — similar grooming your beloved favored with treats, but the “pet” successful this script is simply a neural web and the “treat” is simply a mathematical awesome indicating success. Unlike supervised learning, wherever a exemplary is trained connected a fixed dataset of labeled examples, reinforcement learning lets a exemplary research its environment, instrumentality actions, and continuously update its behaviour based connected the feedback it receives. This attack has proven particularly almighty for grooming AI to play games, power robots, and, much recently, sharpen the reasoning quality of ample connection models. Techniques similar reinforcement learning from quality feedback, oregon RLHF, are present cardinal to however starring AI labs fine-tune their models to beryllium much helpful, accurate, and safe.
Token
When it comes to human-machine communication, determination are immoderate evident challenges — radical pass utilizing quality language, portion AI programs execute tasks done analyzable algorithmic processes informed by data. Tokens span that gap: they are the basal gathering blocks of human-AI communication, representing discrete segments of information that person been processed oregon produced by an LLM. They are created done a process called tokenization, which breaks down earthy substance into bite-sized units a connection exemplary tin digest, akin to however a compiler translates quality connection into binary codification a machine tin understand. In endeavor settings, tokens besides find outgo — astir AI companies complaint for LLM usage connected a per-token basis, meaning the much a concern uses, the much it pays.
Token throughput
So again, tokens are the tiny chunks of substance — often parts of words alternatively than full ones — that AI connection models interruption connection into earlier processing it; they are astir analogous to “words” for the purposes of knowing AI workloads. Throughput refers to however overmuch tin beryllium processed successful a fixed play of time, truthful token throughput is fundamentally a measurement of however overmuch AI enactment a strategy tin grip astatine once. High token throughput is simply a cardinal extremity for AI infrastructure teams, since it determines however galore users a exemplary tin service simultaneously and however rapidly each of them receives a response. AI researcher Andrej Karpathy has described feeling anxious erstwhile his AI subscriptions beryllium idle — echoing the feeling helium had arsenic a grad pupil erstwhile costly machine hardware wasn’t being afloat utilized — a sentiment that captures wherefore maximizing token throughput has go thing of an obsession successful the field.
Training
Developing instrumentality learning AIs involves a process known arsenic training. In elemental terms, this refers to information being fed successful in bid that the exemplary tin larn from patterns and make utile outputs. Essentially, it’s the process of the strategy responding to characteristics successful the information that enables it to accommodate outputs towards a sought-for extremity — whether that’s identifying images of cats oregon producing a haiku connected demand.
Training tin beryllium costly due to the fact that it requires lots of inputs, and the volumes required person been trending upwards — which is wherefore hybrid approaches, specified arsenic fine-tuning a rules-based AI with targeted data, tin assistance negociate costs without starting wholly from scratch.
[See: Inference]
Transfer learning
A method wherever a antecedently trained AI exemplary is utilized arsenic the starting constituent for processing a caller exemplary for a antithetic but typically related task – allowing cognition gained successful erstwhile grooming cycles to beryllium reapplied.
Transfer learning tin thrust ratio savings by shortcutting exemplary development. It tin besides beryllium utile erstwhile information for the task that the exemplary is being developed for is somewhat limited. But it’s important to enactment that the attack has limitations. Models that trust connected transportation learning to summation generalized capabilities volition apt necessitate grooming connected further information successful bid to execute good successful their domain of focus
(See: Fine tuning)
Weights
Weights are halfway to AI training, arsenic they find however overmuch value (or weight) is fixed to antithetic features (or input variables) successful the information utilized for grooming the strategy — thereby shaping the AI model’s output.
Put different way, weights are numerical parameters that specify what’s astir salient successful a dataset for the fixed grooming task. They execute their relation by applying multiplication to inputs. Model grooming typically begins with weights that are randomly assigned, but arsenic the process unfolds, the weights set arsenic the exemplary seeks to get astatine an output that much intimately matches the target.
For example, an AI exemplary for predicting lodging prices that’s trained connected humanities existent property information for a people determination could see weights for features specified arsenic the fig of bedrooms and bathrooms, whether a spot is detached oregon semi-detached, whether it has parking, a garage, and truthful on.
Ultimately, the weights the exemplary attaches to each of these inputs bespeak however overmuch they power the worth of a property, based connected the fixed dataset.
Validation loss
Validation nonaccomplishment is simply a fig that tells you however good an AI exemplary is learning during grooming — and little is better. Researchers way it intimately arsenic a benignant of real-time study card, utilizing it to determine erstwhile to halt training, erstwhile to set hyperparameters, oregon whether to analyse a imaginable problem. One of the cardinal concerns it helps emblem is overfitting, a information successful which a exemplary memorizes its grooming information alternatively than genuinely learning patterns it tin generalize to caller situations. Think of it arsenic the quality betwixt a pupil who genuinely understands the worldly and 1 who simply memorized past year’s exam — validation nonaccomplishment helps uncover which 1 your exemplary is becoming.
This nonfiction is updated regularly with caller information.
When you acquisition done links successful our articles, we whitethorn gain a tiny commission. This doesn’t impact our editorial independence.















English (US) ·