π Add to Chrome β Itβs Free - YouTube Summarizer
Category: AI Education
Tags: AIconferenceinnovationlearningtechnology
Entities: agentic AIAIartificial superintelligenceIBMlarge reasoning modelsMixture of ExpertsModel Context Protocolretrieval augmented generationTechXchangevector database
00:00
There are two things that hold true when it comes to artificial intelligence. One.
It's everywhere. My toothbrush just got an AI update this week.
And two. The field is changing rapidly, making it hard to keep up even for those of us who work in tech.
So, I've put together
00:16
my top seven AI terms that I think are important to be familiar with as AI continues to progress. How many do you already know?
Well, let's find out. And I'm gonna to start at number one with something that I'm quite sure that you have heard of.
00:33
And that's agentic AI. Everybody and their grandmother seems to be building the next generation of AI agents.
But what exactly are they? Well, AI agents, they can reason and act autonomously to achieve goals.
Something like a chatbot that only responds one prompt at a time,
00:51
AI agents, they run autonomously. They go through a number of stages.
So, first of all, they perceive their environment. Once they've done that, they move on to a reasoning stage, and that's where they look to see what the next best steps forwards are.
01:10
Then they move on to act on the plan that it's built through the reasoning, and then observes the results of that action. And around and around we go.
Now agents can work as well in all sorts of roles.
01:27
They could be your travel agent to book a trip. They could be your data analyst to spot trends in quarterly reports.
Or they could perform the role of a DevOps engineer, detecting anomalies in logs and spinning up containers to test fixes and rolling back faulty deployments.
01:44
And AI agents are typically built using a particular form of large language models, and that is known as number two, large reasoning models. Now these are specialized LLMs that have undergone reasoning-focused fine tuning.
So unlike regular
02:00
LLMs that generate responses immediately, reasoning models, they're trained to work through problems step by step, which is exactly what agents need when planning complex, multistep tasks. Now, the reasoning model is trained on problems with verifiably correct answers.
02:19
So math problems or code that can be tested by compilers and through reinforcement learning, the model learns to generate reasoning sequences that lead to correct final answers. So, every time you see a chatbot pause before it responds back to you by saying, thinking.
Well,
02:38
that's the reasoning model at work, generating an internal chain of thought to break down a problem step by step before generating a response. Now let's get a bit lower level and talk about number three, which is vector database.
02:55
So, in a vector database, we don't store raw data like text files and like images just as blobs of data. We actually use something called an embedding model.
03:12
And that embedding model is used to convert that data from these images here into actually a vector. Now, what is a vector?
Well, a vector is essentially just kind of a a long list of numbers.
03:28
And that long list of numbers captures the semantic meaning of the context. Now, what's the benefit of doing that?
Well, in a vector database, we can perform searches as mathematical operations, looking for vector embeddings
03:43
that are close to each other. And that translates to finding semantically similar content.
So, we might start, with let's say a picture of a mountain vista. Something like this.
And then that picture is broken down by the embedding model
04:01
into vectors, a multidimensional numeric vector. And we can perform a similarity search to find items that are similar to that mountain picture by finding the closest vectors
04:16
in the embedding space. Or it could be similar text articles, or it could be similar music files.
Whatever you want. Now vector databases, they play a big role in implementing number four.
And that is RAG or retrieval augmented
04:32
generation. Now, RAG makes use of these vector databases.
And it uses it to enrich prompts to an LLM. So, we start here with a RAG retriever component.
04:49
Now that might take in an input prompt from a user. And it's going to turn it into a vector using an embedding model.
That's the thing that ties it into that series of numbers.
05:05
And then, once we've done that, we can perform a similarity search in the vector database. Now that vector database will return something, and we'll return that all the way back to the large language model prompt that we started with.
05:21
And we'll embed into that prompt now the stuff that came out of that vector database. So, I can I can ask a question about let's say company policy.
And then this RAG system is going to pull the relevant section from the employee handbook to include in the prompt. Now, number five,
05:38
that Model Context Protocol or MCP. This is a really exciting one because for large language models to be truly useful, they need to interact with external data sources and services and tools.
And MCP standardizes how applications provide context to LLMs.
05:56
So, if you want your large language model here to be able to connect to stuff. Perhaps we want to connect to an external database, or maybe we want to go to some kind of code repository, or maybe even to an email server,
06:15
or really any kind of external system. Well, MCP makes that connection standardized.
So, instead of developers having to build one off connections for each new tool, MCP provides a standardized way for AI to access your systems.
06:30
So basically we have here an MCP server. And that is how the AI knows exactly what to do to get through to any one of these tools.
It connects through that MCP server connection. Okay.
Now, for number six.
06:48
That's a mixture of experts or MOE. And we've had the idea of MOE for a good while, actually, since the paper was published in a scientific journal in 1991.
But, it's never been more relevant than it is today. You see, MoE divides a large language model
07:05
into a series of experts. I'm just gonna to draw three, but there could be 100 plus of these.
These are specialized neural subnetworks. And then it uses a routing mechanism to activate
07:22
only the agents it needs for a particular task, or only the experts in this case that it needs for a task. And then, well, then it's going to perform a merge process.
So, because we activated these two experts, we'll merge these two.
07:38
And this performs mathematical operations to combine the output from these different experts into a single representation that continues through the rest of the model. And it's a really efficient way to scale up model size without proportional increases in compute costs.
07:54
So, for example, MoE models, like IBM Granite's 4.0 series, that can have dozens of different experts here. But for any given token, it will only activate these specific experts it needs.
And that means, though, while the whole model
08:09
might have billions of total parameters, it only uses a fraction of those active parameters at inference time. And look, for number seven, I'm gonna throw in a big one, ASI, artificial superintelligence.
08:27
It's the goal of all the frontier AI labs. And at this point, it is purely theoretical.
It doesn't actually exist and we don't know if it ever will. Now, today's best models, they're slowly approaching a different standard, which is AGI.
08:42
That's artificial general intelligence. Now that's also theoretical.
But, if realized, AGI will be able to complete all cognitive tasks as well as any human expert. ASI is one step beyond that.
So, ASI systems would have an intellectual scope
09:00
beyond human level intelligence, potentially capable of recursive self-improvement. So, basically an ASI system could redesign and upgrade itself, becoming ever smarter in an endless cycle.
It's the kind of development that would either solve humanity's biggest problems
09:18
or create entirely new ones that we can't even imagine yet. And if that happens, well, I think it's probably a pretty good idea that we keep the term ASI on our radar.
So, that's my seven. But I'm curious, what's the AI term
09:35
you think that should have made it onto this list? Let me know in the comments.
Hey, Martin. Hey Graeme.
Hey, this is really cool stuff. This AI and these terms.
Fascinating. Yeah, and I came up with seven, but I could have come up with 70.
09:50
There's so much going on in this space. I I I bet you could.
And you know what? There is so much going on.
We are actually going to be talking about AI a lot at the IBM TechXchange conference in Orlando this October. And you know what?
I'm gonna be there as well. I know it's gonna be so exciting!
There's going to be so much going on.
10:07
We are going to have. Let's see.
We're gonna have boot camps. We're going to have workshops.
There's going to be sessions. There's going to be live demos, certifications, all kinds of things going on.
So much more when it comes to AI. But but Martin, what are you going to be doing there?
10:22
Well, I'm going to be bringing my light board pens and this light board to the sessions as well. Oh my God!
It's so exciting! I'm so excited have you there!
Yes. So we're actually going to have a light board studio set up.
And ah we're going to be performing light boards live. So if you always wondered how do I write backwards,
10:38
you're going to find out in person at the event. And da also we'll be kind of teaching how to perform a light board video yourself.
The sort of things that you need to know for that. Wow.
So you get to meet a celebrity and maybe become one yourself. That sounds really exciting.
I can't wait to welcome you down to Orlando.
10:55
It's just going to be a blast. So looking forward to it.
Can't wait. All right.
Hope we see you there too. So, go to ibm.com/techXchange and we'll see you down there.