7 AI Terms You Need to Know: Agents, RAG, ASI & More

πŸš€ Add to Chrome – It’s Free - YouTube Summarizer

Category: AI Education

Tags: AIconferenceinnovationlearningtechnology

Entities: agentic AIAIartificial superintelligenceIBMlarge reasoning modelsMixture of ExpertsModel Context Protocolretrieval augmented generationTechXchangevector database

Building WordCloud ...

Summary

    Introduction to AI
    • AI is ubiquitous and rapidly evolving, making it challenging to keep up with.
    • The video introduces seven important AI terms to understand current developments.
    AI Agents and Reasoning Models
    • Agentic AI refers to AI agents that can reason and act autonomously to achieve goals.
    • Large reasoning models are specialized LLMs trained to work through problems step by step, essential for planning complex tasks.
    Vector Databases and RAG
    • Vector databases store data as vectors, enabling searches for semantically similar content.
    • Retrieval Augmented Generation (RAG) uses vector databases to enhance prompts to LLMs by retrieving relevant data.
    Model Context Protocol and Mixture of Experts
    • Model Context Protocol (MCP) standardizes how applications provide context to LLMs, facilitating connections to external systems.
    • Mixture of Experts (MOE) divides a model into neural subnetworks, activating only necessary experts for tasks, optimizing efficiency.
    Artificial Superintelligence
    • Artificial Superintelligence (ASI) is a theoretical concept of AI surpassing human intelligence, capable of recursive self-improvement.
    • ASI could potentially solve major problems or create unforeseen challenges.
    Takeaways
    • Stay informed about AI developments as the field evolves rapidly.
    • Understand key AI terms like agentic AI, reasoning models, and vector databases.
    • Recognize the role of RAG in enhancing AI prompts with relevant data.
    • Appreciate the importance of MCP in standardizing AI interactions with external systems.
    • Be aware of the theoretical concept of ASI and its potential implications.

    Transcript

    00:00

    There are two things that hold true when it comes to artificial intelligence. One.

    It's everywhere. My toothbrush just got an AI update this week.

    And two. The field is changing rapidly, making it hard to keep up even for those of us who work in tech.

    So, I've put together

    00:16

    my top seven AI terms that I think are important to be familiar with as AI continues to progress. How many do you already know?

    Well, let's find out. And I'm gonna to start at number one with something that I'm quite sure that you have heard of.

    00:33

    And that's agentic AI. Everybody and their grandmother seems to be building the next generation of AI agents.

    But what exactly are they? Well, AI agents, they can reason and act autonomously to achieve goals.

    Something like a chatbot that only responds one prompt at a time,

    00:51

    AI agents, they run autonomously. They go through a number of stages.

    So, first of all, they perceive their environment. Once they've done that, they move on to a reasoning stage, and that's where they look to see what the next best steps forwards are.

    01:10

    Then they move on to act on the plan that it's built through the reasoning, and then observes the results of that action. And around and around we go.

    Now agents can work as well in all sorts of roles.

    01:27

    They could be your travel agent to book a trip. They could be your data analyst to spot trends in quarterly reports.

    Or they could perform the role of a DevOps engineer, detecting anomalies in logs and spinning up containers to test fixes and rolling back faulty deployments.

    01:44

    And AI agents are typically built using a particular form of large language models, and that is known as number two, large reasoning models. Now these are specialized LLMs that have undergone reasoning-focused fine tuning.

    So unlike regular

    02:00

    LLMs that generate responses immediately, reasoning models, they're trained to work through problems step by step, which is exactly what agents need when planning complex, multistep tasks. Now, the reasoning model is trained on problems with verifiably correct answers.

    02:19

    So math problems or code that can be tested by compilers and through reinforcement learning, the model learns to generate reasoning sequences that lead to correct final answers. So, every time you see a chatbot pause before it responds back to you by saying, thinking.

    Well,

    02:38

    that's the reasoning model at work, generating an internal chain of thought to break down a problem step by step before generating a response. Now let's get a bit lower level and talk about number three, which is vector database.

    02:55

    So, in a vector database, we don't store raw data like text files and like images just as blobs of data. We actually use something called an embedding model.

    03:12

    And that embedding model is used to convert that data from these images here into actually a vector. Now, what is a vector?

    Well, a vector is essentially just kind of a a long list of numbers.

    03:28

    And that long list of numbers captures the semantic meaning of the context. Now, what's the benefit of doing that?

    Well, in a vector database, we can perform searches as mathematical operations, looking for vector embeddings

    03:43

    that are close to each other. And that translates to finding semantically similar content.

    So, we might start, with let's say a picture of a mountain vista. Something like this.

    And then that picture is broken down by the embedding model

    04:01

    into vectors, a multidimensional numeric vector. And we can perform a similarity search to find items that are similar to that mountain picture by finding the closest vectors

    04:16

    in the embedding space. Or it could be similar text articles, or it could be similar music files.

    Whatever you want. Now vector databases, they play a big role in implementing number four.

    And that is RAG or retrieval augmented

    04:32

    generation. Now, RAG makes use of these vector databases.

    And it uses it to enrich prompts to an LLM. So, we start here with a RAG retriever component.

    04:49

    Now that might take in an input prompt from a user. And it's going to turn it into a vector using an embedding model.

    That's the thing that ties it into that series of numbers.

    05:05

    And then, once we've done that, we can perform a similarity search in the vector database. Now that vector database will return something, and we'll return that all the way back to the large language model prompt that we started with.

    05:21

    And we'll embed into that prompt now the stuff that came out of that vector database. So, I can I can ask a question about let's say company policy.

    And then this RAG system is going to pull the relevant section from the employee handbook to include in the prompt. Now, number five,

    05:38

    that Model Context Protocol or MCP. This is a really exciting one because for large language models to be truly useful, they need to interact with external data sources and services and tools.

    And MCP standardizes how applications provide context to LLMs.

    05:56

    So, if you want your large language model here to be able to connect to stuff. Perhaps we want to connect to an external database, or maybe we want to go to some kind of code repository, or maybe even to an email server,

    06:15

    or really any kind of external system. Well, MCP makes that connection standardized.

    So, instead of developers having to build one off connections for each new tool, MCP provides a standardized way for AI to access your systems.

    06:30

    So basically we have here an MCP server. And that is how the AI knows exactly what to do to get through to any one of these tools.

    It connects through that MCP server connection. Okay.

    Now, for number six.

    06:48

    That's a mixture of experts or MOE. And we've had the idea of MOE for a good while, actually, since the paper was published in a scientific journal in 1991.

    But, it's never been more relevant than it is today. You see, MoE divides a large language model

    07:05

    into a series of experts. I'm just gonna to draw three, but there could be 100 plus of these.

    These are specialized neural subnetworks. And then it uses a routing mechanism to activate

    07:22

    only the agents it needs for a particular task, or only the experts in this case that it needs for a task. And then, well, then it's going to perform a merge process.

    So, because we activated these two experts, we'll merge these two.

    07:38

    And this performs mathematical operations to combine the output from these different experts into a single representation that continues through the rest of the model. And it's a really efficient way to scale up model size without proportional increases in compute costs.

    07:54

    So, for example, MoE models, like IBM Granite's 4.0 series, that can have dozens of different experts here. But for any given token, it will only activate these specific experts it needs.

    And that means, though, while the whole model

    08:09

    might have billions of total parameters, it only uses a fraction of those active parameters at inference time. And look, for number seven, I'm gonna throw in a big one, ASI, artificial superintelligence.

    08:27

    It's the goal of all the frontier AI labs. And at this point, it is purely theoretical.

    It doesn't actually exist and we don't know if it ever will. Now, today's best models, they're slowly approaching a different standard, which is AGI.

    08:42

    That's artificial general intelligence. Now that's also theoretical.

    But, if realized, AGI will be able to complete all cognitive tasks as well as any human expert. ASI is one step beyond that.

    So, ASI systems would have an intellectual scope

    09:00

    beyond human level intelligence, potentially capable of recursive self-improvement. So, basically an ASI system could redesign and upgrade itself, becoming ever smarter in an endless cycle.

    It's the kind of development that would either solve humanity's biggest problems

    09:18

    or create entirely new ones that we can't even imagine yet. And if that happens, well, I think it's probably a pretty good idea that we keep the term ASI on our radar.

    So, that's my seven. But I'm curious, what's the AI term

    09:35

    you think that should have made it onto this list? Let me know in the comments.

    Hey, Martin. Hey Graeme.

    Hey, this is really cool stuff. This AI and these terms.

    Fascinating. Yeah, and I came up with seven, but I could have come up with 70.

    09:50

    There's so much going on in this space. I I I bet you could.

    And you know what? There is so much going on.

    We are actually going to be talking about AI a lot at the IBM TechXchange conference in Orlando this October. And you know what?

    I'm gonna be there as well. I know it's gonna be so exciting!

    There's going to be so much going on.

    10:07

    We are going to have. Let's see.

    We're gonna have boot camps. We're going to have workshops.

    There's going to be sessions. There's going to be live demos, certifications, all kinds of things going on.

    So much more when it comes to AI. But but Martin, what are you going to be doing there?

    10:22

    Well, I'm going to be bringing my light board pens and this light board to the sessions as well. Oh my God!

    It's so exciting! I'm so excited have you there!

    Yes. So we're actually going to have a light board studio set up.

    And ah we're going to be performing light boards live. So if you always wondered how do I write backwards,

    10:38

    you're going to find out in person at the event. And da also we'll be kind of teaching how to perform a light board video yourself.

    The sort of things that you need to know for that. Wow.

    So you get to meet a celebrity and maybe become one yourself. That sounds really exciting.

    I can't wait to welcome you down to Orlando.

    10:55

    It's just going to be a blast. So looking forward to it.

    Can't wait. All right.

    Hope we see you there too. So, go to ibm.com/techXchange and we'll see you down there.