The Agentic Shift: AI Isn't Just Thinking Anymore, It's Doing!
- Tushar Prasad

- Jun 2
- 7 min read
Alright, let's talk AI. It feels like just yesterday we were marveling at AI that could write a poem, paint a picture, or chat like an old friend. That’s amazing stuff, no doubt! But hold onto your hats, because the AI world is taking another giant leap. We're moving beyond AI that just creates to AI that does stuff on its own. This is what many are calling the "Agentic Shift," and it’s pretty exciting.
So, what's the big deal? This post is all about unpacking that. We'll look at what this new breed of "Agentic AI" is, peek under the hood to see how it works (with a fun example!), and even touch on some of the nitty-gritty like the "Model Context Protocol" – a fancy name for making sure these AI agents can talk to the rest of the digital world.

Quick Rewind: The AI That Creates (Generative AI)
You’ve probably bumped into Generative AI (GenAI). Think ChatGPT, Gemini, DALL-E – these are the wizards that conjure up new content. Whether it’s text, code, images, or even music, they learn from mountains of data and then cleverly predict what should come next.
GenAI is a game-changer for brainstorming, summarizing, translating, you name it. It’s like having a super-smart, creative assistant. But it mostly waits for us to tell it what to do. It's reactive. What if AI could take the initiative?
Agentic AI: The AI That Gets Things Done
This is where Agentic AI struts onto the stage. If GenAI is your brilliant in-house writer or artist, think of Agentic AI as your super-efficient, autonomous project manager or strategist.
The core idea? Agentic AI is designed to act on its own to achieve complex goals, often with multiple steps, and without needing constant hand-holding. These AI systems are proactive. They don't just dream up a plan; they can actually execute it, interact with the world around them, use different tools, and even learn as they go.
This isn't just a small step; it’s a whole new way of thinking about AI – moving it from a creative tool to a genuine partner in getting things accomplished. But what does one of these "agents" even look like?
You might have a single, really smart agent tackling a job, or even a team of specialized AI agents working together, sharing information, and coordinating their efforts. They need to sense what's going on, reason about it, make plans, and then actually do something, often using a powerful AI like an LLM as their central "brain."

Peeking Under the Hood: What Makes an AI Agent Tick?
So, how does an AI agent actually pull this off? While the flashy LLMs (Large Language Models like GPT) are often the "brains" of the operation, they don't work in a vacuum.
They're surrounded by a support system of other crucial bits and pieces:
The LLM Core (The "Brain"): This is where the heavy lifting of understanding, reasoning, planning, and decision-making happens.
The Planning Module: This is like the agent's to-do list creator. It breaks down big, hairy goals into smaller, manageable steps. It can also be smart enough to rethink the plan if things change.
The Memory Module: Just like us, agents need memory!
Short-Term Memory: Keeps track of what's happening right now – the current conversation, recent findings, etc.
Long-Term Memory: Stores past experiences, learned facts, user preferences, and general knowledge (often using cool tech like Vector Databases to find relevant info quickly).
The Tool Use Module (The "Hands"): This is super important. It's how the agent interacts with the outside world. Think of it as giving the AI hands to use various digital tools – APIs for booking flights, databases for looking up customer info, web search for current events, or even running bits of code.

These parts all work together in a kind of loop, often described as
Perceive -> Reason -> Plan -> Act -> Learn/Observe.
The agent perceives what you're asking or what's happening, reasons about it, makes a plan, acts on that plan (often by using its tools), and then observes what happened to learn and do better next time. It keeps going through this loop until the job's done.
Let's Get Real: An AI Agent Planning Your Holiday
Okay, enough theory! Let's imagine you want your AI agent to plan a 7-day trip to Italy. You tell it: "Hey AI, plan a 7-day trip to Italy for two, sometime in July. My budget is $5000, and we love history and food!"
Perceive: The AI agent gets your request. "Got it! Italy, July, 2 people, $5k, history & food focus."
Reason/Plan:
The LLM brain starts whirring: "Italy in July... popular! Okay, Rome and Florence are great for history and food. Must-sees: Colosseum, Vatican, Uffizi. Food: pasta class, wine tasting. Need to juggle flights, hotels, activities, and that budget."
The Planning module breaks it down: Find flights, book hotels, draft an itinerary, find cool food experiences, and keep an eye on the cash.
Act (Using its Tools!):
Flight Tool: The agent connects to a flight API (think Skyscanner's backend). "Searching for flights to Rome/Florence for two in July..." It saves good options to its short-term memory.
Hotel Tool: Next, it pings a hotel booking API (like Booking.com's system). "Finding well-rated hotels near historical sites, within budget..."
Activities/Search Tool: It then uses another API or a smart web search. "Looking for Colosseum tours, Uffizi tickets, pasta-making classes, top-rated local restaurants..."
Internal Budget Tracker: All along, it’s adding up the costs. "Flights: $1200, Hotel: $1500..."
Learn/Observe (And maybe rethink):
The agent looks at what it found: "Okay, flights are $1200, hotel is $1500, tours might be $800. That leaves $1500 for food and getting around. Looking good!"
Or, it might hit a snag: "Whoa, flights are way pricier than I thought! We're over budget." Now it reflects: "Should I suggest different dates? A slightly less fancy hotel? Or just show the user the current options and ask if the budget is flexible or if they want to cut something?"
Based on this, it might tweak the plan, maybe even go back and re-search with new criteria.
Finally, it comes back to you: "Here's a plan! Flights to Rome, 3 nights at Hotel XYZ, then a train to Florence for 3 nights at Hotel ABC. I've penciled in a Colosseum tour, a pasta class, and some awesome restaurant suggestions. Total estimated cost: $4850. What do you think?"
That whole back-and-forth, using tools, and adjusting the plan? That's Agentic AI in action!
The "Lost in Translation" Problem: Enter MCP
Now, for an agent to do all that cool stuff – book flights, check hotel availability, access your calendar – it needs to talk to a LOT of different systems. Each system might speak a slightly different "language" (API). Imagine trying to plug a US appliance into a European socket without an adapter – it just doesn't work!
Historically, connecting an LLM to each new tool meant building a custom "adapter" every single time. That's slow, clunky, and a real headache for developers.

This is where something called the Model Context Protocol (MCP) comes in. Think of it as a universal adapter or a common language for AI. Initiated by folks like Anthropic, it's an open standard trying to solve this "lost in translation" problem.
Why MCP is a Big Deal
LLMs are smart, but they're often stuck in their own little worlds, cut off from the real-time, specific info they need to be truly helpful. MCP aims to change that.
It gives LLM-powered applications (the "MCP Hosts") a standardized way to connect with all sorts of data sources and tools (the "MCP Servers").
This means less custom coding, making it easier and faster to integrate new tools.
It helps LLMs get dynamic access to external information, making them much more context-aware and, frankly, more useful.
How MCP Works (The Simple Version)
You have your MCP Host (like our holiday planning AI app).
Inside it, an MCP Client (a special connector) sends out requests.
These requests go to an MCP Server (which could be a flight booking service, your company's database, or your calendar, as long as it "speaks MCP").
The MCP Server does its thing and sends a structured response back.

The Upside of MCP
Simpler integration, tools become reusable across different AI apps, LLMs get smarter with real-time context, and the whole system can scale much more easily. It's still early days for MCP, but it's a vital step if we want this whole agentic AI thing to really take off. (Just a note: MCP helps with connection; security like who can access what still needs separate layers!)
Of course, it's not all sunshine and rainbows. There are real challenges to tackle:
Oops, Did I Make That Up? (Reliability & Hallucination): LLMs can still make mistakes, and in a multi-step process, those errors can snowball.
Keeping it on the Rails (Controllability & Safety): We need to be sure these agents do what we want them to do and don't go off-script in harmful ways.
The Price of Power (Scalability & Cost): All those AI brain-cycles can get expensive and slow things down.
What Just Happened? (Complexity & Debugging): Figuring out why an autonomous agent did something unexpected can be tricky.
The journey of AI has been incredible – from crunching numbers, to understanding language, to creating art, and now, to taking action. This "Agentic Shift" is a huge step. AI isn't just thinking anymore; it's starting to do.
With powerful AI brains, smart architectures, and crucial connectors like MCP, these Agentic AI systems are set to change how we work, how businesses operate, and how we interact with technology every day. The possibilities are massive, from handling tedious tasks to creating amazing, personalized services.
The road ahead will have its bumps, especially around making these agents reliable, safe, and understandable. But the push towards more capable, autonomous AI is undeniable. The adventure is in building this future responsibly, making sure these powerful new tools work for us, and with us.
What do you think about AI that can take action on its own? Cool? A little scary? How do you see it changing things in your world? I'd love to hear your thoughts in the comments!



Comments