Learning HubModule 03
🤖 Agentic AI & Automation

Beyond chatbots:
AI that acts, not
just answers.

Agentic AI systems don't just respond to questions — they take sequences of actions to complete complex tasks. Here's what that means, why it matters, and how to deploy it responsibly.

Reading time~16 min
LevelAll levels
Module03 of 06
ByAna Rubio Herrera
01

What is agentic AI — and why does it matter now?

Most people's experience of AI is conversational: you ask, it answers. You prompt, it generates. This is useful — but it's a fraction of what AI systems can now do.

Agentic AI refers to systems that don't just respond — they act. Given a goal, an agentic system breaks it into steps, uses tools to complete each step, evaluates the results, and continues until the task is done. It can search the web, read documents, write and run code, call APIs, send emails, update databases — all autonomously, in sequence, without a human prompting each step.

This is a meaningful shift. It moves AI from assistant to executor. And it changes the nature of the governance, oversight, and design questions you need to be asking.

A simple example

Instead of asking an AI to "write a competitive analysis," an agentic system can be given the goal of producing one — and will autonomously search for competitors, read their websites, extract relevant data, structure a comparison, and deliver a finished document. No prompt engineering at each step. Just a goal and a result.

Key takeaway
  • Agentic AI completes multi-step tasks autonomously — it's not a chatbot, it's a workflow executor.
  • This changes the risk profile significantly: mistakes compound across steps, and a human may not be present to catch them.
  • The organisations deploying agentic AI effectively today will have a significant head start within 12–18 months.
02

GenAI vs agentic AI — what's actually different?

The distinction matters for how you design, deploy, and govern these systems.

Standard GenAI
  • Single prompt → single response
  • Human drives every step
  • No memory between conversations
  • No access to external tools by default
  • Low autonomy, high human control
  • Risk contained to one output
Agentic AI
  • Goal → autonomous multi-step execution
  • AI drives the steps, human sets the goal
  • Persistent memory across tasks
  • Accesses tools: web, APIs, databases, code
  • High autonomy, variable human control
  • Risk compounds across a chain of actions

The key implication: standard GenAI is a tool you use. Agentic AI is a system you deploy. The design, testing, oversight, and governance requirements are fundamentally different — and significantly more demanding.

Key takeaway
  • Don't apply GenAI governance thinking to agentic systems — they need their own framework.
  • The higher the autonomy, the more critical the upfront design and testing.
  • Human-in-the-loop is not optional for high-stakes agentic workflows — build it into the architecture from the start.
03

How agentic systems work — the key components

Understanding the architecture helps you ask better questions when evaluating or commissioning agentic systems. Here are the core components of a typical agentic AI workflow.

🎯
Goal definition
A human defines the objective — either through a prompt, a form, or a trigger event. The clearer and more constrained the goal, the more reliably the agent performs. Vague goals produce unpredictable behaviour.
🧠
Planning & reasoning
The LLM at the core of the agent breaks the goal into a sequence of steps and decides which tools to use for each one. This is where the quality of the underlying model matters most.
🔧
Tool use
The agent executes each step using available tools — web search, code execution, database queries, API calls, file operations. The set of tools available defines what the agent can and can't do.
💾
Memory & context
Agents maintain context across steps — short-term (within a task) and sometimes long-term (across tasks). This enables coherent multi-step execution but also introduces data retention considerations.
Evaluation & output
The agent evaluates whether the goal has been achieved and delivers the result — or loops back to try a different approach. Human review at this stage is critical for high-stakes outputs.
Key takeaway
  • The quality of goal definition determines the quality of agentic output — invest time in specifying what "done" looks like.
  • Tool access is a governance decision: only give agents access to the tools they need for the task, nothing more.
  • Build evaluation checkpoints into agentic workflows — don't let them run fully unmonitored for high-stakes tasks.
04

Real use cases — where agentic AI creates value today

Agentic AI is moving fast from research to production. Here are the use cases delivering real value right now — including in contexts I've worked in directly.

📊
Automated reporting
Agents that pull data from multiple sources, run analysis, generate narrative commentary, and produce a finished report — triggered on a schedule or by an event.
🔍
Research & due diligence
Given a company or topic, an agent searches multiple sources, extracts key information, flags risks, and delivers a structured briefing — in minutes rather than days.
📋
Project workflow automation
Agents that generate status reports, update project documentation, draft meeting summaries, and route actions to the right people — triggered by calendar events or task completions.
🛠
Code generation & review
Agents that write, test, debug, and document code — reducing development time for well-scoped tasks significantly. Most effective when the scope is tight and the tests are clear.
📩
Inbox & comms triage
Agents that read, classify, summarise, and draft responses to incoming communications — flagging what needs human attention and handling routine queries autonomously.
🔄
Data pipeline orchestration
Agents that monitor data pipelines, detect anomalies, trigger remediation actions, and notify the right people — reducing the manual overhead of data operations significantly.
📍 SDG Group — Agentic AI in transformation programs

In insurance transformation programs at SDG Group, agentic AI frameworks were deployed to automate project workflows and deliverable generation. The results were significant — tasks that previously required hours of manual work were completed autonomously in minutes, with human review at the output stage. The key design decision was defining tight scope: each agent had a clearly bounded task, access only to the data it needed, and a human checkpoint before any output was shared externally. This constraint-first design is what made it safe to deploy at scale.

Key takeaway
  • The highest-value agentic use cases are those with clear inputs, clear outputs, and well-defined success criteria.
  • Start with internal, lower-stakes workflows before deploying agents in customer-facing or regulated contexts.
  • Tight scope + human review at output = the minimum viable governance for agentic deployment.
05

Governance & risk — what's different with agents

Agentic AI introduces governance challenges that simply don't exist with standard GenAI. The autonomy that makes agents powerful is also what makes them riskier. Here's what you need to think about.

⚠ Compounding errors

In a multi-step agentic workflow, an error in step 2 propagates through steps 3, 4, and 5. By the time a human sees the output, the mistake may be deeply embedded in the result. Build validation checkpoints into long chains — don't let agents run 20 steps without a review gate.

⚠ Scope creep

Agents given broad goals and broad tool access will sometimes take actions outside the intended scope — accessing data they shouldn't, triggering unintended side effects, or making decisions that should have been escalated. Minimum necessary access is not just a security principle — it's an agentic design principle.

⚠ Auditability

When an agent takes 15 actions to complete a task, can you reconstruct exactly what it did and why? If not, you have an auditability problem — especially in regulated industries. Every agentic deployment in a regulated context needs a full action log.

The governance principles for agentic AI build on the framework from Module 02, with three additional requirements: minimum tool access, mandatory checkpoints, and full auditability. These are non-negotiable for any production agentic deployment.

Key takeaway
  • Give agents the minimum tool access necessary — not everything they could theoretically use.
  • Build human review gates into long agentic chains, especially before any external action (sending, publishing, updating).
  • Ensure every production agentic system produces a full, queryable action log — this is your audit trail.
  • For regulated industries: get legal and compliance involved before deployment, not after.
06

How to start — your first agentic deployment

The most common mistake I see is organisations trying to build a complex, multi-system agentic deployment before they understand how agents actually behave in practice. Start small. Learn fast. Scale what works.

Action plan — first agentic deployment
  • Pick one internal, low-stakes workflow. Something that is repetitive, time-consuming, and has clear inputs and outputs. Internal reporting, meeting summaries, or research briefings are good starting points.
  • Define "done" precisely. What does a successful output look like? What format? What information must it include? What would make it wrong? Write this down before you build anything.
  • Start with existing tools. Before building custom agents, explore what tools like Microsoft Copilot, Claude, or n8n can do out of the box for your use case. You may not need to build anything custom for your first deployment.
  • Run it in parallel with the manual process. For the first 2–4 weeks, have a human do the task manually and compare with the agent output. This surfaces errors and builds confidence before you remove the manual step.
  • Document what went wrong. Every agentic deployment will have failure modes you didn't anticipate. Capture them, fix them, and build your institutional knowledge about how agents behave in your specific context.

Agentic AI is not science fiction — it is being deployed in production today, in organisations of all sizes. The question is not whether to engage with it, but how to do so in a way that builds capability without creating unmanaged risk. Start now, start small, and build the muscle before you need it at scale.

Up next
Module 04: Data Platform Modernisation
Module 04: Data Platform Modernisation