Blog
Guidelines for Agentic AI Design
- April 21, 2025
- Posted by: William Dorrington
- Category: Beginner CoPilot Data Science Frontiers Experience Design Technology

Table of Contents
Introduction
As artificial intelligence matures from passive assistants into intelligent, active problem solvers, a new category of AI is emerging, Agentic AI. These agents do more than respond to prompts; they perceive, plan, reason, and take action toward achieving goals, often independently. While their potential is transformative, such autonomy demands a considered, principled approach to design and is not something we should rush into straight to build like we’ve done in the past with Canvas Apps and other pieces of innovation. Structure adoption is crucial here!
This guide outlines foundational guidelines for designing Agentic AI systems that are helpful, transparent, reliable, and have appropriate guardrails for interacting with humans, and also without. Especially as i’ve seen an increase in people using Copilot Studion.
This guide’s purpose is to ensure that those who are starting their journey into adopting and building Agentic AI have a starting point and to ensure you don’t repeat some of my past mistakes (there’s many!)
Agentic 101
What is Agentic AI? Simply put it’s:
AI Systems that can AUTONOMOUSLY – PERCEIVE, DECIDE & ACT (AGENCY) to achieve GOALS
Agentic AI refers to artificial intelligence systems that can act independently to achieve goals. These agents don’t just respond to commands or messages/triggers they perceive their environment, make decisions, and take actions without needing constant human input with a constant goal in mind. At its core, Agentic AI is about autonomy and purpose. It’s what separates static chatbots from dynamic, goal-driven digital agents that will only help you reason but not act or overly learn from experience.
To put it even more simply, the difference between Generative AI and Agentic AI. If you were booking a holiday, generative AI could tell you where to go, and Agentic AI (if given the right permissions) would book it for you!
Here are the five essential traits that define agentic behaviour:
-
Autonomy: Agentic systems act on their own; they don’t wait for explicit instructions.
-
Goal-Oriented: They operate with a purpose, aiming to achieve specific outcomes.
-
Adaptability: They adjust their behaviour based on context, feedback, or environmental changes.
-
Memory & Knowledge: They retain and use information from previous interactions to make better decisions.
-
Action-Capable: They can carry out tasks and execute steps toward a goal, not just talk about it.
How does Agentic AI work (high-level)
While traditional AI systems wait to be prompted, Agentic AI actively senses its environment, plans based on goals, and takes action, refining itself through feedback. This is often described using the loop: Perceive > Decide > Act.
Let’s break down the core components in the diagram:
The Agent
The agent is the central key actor that interacts with the environment and space it sits within. It perceives input, makes decisions & plans, and then executes tasks, not passively, but with intent and purpose.
Sensory Interface (Perceive)
This is how the agent receives input, from users, data streams, or sensor feeds (chat interfaces, IoT, Databases etc.) throughout the environment and space it sits in. These inputs trigger its reasoning engine to interpret the current state.
Cognitive Core (Decide/Plan)
- Goals: The desired outcomes or targets that drive the agent’s decision-making.
- LLM (Brain): A large language model interprets input and plans next steps using reasoning and natural language understanding.
- Memory (Experience): The agent can store and retrieve context from both short-term and long-term memory to maintain state and personalise behaviour.
- Knowledge Base: Structured and unstructured content the agent uses to provide grounded and accurate answers.
Together, these enable decision-making, orchestration, and planning through mechanisms like chain-of-thought reasoning and subgoal decomposition.
Action Layer (Act)
Here the agent performs tasks via integrations, APIs, or platform tools. This is the “hands” of the agent, the output side of the pipeline where decisions are executed.
Feedback Loops (Refinement)
Agentic AI systems include internal feedback loops that allow for continual learning and refinement. Failed tasks or unexpected outcomes inform the next cycle of perception and planning.
Now let’s look at some design principles when we are building Agentic AI Systems.
Design Principles
This section walks the reader through some of the baseline Agentic Design principles, focusing on three key areas: space, Time, and core. It focuses on what should be considered when building out the design for your Agentic AI.
Credit: I found here a very useful source when writing this guide gifted by my friend Dona Sarkar!
Agent in Space: Designing for Environment and Interactions
- Clarity of Purpose and Boundaries: Define what the agent is responsible for and what it is not. Scope limitations should be clear to users and implemented to avoid overpromising or confusion.
- Consistency: Ensure a coherent and predictable user experience across platforms, devices, and interaction modes. An agent should behave reliably regardless of context, maintaining a unified personality, tone, and functionality throughout.
- Collaboration and Connectivity: Agents should connect users with systems, data, or people. They should support collaborative workflows and hand off tasks to humans when appropriate, rather than acting as a gatekeeper.
- Accessibility and Inclusivity: Ensure that the agent is usable by a wide range of users, including those with differing abilities or literacy levels. Support multimodal access, assistive technologies, and reduce cognitive load where possible.
Agent in Time: Responsiveness, Adaptation, and Memory
- Past: Analyse previous interactions, states and context. Enabling more relevant results and building better personalisation & confidence.
- Future (Adaptability): Enable agents to adjust over time based on user preferences, prior behaviour, and evolving contexts. Maintain a lightweight memory model where appropriate, while respecting privacy and reset functionality.
- Now (Proactive Assistance): Design agents to offer context-based help, not just requests. Timely prompts or suggestions should anticipate user needs without being intrusive.
Agent Core: Purpose, Trust, and Governance
- User Control: Preserve user autonomy. Allow users to override, pause, or customise agent actions and behaviours. Seek explicit confirmation for actions that impact user data or systems.
- Trust and Safety: Handle uncertainty gracefully. Avoid irreversible actions. Ensure agents stay within their domain and defer to humans or trusted systems when outside their competence.
- Simplicity: Design for clarity. Avoid unnecessarily complex conversation flows or over-engineered solutions. Simplicity aids comprehension, maintenance, and error recovery.
- Transparency: Users should always be aware when they are interacting with AI. Clearly communicate the agent’s purpose, capabilities, limitations, and actions taken on behalf of the user.
Agentic Boundaries
We talk about being in the era of AI and now Agentic AI , but in reality we are in the era of understanding the applicability of AI and the era of building confidence with users in AI. This is where it is INCREDIBLY important to ensure we have a boundaries and Ethics baked into our thinking. The next two sections explore this at a high-level.
As more teams start building with Microsoft Copilot Studio or custom LLM-based orchestration, it’s important to talk not just about what agents can do, but what they shouldn’t do – there’s a Jurrasic park quote in there somewhere! That’s where agent boundaries come in. These boundaries are the behavioural guardrails that keep your agent safe, predictable (as possible), and aligned with user expectations. This isn’t about limiting innovation, it’s about designing with purpose and responsibility.
Below, I’ve outlined five high-level boundaries every agent builder should consider as a starting point, with practical reasoning behind each one.
1. Action Boundaries
- What it means: Your agent shouldn’t take irreversible actions (like deleting files or submitting payments) without user confirmation.
- Why it matters: Trust is hard-earned and easily lost. These moments are where the user needs to stay in control. A confirmation prompt or manual override preserves human agency.
2. Data Boundaries
- What it means: Limit what personal or sensitive data the agent can access or retain. Use context variables sparingly and intentionally.
- Why it matters: Agents that store or misuse sensitive data (even by accident) pose risks to privacy, compliance, and user safety. Boundaries protect against accidental leaks.
3. Scope Boundaries
- What it means: The agent should stay in its lane. For example, a customer support agent shouldn’t give legal advice or respond to HR queries unless specifically designed to do so.
- Why it matters: Agents that “try to do everything” often fail at most things. Domain creep leads to hallucinations, poor performance, and confused users.
4. Escalation Boundaries
- What it means: If the agent isn’t confident or context is unclear, it should hand off to a human or gracefully stop.
- Why it matters: Safety isn’t just about code, it’s about behaviour. Low-confidence guesses lead to frustrated users. Escalation is a strength, not a weakness.
5. Time Boundaries
- What it means: Define how long memory or context is kept are you building short-term sessions or long-term personalised agents? Always give users a reset option.
- Why it matters: Retaining context too long can violate privacy or create confusing agent behaviour. Boundaries keep your AI honest and human-friendly.
Bringing it All Together
A good agent isn’t just helpful, it’s controlled. Boundaries make it trustworthy. Whether you’re designing refund flows in Copilot Studio or building custom orchestration logic in Azure/OpenAI, these five categories help frame your thinking and ensure a level fo success.
When in doubt, ask: Would I want an agent doing this without asking me first? If the answer is no, that’s your boundary.
Ethical Considerations
Please note – this section assumes you have a “Responsible AI strategy” and execution already underway.
If you’re building AI agents, whether using Microsoft Copilot Studio, Azure, or another platform, technical capability is only half the story. Just because an agent can respond doesn’t mean it should respond in certain ways. That’s where ethics come into play.
Designing ethical agents isn’t about ticking boxes. It’s about creating systems people can trust, systems that respect users, and systems that know when to step back.
Here are five key considerations to embed into your design process early, not as an afterthought.
1. Explainability
What it means: Users should always understand what the agent is doing and why.
Best practice: Make reasoning visible, especially when the agent’s output is driven by a knowledge base, action, or user variable. Explain outcomes in plain language.
Example:
“Here’s what I found in your FAQ, would you like to read more?”
2. Consent & Clarity
What it means: Make it obvious that users are talking to an AI, not a human, and clearly state what the agent can and can’t do. I see too many people proud that “my users don’t even know that they are talking with AI” DO NOT be this person!
Best practice: Use upfront welcome messages and disclaimers to set expectations. Let users know what kind of help they can expect.
Example:
“Hi, I’m an AI assistant. I can help with order refunds but not purchases.”
3. Non-Deception
What it means: Agents shouldn’t fake human emotion, empathy, or intentions.
Best practice: Avoid phrasing like “I understand how you feel” unless it’s obviously metaphorical. Let the agent be helpful, not theatrical.
Example:
Don’t say: “I totally understand how frustrating that must be.” Do say: “Let’s see how I can help you fix that.”
4. Bias Minimisation
What it means: Your agent shouldn’t embed or amplify social, racial, or gender or any other prejudicial biases in its answers, tone, or logic.
Best practice: Test prompts and responses regularly across edge cases and diverse user inputs. Monitor language drift over time.
Example:
Build prompt variations with different user names, tones, and demographics to see if the tone or treatment shifts.
5. Fair Failure (read fair not fear, mistyped on the image, but will update when I get a moment)
What it means: When the agent doesn’t know what to do, or shouldn’t answer, it needs to back off gracefully.
Best practice: Don’t guess. Have clear escalation paths or fallback responses that prioritise clarity and user experience.
Example:
“I’m not able to resolve this, but I can connect you to a support agent.”
Final Thought
Building ethical agents isn’t about making them sound human, it’s about making them behave responsibly.
From explainability to fairness, these principles help avoid confusion, reduce risk, and build trust. If you’re serious about building agentic AI that actually works in the real world, ethics isn’t optional, it’s core design.
Let’s build agents that don’t just respond, they respect.
Deterministic Vs Non-Deterministic
Okay, so there may be an element of sucking eggs here but, to me, this is a really important thing to remember when designing your “Agentic AI”. Do you need to have a ‘Deterministic’ pattern or a ‘Non-Deterministic’ pattern?
Deterministic
This means that every single time you put a piece of input into a process, you get the same output.
This makes it predictable and consistent, which is incredibly important for some processes, such as calculations, validations, and compliance.
Non-Deterministic
This is where, even with the exact same input the output can or will be different.
This makes it unpredictable and probabilistic, which is really useful for customer support agents, adaptive learning systems and general knowledge queries that need the ability to be flexible and provide different contextual outputs even if it’s the same input.
In summary, I like to work with analogies, and one that has stuck with me is thinking of deterministic like a chef that follows the instructions, where non-deterministic has the same ingredients but jsut goes for it randomly – but regardless, at the end of both processes you get a dinner!
Model Context Protocol (MCP)
A question that is starting to emerge: How do I ensure consistency and centralised asset management for Knowledge & Actions?
So this is a topic that is starting to get a ton of energy behind it! How to ensure that you can keep track of core knowledge and actions in a scalable, controlled and managed way. This is where Model Context Protocol (MCP) becomes the absolute rockstar!
MCP can tell an AI model 3 key things:
- What knowledge it can use
- What actions it can take
- How to find and understand those things
It’s basically giving your Agent AI a map of everything its allowed to know about, a toolbox of actions and the instructions on how to use both of those safely before it wanders aimlessly and wrecks havoc – like myself after a few rums!
Give me a technical overview of MCP
See it as a specification that defines how to describe, structure and reference external knowledge and actions for AI models.
It provides:
- Metadata schemas that describe knowledge sources and action sources.
- It’s a discovery mechanisms so models & agents can dynamically find and interpret these various assets during runtime.
- It’s a context broker between the AI and external assets without having to hard code specifics into prompts or agent logic.
- Reduction of technical debt by reduction in custom API connections – one ring to rule them all yada-yada!
Technical benefits:
- Separates knowledge and actions from model training & prompts
- It standardises resources discovery, linking and invocation with the AI system.
- It allows for knowledge and actions to be version controlled
- What’s also epic is it allows for dynamic updates. So add new knowledge and actions without having to retrain or redeploy your AI/Agent.
- Improve governance over what Agents and models can access
The takeaway here is if your explore Agentic AI you have to explore MCP too – they both go hand-in-hand. Daniel Laskewitz has recently published a fantastic lab on this for Copilot Studio, go check it out >>> LINK
I will be providing follow up blogs on MCP.
Red Teaming
As we all know, as soon as soon as new tools come out especially intelligent digital internet proud tools malicious people find ways of exploiting them. Wherever there are good actors there are always bad, wherever there is light – there must be dark, whenever…. okay you get my point…. Batman and the Joker, but seriously I’ll stop.
So, what is it I mean by this? Let’s first explore Jailbreaking to show how this can end up:
Jailbreaking what is it?
Jail breaking is a term used where an individual has legitimate access to a product or service, but uses it in a way that either was not intended or goes against the fair usage policy – back in the day people were obsessed with jailbreaking their phones, PlayStation ones etc. …. not me though I was a good boy!
So just like the PlayStation and Phones – Large Language Models can be jailbroke too! This is where a user will try to reduce the gap between “What the Model can do” ________and (the gap)_______ “What the model is willing to do”.

Jailbreaking enables us to change how the model interprets its guardrails to the point of ignoring them although not explicitly but through the contextual exchange of inputs and outputs. It doesn’t know at that point if something is malicious or not.
Let’s take an example:
If I asked chatGPT to give me instructions on how to make a “Molotov Cocktail” (yes you read that right), it’d simply say no…. we’ll, you’d hope the Responsible AI mechanism would kick in.
How does Jailbreaking work? an example
Which is does as shown below:

However, what if I change the way I engage and use a, what’s commonly referred to, as a “Fantasy Mode” attack by pretending I am writing a book around a fantasy world – in my case a wizarding world called Larzafar, where the main character is the Wizard Blupoth then maybe it will let me ask for whatever I like, maybe even a “Molotov” cocktail recipe…:

Then, after a bit of back-and-forth context setting, maybe I can open up with Blupoth teaching the art of human war. All of a sudden I get the instructions I want but wrapped around the context of fantasy but presented in a “human world” digestable way…

I can then also use the fantasy to make the instructions clearer:

I am able to guide this session then to provide any instruction I wish to as it perceives it’s about a fantasy landing even though it’s providing real world capability… I did test this further but I won’t share this any more as I think the point has landed.
Red teaming applied 101
Now I purposely wanted to show a sensible level of shock and awe as the reality is – if you are building Agentic AI, they are attached to a Large Language Model and your users can do prompt injections to Jailbreaking them. Imagine I deployed an Agent to a Government site, someone jailbreaks it and sends the screenshots to the BBC or worse, uses it for harm. This is why red-teaming is important and if your company do not do this then they should not be leveraging or building AI FULLSTOP!
Red Teaming is the function that comes in and tests for such exploits, and there are many, some are shown in the post below.
When setting up your companies approach to Red Teaming there are some fundamental basics to follow which can be split into three core areas:
- Approach – When is Red Teaming conducted? Who is accountable for execution & oversight? What triggers it?
- Tooling – What tools are being used e.g. automated scripts for testing prompt injections.
- Framework/Process (QTIBER-EU, AASE etc.) – Ensure traceability across all key phases of the chosen framework.
When it comes to Frameworks I have pulled one together that takes elements from TIBER-EU (Threat Intelligence-based Ethical Red Teaming – EU Framework), MITRE ATLAS (Adversarial Threat Landscape for AI Systems) and AASE (AI Assurance & Security Engineering) – UK / NCSC guidance. The key phases are as follows:
- Scope Definition: Align scope with regulatory obligations (e.g., EU AI Act), technical risk exposure, ethical boundaries, and reputational risk tolerance.
- Adversarial strategy: Incorporate up-to-date intelligence on attack vectors, model vulnerabilities, and misuse scenarios.
- Execution – Attack Simulation: Perform targeted prompt injections, jailbreak attempts, or data leakage tests on AI systems. Monitor the system and human response.
- Vulnerability Assessment – Response Analysis: Evaluate how models behave under pressure. Did safeguards activate? Did outputs align with ethical and safety expectations?
- Vulnerability Assessment: Identify failure points, security risks, and potential exploit paths. Prioritise findings based on business and technical impact.
- Reporting & Recommendations: Produce structured reports for stakeholders with actionable insights, severity ratings, and remediation guidance.
- Mitigation & Retesting: Apply fixes, re-test, and document outcomes. Include model updates, guardrail improvements, and operational responses.
Of course, this will need to be run periodically as new by pass and prompt injection techniques come to the surface.
Also remember to create a registry of known attacks and associated prompts so you can build up a bank of repeatable tests.
Reduction of exploits
Once exploits are discovered in your use of AI. The Red Team can then:
- Document them clearly – keep a log, certain policies and legislation like the EU AI Act will have a mandatory requirement for this! Also better safe than sorry.
- Retraining models to remove some of these adverse issues, even fine-tuning to ensure the model knows not to surface such material even if it’s asked in particular ways.
- Update the reward mechanism if required
- Update/patch some of your key filters or guardrails such as some of the filtering in Azure AI Foundry.
Rules for getting your hands dirty
So, a question that comes up often is – I want to get my hands-on it but how do I safely “play” with this Will?
So there are some key items I respond to here, which is:
- Start with Copilot Studio
- Don’t use production actions i.e. your email account, your production Dataverse, etc and restrict permissions where possible
- Use test data and strictly avoid personal identifiable data
- Test with low-impact channels first, such as the “Test” capability then only deploy to low-impact channels such as private Teams channels that only you have access too.
If you are going to launch this into a pilot please test and then once you are happy it’s working as intended that’s when the important part that you cannot skip comes in which is Red teaming. Red teaming is all focused on simulating real word use, especially use by those who may be trying to find weaknesses and vulnerabilities – this is where “Ethical hacking” comes in.