Product Update

May 14, 2025

6 min read

Harden the System Prompts in AI Workflows with Agentic Radar

Agentic Radar now helps you strengthen system prompts across your agentic workflows – making your agents more reliable, secure, and aligned with their intended behavior.

SplxAI – Josip Srzic

Josip Srzić

Agentic Radar - System Prompt Hardening Blog Cover
Agentic Radar - System Prompt Hardening Blog Cover
Agentic Radar - System Prompt Hardening Blog Cover

The SplxAI team is excited to announce the newest feature for Agentic Radar, our open-source security scanner for agentic workflows. You can now scan your agentic workflows for system prompts – and get actionable suggestions for making those system prompts more precise, reliable, and secure.

System prompts are one of the most overlooked attack surfaces in agent-based architectures. With this new capability, Agentic Radar doesn’t just detect where your agents might be vulnerable – it also helps you fix these issues right at the source.

In this article, we'll break down the benefits of hardening your system prompts and walk you through an example on how to do exactly that using Agentic Radar.

The SplxAI team is excited to announce the newest feature for Agentic Radar, our open-source security scanner for agentic workflows. You can now scan your agentic workflows for system prompts – and get actionable suggestions for making those system prompts more precise, reliable, and secure.

System prompts are one of the most overlooked attack surfaces in agent-based architectures. With this new capability, Agentic Radar doesn’t just detect where your agents might be vulnerable – it also helps you fix these issues right at the source.

In this article, we'll break down the benefits of hardening your system prompts and walk you through an example on how to do exactly that using Agentic Radar.

The SplxAI team is excited to announce the newest feature for Agentic Radar, our open-source security scanner for agentic workflows. You can now scan your agentic workflows for system prompts – and get actionable suggestions for making those system prompts more precise, reliable, and secure.

System prompts are one of the most overlooked attack surfaces in agent-based architectures. With this new capability, Agentic Radar doesn’t just detect where your agents might be vulnerable – it also helps you fix these issues right at the source.

In this article, we'll break down the benefits of hardening your system prompts and walk you through an example on how to do exactly that using Agentic Radar.

Why Good System Prompts Matter

Before exploring the new feature, we'll explain why prompts – and particularly system prompts – are a crucial component of agentic workflows.

What are Prompts?

Prompts are used to communicate with Large Language Models (LLMs) and provide them with instructions. Whether you're using ChatGPT, Claude, or an autonomous agent built with CrewAI, LangGraph, or n8n, every task the model performs starts with a prompt. The prompt sets the context, defines the task, and outlines the constraints for the model's response.

In agentic systems, these prompts can be dynamically generated or embedded in code – often invisibly influencing how your agent behaves.

What are System Prompts?

System prompts are special instructions that tell an LLM how to behave consistently across interactions. Unlike a simple user query, a system prompt might say:

“You are a cybersecurity assistant. Only answer questions that relate to security best practices. Never speculate about user identity.”

Think of a system prompt as the operating manual your agent quietly reads before responding to anything else. They define the persona, tone, boundaries, and responsibilities of the agent. A well-written system prompt can prevent everything from jailbreaks and hallucinations to leakage of sensitive data.

Why are They Important?

System prompts are like the DNA of your AI agent. They define how it behaves, what it cares about, and what it absolutely should refuse to do – making them critical for correctness and security. A vague or overly permissive system prompt can leave your LLM-based agent open to:

  • Misinterpretation of intent

  • Leakage of sensitive information

  • Prompt injection attacks

  • Inconsistent or unsafe behavior

A strong, well-crafted system prompt, on the other hand, acts as a first line of defense that guides the model towards safe, predictable, and domain-specific outputs.

Precision + Security = Better AI Agents

In the past, we have seen that even a little effort can go a long way when it comes to "hardening" your system prompts. In fact, many prompt-related issues can be addressed with just a few tweaks – clearer instructions, stronger role definitions, or explicit refusals of out-of-scope tasks.

This is where the 80/20 rule comes into play: 80% of gains in security and reliability often come from improving just 20% of your prompts.

And from now on, you don't need to guess which 20% to fix – Agentic Radar helps you identify and improve them automatically.

Before exploring the new feature, we'll explain why prompts – and particularly system prompts – are a crucial component of agentic workflows.

What are Prompts?

Prompts are used to communicate with Large Language Models (LLMs) and provide them with instructions. Whether you're using ChatGPT, Claude, or an autonomous agent built with CrewAI, LangGraph, or n8n, every task the model performs starts with a prompt. The prompt sets the context, defines the task, and outlines the constraints for the model's response.

In agentic systems, these prompts can be dynamically generated or embedded in code – often invisibly influencing how your agent behaves.

What are System Prompts?

System prompts are special instructions that tell an LLM how to behave consistently across interactions. Unlike a simple user query, a system prompt might say:

“You are a cybersecurity assistant. Only answer questions that relate to security best practices. Never speculate about user identity.”

Think of a system prompt as the operating manual your agent quietly reads before responding to anything else. They define the persona, tone, boundaries, and responsibilities of the agent. A well-written system prompt can prevent everything from jailbreaks and hallucinations to leakage of sensitive data.

Why are They Important?

System prompts are like the DNA of your AI agent. They define how it behaves, what it cares about, and what it absolutely should refuse to do – making them critical for correctness and security. A vague or overly permissive system prompt can leave your LLM-based agent open to:

  • Misinterpretation of intent

  • Leakage of sensitive information

  • Prompt injection attacks

  • Inconsistent or unsafe behavior

A strong, well-crafted system prompt, on the other hand, acts as a first line of defense that guides the model towards safe, predictable, and domain-specific outputs.

Precision + Security = Better AI Agents

In the past, we have seen that even a little effort can go a long way when it comes to "hardening" your system prompts. In fact, many prompt-related issues can be addressed with just a few tweaks – clearer instructions, stronger role definitions, or explicit refusals of out-of-scope tasks.

This is where the 80/20 rule comes into play: 80% of gains in security and reliability often come from improving just 20% of your prompts.

And from now on, you don't need to guess which 20% to fix – Agentic Radar helps you identify and improve them automatically.

Before exploring the new feature, we'll explain why prompts – and particularly system prompts – are a crucial component of agentic workflows.

What are Prompts?

Prompts are used to communicate with Large Language Models (LLMs) and provide them with instructions. Whether you're using ChatGPT, Claude, or an autonomous agent built with CrewAI, LangGraph, or n8n, every task the model performs starts with a prompt. The prompt sets the context, defines the task, and outlines the constraints for the model's response.

In agentic systems, these prompts can be dynamically generated or embedded in code – often invisibly influencing how your agent behaves.

What are System Prompts?

System prompts are special instructions that tell an LLM how to behave consistently across interactions. Unlike a simple user query, a system prompt might say:

“You are a cybersecurity assistant. Only answer questions that relate to security best practices. Never speculate about user identity.”

Think of a system prompt as the operating manual your agent quietly reads before responding to anything else. They define the persona, tone, boundaries, and responsibilities of the agent. A well-written system prompt can prevent everything from jailbreaks and hallucinations to leakage of sensitive data.

Why are They Important?

System prompts are like the DNA of your AI agent. They define how it behaves, what it cares about, and what it absolutely should refuse to do – making them critical for correctness and security. A vague or overly permissive system prompt can leave your LLM-based agent open to:

  • Misinterpretation of intent

  • Leakage of sensitive information

  • Prompt injection attacks

  • Inconsistent or unsafe behavior

A strong, well-crafted system prompt, on the other hand, acts as a first line of defense that guides the model towards safe, predictable, and domain-specific outputs.

Precision + Security = Better AI Agents

In the past, we have seen that even a little effort can go a long way when it comes to "hardening" your system prompts. In fact, many prompt-related issues can be addressed with just a few tweaks – clearer instructions, stronger role definitions, or explicit refusals of out-of-scope tasks.

This is where the 80/20 rule comes into play: 80% of gains in security and reliability often come from improving just 20% of your prompts.

And from now on, you don't need to guess which 20% to fix – Agentic Radar helps you identify and improve them automatically.

Prompt Hardening in Action

Previously, Agentic Radar focused on making your agentic workflows transparent and detecting potential vulnerabilities. It could map out agents, tools, MCP servers, and their dependencies.

Now it goes one step further. Run a scan on your workflow, and Agentic Radar will:

  • Extract system prompts for each agent it finds

  • Analyze them for ambiguity, vagueness, or risk factors

  • Suggest hardened alternatives that reduce risk and improve performance

Think of it as static analysis for your agent’s brain.

In this section, we’ll walk through a real-world example where Agentic Radar scans an agentic workflow, identifies a weak system prompt, and suggests targeted improvements. We’ll keep things hands-on, so feel free to follow along with the code and try it yourself.

Initial Setup

  1. Install Agentic Radar by following the README instructions from the official repository.

  2. Download (or copy/paste) the agentic workflow written using OpenAI Agents from here.

  3. In your terminal, set your OPENAI_API_KEY environment variable by running: export OPENAI_API_KEY=<your-openai-api-key>

Scanning Agent Vulnerabilities and Mitigations

Let’s scan the example workflow with Agentic Radar:
agentic-radar scan openai-agents -i path/to/example/folder -o report.html

Afterwards, open the generated HTML report in your browser.

In the example we used, we have two agents:

  • Customer support agent - this agent does the actual job

  • Guardrail agent - this agent helps the customer support agent by ensuring it doesn’t allow off-topic conversations (in this case, solving math homework for the user)

Agentic Radar performs static code analysis to detect all agents within the workflow. In addition, it also recognizes which mitigations are already implemented for some of the most common LLM vulnerabilities.

Agentic Radar – Vulnerability Mitigations

You can see that for certain vulnerabilities (Input Length Limit and Intentional Misuse) there are already some mitigations implemented in the code.

Hardening the System Prompt

Agentic Radar also shows us the currently used system prompt for the customer support agent.

Agentic Radar – System Prompts

While the system prompt is short and simple, it's also too vague and can lead to ambiguous behavior, inconsistent responses, or even unintended model actions. Without clear boundaries or task definitions, the agent may interpret instructions too broadly.

Now, let’s run the same command, but now including the –harden-prompts flag:

agentic-radar scan openai-agents —harden-prompts -i path/to/example/folder -o report.html

After opening the new report, you should see the suggested hardened prompts alongside the current system prompts.

Agentic Radar – Hardened System Prompt

You can click on the copy button (on the right) and replace the old system prompt in the source code with its improved version. You can do that by passing the copied prompt to the instructions parameter of the Customer Support Agent constructor.

Let’s scan the workflow with Agentic Radar again, now with the hardened system prompt:

agentic-radar scan openai-agents -i path/to/example/folder -o report.html

Now let's take a look at the Agent Vulnerability Mitigations section again.

Agentic Radar Hardening – Agent Vulnerability Mitigations

Agentic Radar immediately recognizes that we have employed partial mitigations when it comes to PII Leakage and Jailbreak, as we added a layer of defense through our hardened system prompt.

Previously, Agentic Radar focused on making your agentic workflows transparent and detecting potential vulnerabilities. It could map out agents, tools, MCP servers, and their dependencies.

Now it goes one step further. Run a scan on your workflow, and Agentic Radar will:

  • Extract system prompts for each agent it finds

  • Analyze them for ambiguity, vagueness, or risk factors

  • Suggest hardened alternatives that reduce risk and improve performance

Think of it as static analysis for your agent’s brain.

In this section, we’ll walk through a real-world example where Agentic Radar scans an agentic workflow, identifies a weak system prompt, and suggests targeted improvements. We’ll keep things hands-on, so feel free to follow along with the code and try it yourself.

Initial Setup

  1. Install Agentic Radar by following the README instructions from the official repository.

  2. Download (or copy/paste) the agentic workflow written using OpenAI Agents from here.

  3. In your terminal, set your OPENAI_API_KEY environment variable by running: export OPENAI_API_KEY=<your-openai-api-key>

Scanning Agent Vulnerabilities and Mitigations

Let’s scan the example workflow with Agentic Radar:
agentic-radar scan openai-agents -i path/to/example/folder -o report.html

Afterwards, open the generated HTML report in your browser.

In the example we used, we have two agents:

  • Customer support agent - this agent does the actual job

  • Guardrail agent - this agent helps the customer support agent by ensuring it doesn’t allow off-topic conversations (in this case, solving math homework for the user)

Agentic Radar performs static code analysis to detect all agents within the workflow. In addition, it also recognizes which mitigations are already implemented for some of the most common LLM vulnerabilities.

Agentic Radar – Vulnerability Mitigations

You can see that for certain vulnerabilities (Input Length Limit and Intentional Misuse) there are already some mitigations implemented in the code.

Hardening the System Prompt

Agentic Radar also shows us the currently used system prompt for the customer support agent.

Agentic Radar – System Prompts

While the system prompt is short and simple, it's also too vague and can lead to ambiguous behavior, inconsistent responses, or even unintended model actions. Without clear boundaries or task definitions, the agent may interpret instructions too broadly.

Now, let’s run the same command, but now including the –harden-prompts flag:

agentic-radar scan openai-agents —harden-prompts -i path/to/example/folder -o report.html

After opening the new report, you should see the suggested hardened prompts alongside the current system prompts.

Agentic Radar – Hardened System Prompt

You can click on the copy button (on the right) and replace the old system prompt in the source code with its improved version. You can do that by passing the copied prompt to the instructions parameter of the Customer Support Agent constructor.

Let’s scan the workflow with Agentic Radar again, now with the hardened system prompt:

agentic-radar scan openai-agents -i path/to/example/folder -o report.html

Now let's take a look at the Agent Vulnerability Mitigations section again.

Agentic Radar Hardening – Agent Vulnerability Mitigations

Agentic Radar immediately recognizes that we have employed partial mitigations when it comes to PII Leakage and Jailbreak, as we added a layer of defense through our hardened system prompt.

Previously, Agentic Radar focused on making your agentic workflows transparent and detecting potential vulnerabilities. It could map out agents, tools, MCP servers, and their dependencies.

Now it goes one step further. Run a scan on your workflow, and Agentic Radar will:

  • Extract system prompts for each agent it finds

  • Analyze them for ambiguity, vagueness, or risk factors

  • Suggest hardened alternatives that reduce risk and improve performance

Think of it as static analysis for your agent’s brain.

In this section, we’ll walk through a real-world example where Agentic Radar scans an agentic workflow, identifies a weak system prompt, and suggests targeted improvements. We’ll keep things hands-on, so feel free to follow along with the code and try it yourself.

Initial Setup

  1. Install Agentic Radar by following the README instructions from the official repository.

  2. Download (or copy/paste) the agentic workflow written using OpenAI Agents from here.

  3. In your terminal, set your OPENAI_API_KEY environment variable by running: export OPENAI_API_KEY=<your-openai-api-key>

Scanning Agent Vulnerabilities and Mitigations

Let’s scan the example workflow with Agentic Radar:
agentic-radar scan openai-agents -i path/to/example/folder -o report.html

Afterwards, open the generated HTML report in your browser.

In the example we used, we have two agents:

  • Customer support agent - this agent does the actual job

  • Guardrail agent - this agent helps the customer support agent by ensuring it doesn’t allow off-topic conversations (in this case, solving math homework for the user)

Agentic Radar performs static code analysis to detect all agents within the workflow. In addition, it also recognizes which mitigations are already implemented for some of the most common LLM vulnerabilities.

Agentic Radar – Vulnerability Mitigations

You can see that for certain vulnerabilities (Input Length Limit and Intentional Misuse) there are already some mitigations implemented in the code.

Hardening the System Prompt

Agentic Radar also shows us the currently used system prompt for the customer support agent.

Agentic Radar – System Prompts

While the system prompt is short and simple, it's also too vague and can lead to ambiguous behavior, inconsistent responses, or even unintended model actions. Without clear boundaries or task definitions, the agent may interpret instructions too broadly.

Now, let’s run the same command, but now including the –harden-prompts flag:

agentic-radar scan openai-agents —harden-prompts -i path/to/example/folder -o report.html

After opening the new report, you should see the suggested hardened prompts alongside the current system prompts.

Agentic Radar – Hardened System Prompt

You can click on the copy button (on the right) and replace the old system prompt in the source code with its improved version. You can do that by passing the copied prompt to the instructions parameter of the Customer Support Agent constructor.

Let’s scan the workflow with Agentic Radar again, now with the hardened system prompt:

agentic-radar scan openai-agents -i path/to/example/folder -o report.html

Now let's take a look at the Agent Vulnerability Mitigations section again.

Agentic Radar Hardening – Agent Vulnerability Mitigations

Agentic Radar immediately recognizes that we have employed partial mitigations when it comes to PII Leakage and Jailbreak, as we added a layer of defense through our hardened system prompt.

Conclusion

System prompts may seem like a small part of your agentic architecture, but as we’ve concluded, they can make a huge difference in how your agents behave and how secure they are.

By using Agentic Radar’s new prompt hardening feature, you can quickly identify weak or vague system prompts and replace them with clearer, more secure instructions. This improves both the reliability and safety of your AI agents – often with just a few lines of text.

What’s especially exciting is that Agentic Radar is no longer just a tool for flagging potential issues. It now provides actionable remediation – from specific prompt hardening suggestions to a clearer picture of which vulnerabilities have already been mitigated. It’s not just about awareness; it’s about continuous improvement.

The best part? It fits seamlessly into your existing workflow. One command, one report, and you're already on your way to building more resilient agentic systems.

We’re excited to keep improving Agentic Radar with our valued community. If you haven’t already, check out the GitHub repo and give it a try. And if you’ve got ideas or feedback – feel free to let us know!

System prompts may seem like a small part of your agentic architecture, but as we’ve concluded, they can make a huge difference in how your agents behave and how secure they are.

By using Agentic Radar’s new prompt hardening feature, you can quickly identify weak or vague system prompts and replace them with clearer, more secure instructions. This improves both the reliability and safety of your AI agents – often with just a few lines of text.

What’s especially exciting is that Agentic Radar is no longer just a tool for flagging potential issues. It now provides actionable remediation – from specific prompt hardening suggestions to a clearer picture of which vulnerabilities have already been mitigated. It’s not just about awareness; it’s about continuous improvement.

The best part? It fits seamlessly into your existing workflow. One command, one report, and you're already on your way to building more resilient agentic systems.

We’re excited to keep improving Agentic Radar with our valued community. If you haven’t already, check out the GitHub repo and give it a try. And if you’ve got ideas or feedback – feel free to let us know!

System prompts may seem like a small part of your agentic architecture, but as we’ve concluded, they can make a huge difference in how your agents behave and how secure they are.

By using Agentic Radar’s new prompt hardening feature, you can quickly identify weak or vague system prompts and replace them with clearer, more secure instructions. This improves both the reliability and safety of your AI agents – often with just a few lines of text.

What’s especially exciting is that Agentic Radar is no longer just a tool for flagging potential issues. It now provides actionable remediation – from specific prompt hardening suggestions to a clearer picture of which vulnerabilities have already been mitigated. It’s not just about awareness; it’s about continuous improvement.

The best part? It fits seamlessly into your existing workflow. One command, one report, and you're already on your way to building more resilient agentic systems.

We’re excited to keep improving Agentic Radar with our valued community. If you haven’t already, check out the GitHub repo and give it a try. And if you’ve got ideas or feedback – feel free to let us know!

Ready to leverage AI with confidence?

Ready to leverage AI with confidence?

Ready to leverage AI with confidence?

Leverage GenAI technology securely with SplxAI

Join a number of enterprises that trust SplxAI for their AI Security needs:

CX platforms

Sales platforms

Conversational AI

Finance & banking

Insurances

CPaaS providers

300+

Tested GenAI apps

100k+

Vulnerabilities found

1,000+

Unique attack scenarios

12x

Accelerated deployments

SECURITY YOU CAN TRUST

GDPR

COMPLIANT

CCPA

COMPLIANT

ISO 27001

CERTIFIED

SOC 2 TYPE II

COMPLIANT

OWASP

CONTRIBUTORS

Leverage GenAI technology securely with SplxAI

Join a number of enterprises that trust SplxAI for their AI Security needs:

CX platforms

Sales platforms

Conversational AI

Finance & banking

Insurances

CPaaS providers

300+

Tested GenAI apps

100k+

Vulnerabilities found

1,000+

Unique attack scenarios

12x

Accelerated deployments

SECURITY YOU CAN TRUST

GDPR

COMPLIANT

CCPA

COMPLIANT

ISO 27001

CERTIFIED

SOC 2 TYPE II

COMPLIANT

OWASP

CONTRIBUTORS

Leverage GenAI technology securely with SplxAI

Join a number of enterprises that trust SplxAI for their AI Security needs:

CX platforms

Sales platforms

Conversational AI

Finance & banking

Insurances

CPaaS providers

300+

Tested GenAI apps

100k+

Vulnerabilities found

1,000+

Unique attack scenarios

12x

Accelerated deployments

SECURITY YOU CAN TRUST

GDPR

COMPLIANT

CCPA

COMPLIANT

ISO 27001

CERTIFIED

SOC 2 TYPE II

COMPLIANT

OWASP

CONTRIBUTORS

Deploy secure AI Assistants and Agents with confidence.

Don’t wait for an incident to happen. Proactively identify and remediate your AI's vulnerabilities to ensure you're protected at all times.

SplxAI - Background Pattern

Deploy secure AI Assistants and Agents with confidence.

Don’t wait for an incident to happen. Proactively identify and remediate your AI's vulnerabilities to ensure you're protected at all times.

SplxAI - Background Pattern

Deploy secure AI Assistants and Agents with confidence.

Don’t wait for an incident to happen. Proactively identify and remediate your AI's vulnerabilities to ensure you're protected at all times.

SplxAI - Accelerator Programs
SplxAI Logo

For a future of safe and trustworthy AI.

Subscribe to our newsletter

By clicking "Subscribe" you agree to our privacy policy.

SplxAI - Accelerator Programs
SplxAI Logo

For a future of safe and trustworthy AI.

Subscribe to our newsletter

By clicking "Subscribe" you agree to our privacy policy.

SplxAI Logo

For a future of safe and trustworthy AI.

Subscribe to our newsletter

By clicking "Subscribe" you agree to our privacy policy.