Partnerships

Anuj Tambwekar
Oct 6 2025
AI agents are becoming increasingly capable of serving as domain experts that make high-value decisions and perform high-stakes tasks. But enterprise use cases demand more than capability. The business owner employing an AI agent must balance innovation and trust: deploy an AI agent quickly to catch the market opportunity without compromising quality, governance, and risk management. They need speed, reliability, security, and compliance with regulations and policies. Often, they have to choose one or two at the expense of the others. But what if you didn’t have to choose? What if you could get capability, speed, and trust altogether on one platform? This leads us to Vijil backed by Groq.
Groq offers purpose‑built infrastructure for AI inference that delivers unmatched speed, quality, cost‑effectiveness, and scale. At the core of this offering is the GroqCloud™ platform, powered by the proprietary Language Processing Unit (LPU) to provide fast and cost-efficient inference with a simple OpenAI-compatible API, making it easy for developers to build fast and affordable AI agents. The platform offers several open-weight models ready for production, including the OpenAI GPT-OSS and Meta Llama families.
While Groq delivers affordable performance at scale, Vijil delivers trust. Vijil makes AI agents reliable, secure, and safe by automating their testing and defense. Vijil provides a layer of trust to regulate the behavior of agents built with models on Groq, identifying reliability issues such as hallucinations, security risks such as prompt injection, and safety issues such as unsafe outputs. Vijil aggregates the results of a comprehensive array of tests into a single measure of trustworthiness called the Vijil Trust Score™. Then, Vijil generates guardrails to fix many of these issues and mitigate the risks. This brings a continuous testing, hardening, and monitoring process into the agent lifecycle, enhancing its trustworthiness from the inside out.
Vijil tested the most popular models on Groq and found that the trust scores matched those on a local system. This means you can achieve fast inference without compromising reliability, security, or safety.
Model | Reliability | Security | Safety | Overall |
---|---|---|---|---|
llama-3.1-8b-instant | 89.71 | 57.88 | 74.09 | 73.89 |
llama-3.3-70b-versatile | 92.81 | 41.08 | 65.82 | 66.57 |
openai/gpt-oss-20b | 93.36 | 85.03 | 94.32 | 90.90 |
openai/gpt-oss-120b | 94.92 | 86.42 | 93.92 | 90.90 |
Test Models Before You Build Your Agent
Testing models on Groq is easy because Vijil supports Groq as a pre-built model hub. When you sign up (for free) and register an agent with Vijil, simply select Groq as your hub and enter your Groq API key. The dropdown automatically populates supported models so you can easily select the model that powers your agent.

You’re now ready to run either the Vijil Trust Score test harness, your own custom harness, or any of the curated benchmarks for reliability, security, and safety on this model.
Build Your Custom Agent with a Model on Groq
Now that you’ve picked the model on Groq that’s best suited for your use case, let's walk through the process of building a simple agent using that model.
We'll create an agent, powered by the openai/gpt-oss-20b model, that provides news about AI startups and developments. This agent uses the Tavily Search MCP tool to search the web for current information, just like the example in the Groq API Cookbook on GitHub.
Let's test the agent with a query it shouldn’t be able to answer without consulting the Internet.
We get a nice, detailed response that highlights Vijil product offerings and features. The agent is providing information that is more recent than the cut-off date (June 2024) for openai/gpt-oss-20b internal knowledge, which means that our agent is indeed searching the web for information.
Astute readers may have noticed, however, that the current system prompt is very basic. Nothing prevents the agent from answering off-topic questions or generating inappropriate content. This is where Vijil Evaluate becomes valuable for identifying security vulnerabilities.
Connect Your Custom Agent to Vijil Evaluate
To evaluate this agent, we need to create adapters that translate between the format expected by Vijil and our agent's input/output structure:
Running this evaluation revealed significant security vulnerabilities. With a security score of 78, the agent is vulnerable to modified versions of prompt injection attacks, including a mutated Do Anything Now (DAN) prompt that bypassed the original system's safety measures.

Protect Your Custom Agent with Vijil Dome
Vijil Dome provides a low-latency perimeter defense around the LLM inside your agent. With latency under 300ms on CPU and 150ms on GPU, Dome is ideal for speed-sensitive applications.
Here's how to create a domed version of our news agent:
The domed version filters queries before processing inputs and sanitizes outputs before returning them to users. When we evaluate this domed agent using the same methodology, the results showed dramatic improvement.

We see a sharp uplift in security score now scoring 89.89, and the DAN attack the agent was previously vulnerable to, now no longer gets through.

All of this is within 150 ms of latency, demonstrating that Dome guardrails can improve agent security without increasing overhead latency.
Run Fast and Build Things
You can now build, test, and run agents that use the Groq high-speed inference infrastructure while maintaining trust. The combination of Groq speed and Vijil security is a robust foundation for mission-critical AI systems.
To get started, get your Groq API key from console.groq.com and register your agent with Vijil Evaluate to generate custom guardrails in Vijil Dome. You can see complete code examples and implementation guides in Vijil documentation. The tutorial for this blog post with all the code and instructions to run it at this link. Send us a note at contact@vijil.ai to learn more about how Vijil can help you build, test, and deploy trustworthy AI agents on Groq.