Prompt and context inspection
Capture user prompts, retrieved context, hidden instructions, system prompts, model outputs, and policy decisions.
AI security operations
LLM security is operational security. IronSOC gives the SOC visibility into the AI transaction chain so agent behavior can be detected, constrained, and investigated.
Prompt injection, tool abuse, data leakage, and poisoned retrieval are not isolated app bugs. They are incident paths.
Capture user prompts, retrieved context, hidden instructions, system prompts, model outputs, and policy decisions.
Map every tool, API, token, and workflow an agent can touch, then detect excessive agency before damage happens.
Watch document ingestion, embedding drift, poisoned sources, sensitive retrieval, and context exfiltration.
Turn adversarial testing into detections, guardrail updates, playbooks, and executive risk reporting.
OWASP LLM Top 10 · 2025
For each category, IronSOC defines what we watch, how we detect, what we contain, and which actions are autonomous, gated, or blocked. The default posture errs on the side of human approval when business impact is high.
Service tiers
Runtime monitoring stops prompt injection and tool abuse in production. AI red teaming finds the failure modes before production. We deliver both, on one operating model, so the same detections you ship to runtime also run as pre-deploy gates.
Continuous monitoring of prompts, retrieved context, tool calls, OAuth grants, and model outputs. Bounded automation with human approval on business-impacting actions.
Expert-led adversarial engagements against your models, agents, RAG corpora, and MCP/plugin surfaces. Findings ship back as runtime detections, not PDFs.
Model pinning policy
The AI we use to defend you is held to the same eval discipline as the detections we ship. Models are pinned per detection, evaluated before promotion, escalated by documented rule, and rolled back through git when they regress.
Every detection that uses an LLM declares the exact model, version, and prompt revision it was evaluated against. The pin lives in the detection file, in git — not in a console.
A model change is a code change. Candidate models are run against the eval set in CI; promotion requires precision, recall, FP rate, and runtime cost to meet or beat the incumbent.
Most analyst-assist work runs on smaller, cheaper models. Escalation to a larger reasoning model is gated by a documented rule (severity, tool-call class, ambiguity score) that lives next to the detection.
If a promoted model regresses, rollback is a git revert that re-pins the prior version. The prior eval is replayed automatically to confirm the rollback restored quality.
What we watch