pip install.
Quick Start
Install InferenceWall and scan your first input in under 5 minutes.
Deployment Profiles
Choose between Lite, Standard, and Full profiles based on your latency and accuracy needs.
API Reference
Explore the REST API for scanning inputs and outputs via HTTP.
Integrations
Wrap OpenAI, Anthropic, LangChain, and FastAPI with InferenceWall in minutes.
How it works
InferenceWall runs a multi-layer detection pipeline on every scan request. Each layer uses a different technique — pattern matching, ML classification, semantic similarity, or LLM judgment — and contributes to a single anomaly score. When the score crosses a threshold, InferenceWall flags or blocks the content.Install
Install InferenceWall from PyPI. The
lite profile has zero ML dependencies and runs with sub-millisecond latency.Scan inputs
Call
inferwall.scan_input() before forwarding user prompts to your LLM. Check the decision field to allow, flag, or block.Scan outputs
Call
inferwall.scan_output() before returning LLM responses to users. InferenceWall catches PII, API keys, and other sensitive data leakage.Detection capabilities
InferenceWall ships with 100 signatures covering five threat categories, all mapped to the MITRE ATLAS adversarial AI framework.Prompt Injection
Direct and indirect injection, jailbreaks, persona hijacking, obfuscated payloads (base64, ROT13, homoglyphs).
Data Leakage
PII detection, API key and credential exposure, training data exfiltration.
Content Safety
Toxicity, hate speech, violence, self-harm, and other harmful content categories.
Agentic Threats
Tool abuse, context poisoning, host escape attempts, and exfiltration via agent actions.