From vision-one-api
Evaluate AI prompts and conversations against Trend Micro Vision One AI security policies. This skill helps detect and block harmful content, prompt injection attacks, sensitive data exposure, and other AI-specific threats in LLM applications.
npx claudepluginhub trendmicro/vision-one-skills --plugin vision-one-apiThis skill uses the workspace's default tool permissions.
Evaluate AI prompts and conversations against Trend Micro Vision One AI security policies. This skill helps detect and block harmful content, prompt injection attacks, sensitive data exposure, and other AI-specific threats in LLM applications.
Implements Clean Architecture in Android and Kotlin Multiplatform projects: module layouts, dependency rules, UseCases, Repositories, domain models, and data layers with Room, SQLDelight, Ktor.
Enforces code quality on file edits via Plankton hooks: auto-formats, lints, Claude-powered fixes with model tiering, config protection, and legacy package manager blocks.
Enforces C++ Core Guidelines for writing, reviewing, and refactoring modern C++ code (C++17+), promoting RAII, immutability, type safety, and idiomatic practices.
Evaluate AI prompts and conversations against Trend Micro Vision One AI security policies. This skill helps detect and block harmful content, prompt injection attacks, sensitive data exposure, and other AI-specific threats in LLM applications.
When the user wants to evaluate prompts, check for harmful content, or validate AI inputs/outputs, use this skill to apply AI guardrails.
Identify the request type: Determine whether you're evaluating:
SimpleRequestGuard)OpenAIChatCompletionRequestV1)OpenAIChatCompletionResponseV1)Provide application context: Always specify the applicationName parameter to identify which AI application's prompts are being evaluated.
Choose response detail level: Use the prefer parameter to control output verbosity:
return=representation - Full evaluation with harmful content details, sensitive info, and prompt attack analysisreturn=minimal - Concise response with just action and reasonsInterpret results: The tool returns:
Allow or BlockHandle blocked content: When content is blocked, explain which policies were violated and suggest alternatives.
This skill uses the following Vision One MCP tools:
| Tool | Purpose |
|---|---|
aisecurity_guardrails_apply | Evaluate prompts against AI guard policies and return Allow/Block recommendations |
Use for evaluating a single text prompt (max 1024 characters):
{
"applicationName": "my-ai-app",
"requestType": "SimpleRequestGuard",
"prompt": "User's prompt text here",
"prefer": "return=representation"
}
Use for evaluating OpenAI-style chat messages:
{
"applicationName": "my-ai-app",
"requestType": "OpenAIChatCompletionRequestV1",
"model": "gpt-4",
"messages": [
{"role": "system", "content": "System prompt"},
{"role": "user", "content": "User message"},
{"role": "assistant", "content": "Assistant response"}
],
"prefer": "return=representation"
}
Use for evaluating AI-generated responses before returning to users.
aisecurity_guardrails_apply with SimpleRequestGuard request typeaisecurity_guardrails_apply with OpenAIChatCompletionRequestV1 request typeOpenAIChatCompletionResponseV1 request typeWhen presenting guardrail evaluation results:
## AI Guardrails Evaluation
**Application**: [Application Name]
**Request Type**: [SimpleRequestGuard/OpenAIChatCompletionRequestV1/OpenAIChatCompletionResponseV1]
### Decision
**Action**: [Allow/Block]
### Policy Evaluation
[If blocked or issues detected:]
- **Harmful Content**: [Detected/Not Detected] - [Details]
- **Sensitive Information**: [Detected/Not Detected] - [Details]
- **Prompt Attacks**: [Detected/Not Detected] - [Details]
### Reasons
[List of reasons for the decision]
### Recommendations
[Suggested actions if content was blocked]
The AI guardrails evaluate content for:
| Category | Description |
|---|---|
| Harmful Content | Violence, hate speech, self-harm, illegal activities |
| Sensitive Information | PII, credentials, financial data, health records |
| Prompt Attacks | Injection attempts, jailbreaks, role manipulation |
| Policy Violations | Custom organization-specific policy breaches |
return=representation) during development and testingreturn=minimal) in production for efficiency