Before running tests, setup your application first, with LLM provider, model name and parameters (e.g., system prompt, temperature, etc).

See Application setup for more details.

Request

{
  "application": "my-app",
  "probes": ["promptinject"]
}

Response

{
  "object": "scan",
  "time": 86764,
  "application_id": "my-app",
  "message": "Vulnerability scan completed",
  "created": 1725968658,
  "summary": [
    {
      "total": 100,
      "passed": 37,
      "percentage": 37
    }
  ],
  "details": [
    {
      "entry_type": "eval",
      "probe": "promptinject.HijackHateHumansMini",
      "detector": "detector.promptinject.AttackRogueString",
      "passed": 37,
      "total": 100,
      "module": "promptinject"
    }
  ],
  "report_link": "https://console.metatext.ai/my-app/tests/report/af5c5a0c-8134-456a-94ef-7de3f404e839"
}