🛡️AI Verifier

Unifies the verifier docs into one page and switches to a GPT‑Vision‑mini–based approach:

  • Samples frames from the video

  • Sends a small set of frames to a vision‑capable LLM (e.g., gpt-4o-mini)

  • Asks the model to produce a strict JSON with the required metrics in percent and a final verdict

This keeps the interface stable while the intelligence lives in the prompt and the model.

Metrics schema

Required JSON keys (all integers in 0..100, final_pct also 0..100, verdict string):

{
  "accuracy": 0,
  "speed": 0,
  "safety": 0,
  "optimal_track": 0,
  "energy_efficiency": 0,
  "trajectory_stability": 0,
  "final_pct": 0,
  "verdict": "success | failure | inconclusive",
  "reasoning": "one‑paragraph short explanation"
}

Reference implementation (Python)

Requirements:

Notes:

  • We sample ~6 frames to keep token cost under control. Tune as needed.

  • Use JPEG at ~85% quality and width ≤ 640 px for a good cost/quality trade‑off.

  • If your provider requires a different schema for images (e.g., image_url), adapt the images payload.

Aggregation across verifiers

To emulate multiple verifiers (Alpha/Beta/Gamma), call verify_with_gpt_vision 3 times with different random frame subsets or slightly perturbed prompts (“Verifier Alpha perspective” etc.) and assemble a 2D metrics table like in the demo dashboard. The FINAL row can be the average per‑column or an LLM‑based consensus.

Last updated