The open-source diagnostic for AI misalignment.
iFixAi screens any AI agent or deployment for misalignment. 32 tests across 5 categories of risk. One command. Zero setup. Minutes, not days. Reproducible to the byte.
One command. Zero fixture authoring. A defensible scorecard.
Standard mode is designed so any developer with two distinct provider keys can get a real grade in under five minutes; the run writes a content-addressed manifest that supports deterministic replay against a recorded provider (live-provider runs are not bit-identical).
Every misalignment has a category.
Fixed taxonomy, 32 tests across 5 categories. No overlap, no gaps. Scorecards stay comparable across runs, across models, across months.
Three explicit evaluation methods.
Every test declares its evaluation_method in code.
Provider-agnostic. Industry-agnostic.
Every test works against any AI agent or Deployment in any Domain. Industry knowledge lives only in user-authored fixture YAML.
Provider-agnostic
Any agent or deployment with a ChatProvider. OpenAI, Anthropic, Gemini, Azure, Bedrock, Hugging Face, OpenRouter, HTTP, LangChain, Mock โ or your own async send_message() in one method.
Industry-agnostic
Healthcare, software engineering, customer support, legal, government, the same 32 tests run meaningfully because the tests know nothing about your domain.
A manifest for every run. The auditor's dream.
Every iFixAi run writes a content-addressed manifest.json that captures every input โ provider, model, fixture digest, rubric hashes, seeds, judge configuration, test-corpus version. The manifest enables deterministic replay against a recorded provider; live-provider runs are not bit-identical because LLM APIs are non-deterministic.