← Index

io.github.hidai25/evalview-mcp

io.github.hidai25/evalview-mcp·v0.6.0·AI & LLMs
Quality Score
80
/100

Regression testing for AI agents. Golden baselines, CI/CD, LangGraph, CrewAI, OpenAI, Claude.

§01  Install
Claude Desktop (uvx)
{
  "mcpServers": {
    "evalview-mcp": {
      "command": "uvx",
      "args": [
        "evalview"
      ],
      "env": {
        "OPENAI_API_KEY": "<your-openai_api_key>"
      }
    }
  }
}
§02  Environment variables
OPENAI_API_KEY
secret

OpenAI API key for LLM-as-judge output quality scoring. Optional — deterministic tool/sequence evaluation works without it.

§03  MCP Quality Score  ·  methodology
freshness
25
completeness
10
installability
25
documentation
15
stability
5
§04  Alternatives in AI & LLMs