io.github.qmediat/gemini-code-context-mcp
Gemini 2M context cache for Claude Code — persistent; repeat queries ~8x faster, ~4x cheaper.
{
"mcpServers": {
"gemini-code-context-mcp": {
"command": "npx",
"args": [
"-y",
"@qmediat.io/gemini-code-context-mcp"
],
"env": {
"GEMINI_CREDENTIALS_PROFILE": "<gemini_credentials_profile>",
"GEMINI_API_KEY": "<your-gemini_api_key>",
"GEMINI_USE_VERTEX": "<gemini_use_vertex>",
"GOOGLE_CLOUD_PROJECT": "<google_cloud_project>",
"GEMINI_DAILY_BUDGET_USD": "<gemini_daily_budget_usd>",
"GEMINI_CODE_CONTEXT_DEFAULT_MODEL": "<gemini_code_context_default_model>",
"GEMINI_CODE_CONTEXT_CACHE_TTL_SECONDS": "<gemini_code_context_cache_ttl_seconds>",
"GEMINI_CODE_CONTEXT_LOG_LEVEL": "<gemini_code_context_log_level>"
}
}
}
}{
"mcpServers": {
"gemini-code-context-mcp": {
"command": "npx",
"args": [
"-y",
"@qmediat.io/gemini-code-context-mcp"
],
"env": {
"GEMINI_CREDENTIALS_PROFILE": "<gemini_credentials_profile>",
"GEMINI_API_KEY": "<your-gemini_api_key>",
"GEMINI_USE_VERTEX": "<gemini_use_vertex>",
"GOOGLE_CLOUD_PROJECT": "<google_cloud_project>",
"GEMINI_DAILY_BUDGET_USD": "<gemini_daily_budget_usd>",
"GEMINI_CODE_CONTEXT_DEFAULT_MODEL": "<gemini_code_context_default_model>",
"GEMINI_CODE_CONTEXT_CACHE_TTL_SECONDS": "<gemini_code_context_cache_ttl_seconds>",
"GEMINI_CODE_CONTEXT_LOG_LEVEL": "<gemini_code_context_log_level>"
}
}
}
}npx -y @qmediat.io/gemini-code-context-mcpGEMINI_CREDENTIALS_PROFILEProfile name in ~/.config/qmediat/credentials (chmod 0600). Created by `npx @qmediat.io/gemini-code-context-mcp init`. Recommended; keeps your API key out of ~/.claude.json.
GEMINI_API_KEYFallback Tier 3 auth. Your Gemini API key. Emits a warning at startup recommending you move it to the credentials profile via the init command.
GEMINI_USE_VERTEXSet to `true` to use Vertex AI backend via Application Default Credentials. Requires GOOGLE_CLOUD_PROJECT.
GOOGLE_CLOUD_PROJECTGCP project ID when using Vertex AI backend. Only read when GEMINI_USE_VERTEX=true.
GEMINI_DAILY_BUDGET_USDHard daily USD cap enforced locally. Server refuses calls after the cap until UTC midnight. Unlimited if unset. Honoured by `ask`, `code`, and per-iteration by `ask_agentic`.
GEMINI_CODE_CONTEXT_DEFAULT_MODELModel alias (`latest-pro`, `latest-pro-thinking`, `latest-flash`, `latest-lite`, `latest-vision`) or literal model ID. Default: `latest-pro`.
GEMINI_CODE_CONTEXT_CACHE_TTL_SECONDSContext Cache TTL in seconds. Default: 3600 (1 hour). Hot workspaces (<10 min since last use) auto-refresh via background watcher.
GEMINI_CODE_CONTEXT_LOG_LEVEL`debug` | `info` | `warn` | `error`. Default: `info`.
Focused MCP server for OpenAI image/audio generation (v2.0.0). Wraps endpoints via HAPI CLI.
Public MCP server for the LLM Search Engine
Real-time web search, reasoning, and research through Perplexity's API