ask-gemini-mcp
MCP server that enables AI assistants to interact with Google Gemini CLI
Ask Gemini MCP
An MCP server for AI-to-AI collaboration via the Gemini CLI. Available on npm: ask-gemini-mcp. Works with Claude Code, Claude Desktop, Cursor, Warp, Copilot, and 40+ other MCP clients. Leverage Gemini's massive 1M+ token context window for large file and codebase analysis while your primary AI handles interaction and code editing.
Why?
- Get a second opinion — Ask Gemini to review your coding approach before committing to it
- Debate plans — Send architecture proposals to Gemini for critique and alternative suggestions
- Review changes — Have Gemini analyze diffs or modified files to catch issues your primary AI might miss
- Massive context — Gemini reads entire codebases (1M+ tokens) that would overflow other models
Quick Start
Claude Code
# Project scope (available in current project only)
claude mcp add gemini-cli -- npx -y ask-gemini-mcp
# User scope (available across all projects)
claude mcp add --scope user gemini-cli -- npx -y ask-gemini-mcp
Claude Desktop
Add to your config file (~/Library/Application Support/Claude/claude_desktop_config.json on macOS):
{
"mcpServers": {
"gemini-cli": {
"command": "npx",
"args": ["-y", "ask-gemini-mcp"]
}
}
}
Other config file locations
- Windows:
%APPDATA%\Claude\claude_desktop_config.json - Linux:
~/.config/claude/claude_desktop_config.json
Cursor
Add to .cursor/mcp.json in your project (or ~/.cursor/mcp.json for global):
{
"mcpServers": {
"gemini-cli": {
"command": "npx",
"args": ["-y", "ask-gemini-mcp"]
}
}
}
Codex CLI
Add to ~/.codex/config.toml (or .codex/config.toml in your project):
[mcp_servers.gemini-cli]
command = "npx"
args = ["-y", "ask-gemini-mcp"]
Or via CLI:
codex mcp add gemini-cli -- npx -y ask-gemini-mcp
OpenCode
Add to opencode.json in your project (or ~/.config/opencode/opencode.json for global):
{
"mcp": {
"gemini-cli": {
"type": "local",
"command": ["npx", "-y", "ask-gemini-mcp"]
}
}
}
Any MCP Client (STDIO Transport)
{
"transport": {
"type": "stdio",
"command": "npx",
"args": ["-y", "ask-gemini-mcp"]
}
}
Prerequisites
- Node.js v20.0.0 or higher (LTS)
- Google Gemini CLI installed and authenticated
Tools
| Tool | Purpose |
|---|---|
ask-gemini | Send prompts to Gemini CLI. Supports @ file syntax, model selection, sandbox mode, and changeMode for structured edits |
fetch-chunk | Retrieve subsequent chunks from cached large responses |
ping | Connection test — verify MCP setup without using Gemini tokens |
Usage Examples
File analysis (@ syntax):
ask gemini to analyze @src/main.js and explain what it doesuse gemini to summarize @. the current directory
Code review:
ask gemini to review the changes in @src/auth.ts for security issuesuse gemini to compare @old.js and @new.js
General questions:
ask gemini about best practices for React state management
Sandbox mode:
use gemini sandbox to create and run a Python script
Models
| Model | Use Case |
|---|---|
gemini-3.1-pro-preview | Default — best quality reasoning |
gemini-3-flash-preview | Faster responses, large codebases |
The server automatically falls back to Flash when Pro quota is exceeded.
Contributing
Contributions are welcome! See open issues for things to work on.
License
MIT License. See LICENSE for details.
Disclaimer: This is an unofficial, third-party tool and is not affiliated with, endorsed, or sponsored by Google.
Serveurs connexes
Scout Monitoring MCP
sponsorPut performance and error data directly in the hands of your AI assistant.
Alpha Vantage MCP Server
sponsorAccess financial market data: realtime & historical stock, ETF, options, forex, crypto, commodities, fundamentals, technical indicators, & more
GitHub Workflow Debugger MCP
Diagnose and fix GitHub Actions workflow failures using the GitHub API.
OPNsense MCP Server
A comprehensive MCP server for managing OPNsense firewalls, offering over 300 tools for configuration and monitoring.
PolyMarket
Access prediction market data from the PolyMarket API.
Command Executor
Execute pre-approved shell commands securely on a server.
Knowledge Graph Memory Server
Enables persistent memory for Claude using a local knowledge graph of entities, relations, and observations.
DHTMLX MCP Server
The DHTMLX MCP server provides real-time access to official docs and a smart RAG.
Screeny
A macOS-only server that enables LLMs to capture screenshots of specific application windows, providing visual context for development and debugging.
browser-devtools-mcp
A Playwright-based MCP server that exposes a live browser as a traceable, inspectable, debuggable and controllable execution environment for AI agents.
return0
Debug your remote Node.js and Next.js applications directly from your AI IDE like Cursor.
Vibetest Use
A browser-agent QA swarm with an MCP interface for testing AI-generated websites.