ask-gemini-mcp
MCP server that enables AI assistants to interact with Google Gemini CLI
Ask LLM
| Package | Type | Version | Downloads |
|---|---|---|---|
ask-gemini-mcp | MCP Server | ||
ask-codex-mcp | MCP Server | ||
ask-ollama-mcp | MCP Server | ||
ask-llm-mcp | MCP Server | ||
@ask-llm/plugin | Claude Code Plugin | /plugin install |
MCP servers + Claude Code plugin for AI-to-AI collaboration
MCP servers that bridge your AI client with multiple LLM providers for AI-to-AI collaboration. Works with Claude Code, Claude Desktop, Cursor, Warp, Copilot, and 40+ other MCP clients. Leverage Gemini's 1M+ token context, Codex's GPT-5.4, or local Ollama models — all via standard MCP.
Why?
- Get a second opinion — Ask another AI to review your coding approach before committing
- Debate plans — Send architecture proposals for critique and alternative suggestions
- Review changes — Have multiple AIs analyze diffs to catch issues your primary AI might miss
- Massive context — Gemini reads entire codebases (1M+ tokens) that would overflow other models
- Local & private — Use Ollama for reviews where no data leaves your machine
Quick Start
Claude Code
# All-in-one — auto-detects installed providers
claude mcp add --scope user ask-llm -- npx -y ask-llm-mcp
Or install providers individually
claude mcp add --scope user gemini -- npx -y ask-gemini-mcp
claude mcp add --scope user codex -- npx -y ask-codex-mcp
claude mcp add --scope user ollama -- npx -y ask-ollama-mcp
Claude Desktop
Add to claude_desktop_config.json:
{
"mcpServers": {
"ask-llm": {
"command": "npx",
"args": ["-y", "ask-llm-mcp"]
}
}
}
Or install providers individually
{
"mcpServers": {
"gemini": {
"command": "npx",
"args": ["-y", "ask-gemini-mcp"]
},
"codex": {
"command": "npx",
"args": ["-y", "ask-codex-mcp"]
},
"ollama": {
"command": "npx",
"args": ["-y", "ask-ollama-mcp"]
}
}
}
Cursor, Codex CLI, OpenCode, and other clients
Cursor (.cursor/mcp.json):
{
"mcpServers": {
"ask-llm": { "command": "npx", "args": ["-y", "ask-llm-mcp"] }
}
}
Codex CLI (~/.codex/config.toml):
[mcp_servers.ask-llm]
command = "npx"
args = ["-y", "ask-llm-mcp"]
Any MCP Client (STDIO transport):
{ "command": "npx", "args": ["-y", "ask-llm-mcp"] }
Replace ask-llm-mcp with ask-gemini-mcp, ask-codex-mcp, or ask-ollama-mcp for a single provider.
Claude Code Plugin
The Ask LLM plugin adds multi-provider code review, brainstorming, and automated hooks directly into Claude Code:
/plugin marketplace add Lykhoyda/ask-llm
/plugin install ask-llm@ask-llm-plugins
What You Get
| Feature | Description |
|---|---|
/multi-review | Parallel Gemini + Codex review with 4-phase validation pipeline and consensus highlighting |
/gemini-review | Gemini-only review with confidence filtering |
/codex-review | Codex-only review with confidence filtering |
/ollama-review | Local review — no data leaves your machine |
/brainstorm | Multi-LLM brainstorm: Claude Opus researches the topic against real files in parallel with external providers (Gemini/Codex/Ollama), then synthesizes all findings with verified findings weighted higher |
| Pre-commit hook | Reviews staged changes before git commit, warns about critical issues |
The review agents use a 4-phase pipeline inspired by Anthropic's code-review plugin: context gathering, prompt construction with explicit false-positive exclusions, synthesis, and source-level validation of each finding.
See the plugin docs for details.
Prerequisites
- Node.js v20.0.0 or higher (LTS)
- At least one provider:
- Gemini CLI —
npm install -g @google/gemini-cli && gemini login - Codex CLI — installed and authenticated
- Ollama — running locally with a model pulled (
ollama pull qwen2.5-coder:7b)
- Gemini CLI —
MCP Tools
| Tool | Package | Purpose |
|---|---|---|
ask-gemini | ask-gemini-mcp | Send prompts to Gemini CLI with @ file syntax. 1M+ token context |
ask-gemini-edit | ask-gemini-mcp | Get structured OLD/NEW code edit blocks from Gemini |
fetch-chunk | ask-gemini-mcp | Retrieve chunks from cached large responses |
ask-codex | ask-codex-mcp | Send prompts to Codex CLI. GPT-5.4 with mini fallback |
ask-ollama | ask-ollama-mcp | Send prompts to local Ollama. Fully private, zero cost |
ping | all | Connection test — verify MCP setup |
Usage Examples
ask gemini to review the changes in @src/auth.ts for security issues
ask codex to suggest a better algorithm for @src/sort.ts
ask ollama to explain @src/config.ts (runs locally, no data sent anywhere)
use gemini to summarize @. the current directory
Models
| Provider | Default | Fallback |
|---|---|---|
| Gemini | gemini-3.1-pro-preview | gemini-3-flash-preview (on quota) |
| Codex | gpt-5.4 | gpt-5.4-mini (on quota) |
| Ollama | qwen2.5-coder:7b | qwen2.5-coder:1.5b (if not found) |
All providers automatically fall back to a lighter model on errors.
Documentation
- Docs site: lykhoyda.github.io/ask-llm
- AI-readable: llms.txt | llms-full.txt
Contributing
Contributions are welcome! See open issues for things to work on.
License
MIT License. See LICENSE for details.
Disclaimer: This is an unofficial, third-party tool and is not affiliated with, endorsed, or sponsored by Google or OpenAI.
Related Servers
Scout Monitoring MCP
sponsorPut performance and error data directly in the hands of your AI assistant.
Alpha Vantage MCP Server
sponsorAccess financial market data: realtime & historical stock, ETF, options, forex, crypto, commodities, fundamentals, technical indicators, & more
YepCode
Execute any LLM-generated code in the YepCode secure and scalable sandbox environment and create your own MCP tools using JavaScript or Python, with full support for NPM and PyPI packages
LambdaTest MCP Server
LambdaTest MCP Servers ranging from Accessibility, SmartUI, Automation, and HyperExecute allows you to connect AI assistants with your testing workflow, streamlining setup, analyzing failures, and generating fixes to speed up testing and improve efficiency.
Firebase MCP Server
You can use the Firebase MCP server to give AI-powered development tools the ability to work with your Firebase projects and your app's codebase.
PydanticRPC
A Python library for building gRPC/ConnectRPC services with Pydantic models, featuring automatic protobuf generation and AI assistant tool exposure.
PyPI MCP Server
Search and access Python package metadata, version history, and download statistics from the PyPI repository.
Apple Developer Documentation
Search Apple's official developer documentation for iOS, macOS, SwiftUI, WWDC videos, and more.
Drupal Tools
A server providing resources, tools, and prompts for Drupal development.
promptz.dev
Access and manage prompts from promptz.dev for developers.
Remote MCP Server (Authless)
An example of a remote MCP server deployable on Cloudflare Workers, without authentication.
CLI Exec
Execute shell commands with structured output via a powerful CLI server.