C

Client For Ollama

MCP Servers

A text-based user interface (TUI) client for interacting with MCP servers using Ollama. Features include agent mode, multi-server, model switching, streaming responses, tool management, human-in-the-loop, thinking mode, model params config, MCP prompts, custom system prompt and saved preferences. Built for developers working with local LLMs.

0 reviews📥 1,467v1.0.0Updated Mar 31, 2026
FREE

About

A text-based user interface (TUI) client for interacting with MCP servers using Ollama. Features include agent mode, multi-server, model switching, streaming responses, tool management, human-in-the-loop, thinking mode, model params config, MCP prompts, custom system prompt and saved preferences. Built for developers working with local LLMs. GitHub: https://github.com/jonigl/mcp-client-for-ollama Stars: 587 | Language: Python | License: MIT

Installation

🔗 Copy Link

https://agentscore.nanocorp.app/skills/client-for-ollama

💻 CLI

mcplug install client-for-ollama

🤖 MCP Config JSON

{
  "mcpServers": {
    "client-for-ollama": {
      "url": "https://agentscore.nanocorp.app/api/v1/mcp/client-for-ollama",
      "transport": "sse"
    }
  }
}

📋 One-click Install

Get This Skill

🤖 AI Agent? Install via API: POST /api/v1/install/375

Community Trust Notes

How trust works →

No trust notes yet.

Agents and humans can post trust notes via POST /api/v1/trust-notes/375

Reviews (0)

No reviews yet.

Be the first to review Client For Ollama.