Understanding the MCP Host
What is MCP?
MCP is a protocol developed by Anthropic for Claude. MCP, an acronym for Model Context Protocol, enables LLMs to actively request external actions or resources. Since MCP is merely a protocol for literal requests and responses, the process and execution must be handled by the developer.
Regarding Internal Operations
Before explaining internal operations, let us briefly examine Gemini Function Calling. Similar to MCP, Gemini Function Calling allows an LLM to proactively invoke external actions. One might then question why Function Calling is being introduced. The reason for its inclusion is that Function Calling predates MCP, and both utilize the OpenAPI schema, suggesting compatibility and similar operational behavior. Consequently, the explanations provided for Gemini Function Calling are more detailed and are thus presented as a helpful reference.
The overall flow is as follows:
- Define the function.
- Send the function definition to Gemini along with the prompt.
- "Send user prompt along with the function declaration(s) to the model. It analyzes the request and determines if a function call would be helpful. If so, it responds with a structured JSON object."
- Gemini requests a function call if necessary.
- If required by Gemini, the caller receives the name and parameters for the function call.
- The caller can decide whether or not to execute.
- Whether to execute and return a valid value.
- Whether to return data as if called, without actually executing.
- Whether to simply disregard.
- In the process above, Gemini performs and requests actions such as calling multiple functions at once or calling a function and then, based on the result, calling another.
- The process terminates when a refined answer is produced.
This flow generally aligns with MCP. This is similarly explained in the MCP tutorial. This also applies to Ollama tools.
Fortunately, these three tools—Ollama tools, MCP, and Gemini Function Calling—share a nearly identical schema structure, meaning that implementing MCP alone allows for its use across all three.
Furthermore, there is a common drawback shared by all of them. Ultimately, since the model executes the functions, if the model you are using is in a poor state, it may malfunction by not calling functions, calling them incorrectly, or even launching a Denial-of-Service attack on the MCP server.
MCP Host in Go
mark3lab's mcphost
In Go, there is mcphost, which is under development by the mark3lab organization.
Its usage is remarkably straightforward.
1go install github.com/mark3labs/mcphost@latest
After installation, create the $HOME/.mcp.json
file and populate it as follows:
1{
2 "mcpServers": {
3 "sqlite": {
4 "command": "uvx",
5 "args": [
6 "mcp-server-sqlite",
7 "--db-path",
8 "/tmp/foo.db"
9 ]
10 },
11 "filesystem": {
12 "command": "npx",
13 "args": [
14 "-y",
15 "@modelcontextprotocol/server-filesystem",
16 "/tmp"
17 ]
18 }
19 }
20}
Then, execute it with an Ollama model as shown below.
Of course, if necessary, first acquire the model via ollama pull mistral-small
.
While Claude or Qwen 2.5 are generally recommended, I currently suggest mistral-small.
1mcphost -m ollama:mistral-small
However, executing it in this manner restricts its use to a command-line interface for question-and-answer interactions.
Therefore, we will modify the code of mcphost
to enable more programmatic behavior.
mcphost Fork
As already confirmed, mcphost
includes functionalities for extracting metadata and calling functions using MCP. Therefore, components for calling the LLM, managing the MCP server, and handling message history are required.
The Runner
in the following package incorporates these components:
1package runner
2
3import (
4 "context"
5 "encoding/json"
6 "fmt"
7 "log"
8 "strings"
9 "time"
10
11 mcpclient "github.com/mark3labs/mcp-go/client"
12 "github.com/mark3labs/mcp-go/mcp"
13
14 "github.com/mark3labs/mcphost/pkg/history"
15 "github.com/mark3labs/mcphost/pkg/llm"
16)
17
18type Runner struct {
19 provider llm.Provider
20 mcpClients map[string]*mcpclient.StdioMCPClient
21 tools []llm.Tool
22
23 messages []history.HistoryMessage
24}
We shall not examine the internal declarations of this section. However, they are largely self-explanatory by their names.
1func NewRunner(systemPrompt string, provider llm.Provider, mcpClients map[string]*mcpclient.StdioMCPClient, tools []llm.Tool) *Runner {
2 return &Runner{
3 provider: provider,
4 mcpClients: mcpClients,
5 tools: tools,
6 messages: []history.HistoryMessage{
7 {
8 Role: "system",
9 Content: []history.ContentBlock{{
10 Type: "text",
11 Text: systemPrompt,
12 }},
13 },
14 },
15 }
16}
For mcpClients
and tools
used herein, please refer to this file.
For the provider
, which will utilize Ollama's, please refer to this file.
The main component is the Run
method.
1func (r *Runner) Run(ctx context.Context, prompt string) (string, error) {
2 if len(prompt) != 0 {
3 r.messages = append(r.messages, history.HistoryMessage{
4 Role: "user",
5 Content: []history.ContentBlock{{
6 Type: "text",
7 Text: prompt,
8 }},
9 })
10 }
11
12 llmMessages := make([]llm.Message, len(r.messages))
13 for i := range r.messages {
14 llmMessages[i] = &r.messages[i]
15 }
16
17 const initialBackoff = 1 * time.Second
18 const maxRetries int = 5
19 const maxBackoff = 30 * time.Second
20
21 var message llm.Message
22 var err error
23 backoff := initialBackoff
24 retries := 0
25 for {
26 message, err = r.provider.CreateMessage(
27 context.Background(),
28 prompt,
29 llmMessages,
30 r.tools,
31 )
32 if err != nil {
33 if strings.Contains(err.Error(), "overloaded_error") {
34 if retries >= maxRetries {
35 return "", fmt.Errorf(
36 "claude is currently overloaded. please wait a few minutes and try again",
37 )
38 }
39
40 time.Sleep(backoff)
41 backoff *= 2
42 if backoff > maxBackoff {
43 backoff = maxBackoff
44 }
45 retries++
46 continue
47 }
48
49 return "", err
50 }
51
52 break
53 }
54
55 var messageContent []history.ContentBlock
56
57 var toolResults []history.ContentBlock
58 messageContent = []history.ContentBlock{}
59
60 if message.GetContent() != "" {
61 messageContent = append(messageContent, history.ContentBlock{
62 Type: "text",
63 Text: message.GetContent(),
64 })
65 }
66
67 for _, toolCall := range message.GetToolCalls() {
68 input, _ := json.Marshal(toolCall.GetArguments())
69 messageContent = append(messageContent, history.ContentBlock{
70 Type: "tool_use",
71 ID: toolCall.GetID(),
72 Name: toolCall.GetName(),
73 Input: input,
74 })
75
76 parts := strings.Split(toolCall.GetName(), "__")
77
78 serverName, toolName := parts[0], parts[1]
79 mcpClient, ok := r.mcpClients[serverName]
80 if !ok {
81 continue
82 }
83
84 var toolArgs map[string]interface{}
85 if err := json.Unmarshal(input, &toolArgs); err != nil {
86 continue
87 }
88
89 var toolResultPtr *mcp.CallToolResult
90 req := mcp.CallToolRequest{}
91 req.Params.Name = toolName
92 req.Params.Arguments = toolArgs
93 toolResultPtr, err = mcpClient.CallTool(
94 context.Background(),
95 req,
96 )
97
98 if err != nil {
99 errMsg := fmt.Sprintf(
100 "Error calling tool %s: %v",
101 toolName,
102 err,
103 )
104 log.Printf("Error calling tool %s: %v", toolName, err)
105
106 toolResults = append(toolResults, history.ContentBlock{
107 Type: "tool_result",
108 ToolUseID: toolCall.GetID(),
109 Content: []history.ContentBlock{{
110 Type: "text",
111 Text: errMsg,
112 }},
113 })
114
115 continue
116 }
117
118 toolResult := *toolResultPtr
119
120 if toolResult.Content != nil {
121 resultBlock := history.ContentBlock{
122 Type: "tool_result",
123 ToolUseID: toolCall.GetID(),
124 Content: toolResult.Content,
125 }
126
127 var resultText string
128 for _, item := range toolResult.Content {
129 if contentMap, ok := item.(map[string]interface{}); ok {
130 if text, ok := contentMap["text"]; ok {
131 resultText += fmt.Sprintf("%v ", text)
132 }
133 }
134 }
135
136 resultBlock.Text = strings.TrimSpace(resultText)
137
138 toolResults = append(toolResults, resultBlock)
139 }
140 }
141
142 r.messages = append(r.messages, history.HistoryMessage{
143 Role: message.GetRole(),
144 Content: messageContent,
145 })
146
147 if len(toolResults) > 0 {
148 r.messages = append(r.messages, history.HistoryMessage{
149 Role: "user",
150 Content: toolResults,
151 })
152
153 return r.Run(ctx, "")
154 }
155
156 return message.GetContent(), nil
157}
The code itself is a compilation of sections from this file.
The content is approximately as follows:
- Transmit the prompt along with the list of tools to request execution or response generation.
- If a response is generated, stop recursion and return.
- If the LLM requests tool execution, the host invokes the MCP Server.
- Add the response to the history and return to step 1.
Conclusion
Already finished?
Indeed, there is not much more to elaborate upon. This article was composed to aid in comprehending the general operational principles of the MCP Server. It is hoped that this text has contributed, even in a small measure, to your understanding of MCP host operations.