Skip to content

Go

Low cost subscription for open coding models.

OpenCode Go is a low cost $10/month subscription that gives you reliable access to popular open coding models.

Go works like any other provider in OpenCode. You subscribe to OpenCode Go and get your API key. It’s completely optional and you don’t need to use it to use OpenCode.

It is designed primarily for international users, with models hosted in the US, EU, and Singapore for stable global access.


Background

Open models have gotten really good. They now reach performance close to proprietary models for coding tasks. And because many providers can serve them competitively, they are usually far cheaper.

However, getting reliable, low latency access to them can be difficult. Providers vary in quality and availability.

To fix this, we did a couple of things:

  1. We tested a select group of open models and talked to their teams about how to best run them.
  2. We then worked with a few providers to make sure these were being served correctly.
  3. Finally, we benchmarked the combination of the model/provider and came up with a list that we feel good recommending.

OpenCode Go gives you access to these models for $10/month.


How it works

OpenCode Go works like any other provider in OpenCode.

  1. You sign in to OpenCode Zen, subscribe to Go, and copy your API key.
  2. You run the /connect command in the TUI, select OpenCode Go, and paste your API key.
  3. Run /models in the TUI to see the list of models available through Go.

The current list of models includes:

  • GLM-5
  • Kimi K2.5
  • MiniMax M2.5

The list of models may change as we test and add new ones.


Usage limits

OpenCode Go includes the following limits:

  • 5 hour limit — $12 of usage
  • Weekly limit — $30 of usage
  • Monthly limit — $60 of usage

Limits are defined in dollar value. This means your actual request count depends on the model you use. Cheaper models like MiniMax M2.5 allow for more requests, while higher-cost models like GLM-5 allow for fewer.

The table below provides an estimated request count based on typical Go usage patterns:

GLM-5Kimi K2.5MiniMax M2.5
requests per 5 hour1,1501,85020,000
requests per week2,8804,63050,000
requests per month5,7509,250100,000

Estimates are based on observed average request patterns:

  • GLM-5 — 700 input, 52,000 cached, 150 output tokens per request
  • Kimi K2.5 — 870 input, 55,000 cached, 200 output tokens per request
  • MiniMax M2.5 — 300 input, 55,000 cached, 125 output tokens per request

You can track your current usage in the console.

Usage limits may change as we learn from early usage and feedback.


Usage beyond limits

If you also have credits on your Zen balance, you can enable the Use balance option in the console. When enabled, Go will fall back to your Zen balance after you’ve reached your usage limits instead of blocking requests.


Endpoints

You can also access Go models through the following API endpoints.

ModelModel IDEndpointAI SDK Package
GLM-5glm-5https://opencode.ai/zen/go/v1/chat/completions@ai-sdk/openai-compatible
Kimi K2.5kimi-k2.5https://opencode.ai/zen/go/v1/chat/completions@ai-sdk/openai-compatible
MiniMax M2.5minimax-m2.5https://opencode.ai/zen/go/v1/messages@ai-sdk/anthropic

The model id in your OpenCode config uses the format opencode-go/<model-id>. For example, for Kimi K2.5, you would use opencode-go/kimi-k2.5 in your config.


Privacy

The plan is designed primarily for international users, with models hosted in the US, EU, and Singapore for stable global access.

Contact us if you have any questions.


Goals

We created OpenCode Go to:

  1. Make AI coding accessible to more people with a low cost subscription.
  2. Provide reliable access to the best open coding models.
  3. Curate models that are tested and benchmarked for coding agent use.
  4. Have no lock-in by allowing you to use any other provider with OpenCode as well.