For Builders

Build on JimboMesh

One API. Hundreds of GPUs. The same OpenAI-compatible format you already use. Change one line. That's it.

🚀 Quick Start

Choose your setup path and copy-paste to launch a local Holler fast.

Detected OS: Auto

# Download from GitHub Releases
# Visit jimbomesh.ai/Holler for one-click download
🪟 Windows 🍎 macOS 🐧 Linux
docker run -d --name holler \
  -p 1920:1920 \
  --gpus all \
  ghcr.io/ingresstechnology/jimbomesh-holler-server:latest
# Full stack with Ollama + Qdrant
curl -fsSL https://raw.githubusercontent.com/IngressTechnology/jimbomesh-holler-server/main/docker-compose.yml -o docker-compose.yml
docker compose up -d

Community

Join r/JimboMesh

Join the conversation on Reddit — share your setup, get help, show off your Holler.

Having trouble? Ask on r/JimboMesh — the community is here to help.

Join r/JimboMesh

Environment Variables

Variable Purpose Default
ASPNETCORE_ENVIRONMENT Sets runtime environment (Development/Production). Production
HOLLER_PORT Public API/Admin port mapping for Holler. 1920
OLLAMA_BASE_URL Internal Ollama endpoint used for model inference. http://localhost:11434
QDRANT_URL Vector database endpoint for RAG collections. http://localhost:6333
JIMBOMESH_API_KEY Optional key for connecting your local node to the mesh. (unset)

Dead Simple Integration

If your app already talks to OpenAI, it already talks to JimboMesh. We're fully OpenAI API compatible — same request format, same response format.

Change one base URL. Point it at api.jimbomesh.ai. You're done. Seriously.

  • OpenAI Python SDK, Node.js SDK — works out of the box
  • Any HTTP client that speaks JSON
  • Streaming support (SSE)
  • Same auth pattern — API key in the header
  • WebRTC peer-to-peer — your data never touches our servers
python
import openai

client = openai.OpenAI(
    base_url="https://api.jimbomesh.ai/v1",
    api_key="your-moonshine-key"
)

response = client.chat.completions.create(
    model="llama3.3:70b",
    messages=[{
        "role": "user",
        "content": "What is JimboMesh?"
    }]
)

print(response.choices[0].message.content)

Change one line. That's it. If your app talks to OpenAI, it already talks to JimboMesh.

API Examples

OpenAI-compatible. Local Holler or the Mesh — same format.

Local Holler (Port 1920)

curl http://localhost:1920/v1/chat/completions \
  -H "Content-Type: application/json" \
  -d '{"model": "llama3.2:3b", "messages": [{"role": "user", "content": "Hello!"}]}'
$body = @{
    model = "llama3.2:3b"
    messages = @(@{role = "user"; content = "Hello!"})
} | ConvertTo-Json -Depth 3

Invoke-RestMethod -Uri "http://localhost:1920/v1/chat/completions" `
  -Method POST -ContentType "application/json" -Body $body

Your hardware. Your models. No account required. View local Swagger docs →

Via JimboMesh Mesh

curl https://api.jimbomesh.ai/v1/chat/completions \
  -H "Authorization: Bearer your_key_here" \
  -H "Content-Type: application/json" \
  -d '{"model": "llama3.3:70b", "messages": [{"role": "user", "content": "Hello!"}]}'
$headers = @{
    "Authorization" = "Bearer your_key_here"
    "Content-Type" = "application/json"
}
$body = @{
    model = "llama3.3:70b"
    messages = @(@{role = "user"; content = "Hello!"})
} | ConvertTo-Json -Depth 3

Invoke-RestMethod -Uri "https://api.jimbomesh.ai/v1/chat/completions" `
  -Method POST -Headers $headers -Body $body

Distributed GPU mesh. One API key. Hundreds of models. Get your API key →

Try the Playground

Chat with AI models running on real GPUs — no setup required. Three-tab chat UI connected to real Hollers via WebRTC peer-to-peer.

  • Browser-based — no install, no API key needed to try
  • Connected to real Holler nodes on the mesh
  • WebRTC direct — your data never touches our servers
Open the Playground →

WebRTC Peer-to-Peer

Your data never touches our servers. WebRTC connects you directly to the GPU. The platform handles signaling and billing. Your inference data goes peer-to-peer.

8 IDE Integrations

Replace GitHub Copilot, Cursor Pro, or Cody with your own hardware — or the mesh.

Point any IDE at http://localhost:1920 (local) or https://api.jimbomesh.ai (mesh). Save $79-108/month on AI coding subscriptions.

Available Models

The mesh serves what the Hollers run. More Hollers = more models.

Model Type Params Status
llama3.3:70b Chat 70B Live
llama3.2:3b Chat 3B Live
mistral-large Chat 123B Live
qwen3 Chat Various Live
nomic-embed-text Embedding 137M Live

37+ curated models available on the Holler marketplace. Model availability on the mesh depends on network supply. More Hollers = more models. Run a Holler to add capacity.

Ready to Build on Cheap Compute?

Get your API key. Point your IDE at the mesh. Or fire up a Holler and run it all locally.