For Builders
Build on JimboMesh
One API. Hundreds of GPUs. The same OpenAI-compatible format you already use. Change one line. That's it.
🚀 Quick Start
Choose your setup path and copy-paste to launch a local Holler fast.
Detected OS: Auto
# Download from GitHub Releases
# Visit jimbomesh.ai/Holler for one-click download
docker run -d --name holler \
-p 1920:1920 \
--gpus all \
ghcr.io/ingresstechnology/jimbomesh-holler-server:latest
# Full stack with Ollama + Qdrant
curl -fsSL https://raw.githubusercontent.com/IngressTechnology/jimbomesh-holler-server/main/docker-compose.yml -o docker-compose.yml
docker compose up -d
Community
Join r/JimboMesh
Join the conversation on Reddit — share your setup, get help, show off your Holler.
Having trouble? Ask on r/JimboMesh — the community is here to help.
Join r/JimboMeshEnvironment Variables
| Variable | Purpose | Default |
|---|---|---|
ASPNETCORE_ENVIRONMENT |
Sets runtime environment (Development/Production). | Production |
HOLLER_PORT |
Public API/Admin port mapping for Holler. | 1920 |
OLLAMA_BASE_URL |
Internal Ollama endpoint used for model inference. | http://localhost:11434 |
QDRANT_URL |
Vector database endpoint for RAG collections. | http://localhost:6333 |
JIMBOMESH_API_KEY |
Optional key for connecting your local node to the mesh. | (unset) |
Dead Simple Integration
If your app already talks to OpenAI, it already talks to JimboMesh. We're fully OpenAI API compatible — same request format, same response format.
Change one base URL. Point it at api.jimbomesh.ai.
You're done. Seriously.
- OpenAI Python SDK, Node.js SDK — works out of the box
- Any HTTP client that speaks JSON
- Streaming support (SSE)
- Same auth pattern — API key in the header
- WebRTC peer-to-peer — your data never touches our servers
import openai
client = openai.OpenAI(
base_url="https://api.jimbomesh.ai/v1",
api_key="your-moonshine-key"
)
response = client.chat.completions.create(
model="llama3.3:70b",
messages=[{
"role": "user",
"content": "What is JimboMesh?"
}]
)
print(response.choices[0].message.content)
Change one line. That's it. If your app talks to OpenAI, it already talks to JimboMesh.
API Examples
OpenAI-compatible. Local Holler or the Mesh — same format.
Local Holler (Port 1920)
curl http://localhost:1920/v1/chat/completions \
-H "Content-Type: application/json" \
-d '{"model": "llama3.2:3b", "messages": [{"role": "user", "content": "Hello!"}]}'
$body = @{
model = "llama3.2:3b"
messages = @(@{role = "user"; content = "Hello!"})
} | ConvertTo-Json -Depth 3
Invoke-RestMethod -Uri "http://localhost:1920/v1/chat/completions" `
-Method POST -ContentType "application/json" -Body $body
Your hardware. Your models. No account required. View local Swagger docs →
Via JimboMesh Mesh
curl https://api.jimbomesh.ai/v1/chat/completions \
-H "Authorization: Bearer your_key_here" \
-H "Content-Type: application/json" \
-d '{"model": "llama3.3:70b", "messages": [{"role": "user", "content": "Hello!"}]}'
$headers = @{
"Authorization" = "Bearer your_key_here"
"Content-Type" = "application/json"
}
$body = @{
model = "llama3.3:70b"
messages = @(@{role = "user"; content = "Hello!"})
} | ConvertTo-Json -Depth 3
Invoke-RestMethod -Uri "https://api.jimbomesh.ai/v1/chat/completions" `
-Method POST -Headers $headers -Body $body
Distributed GPU mesh. One API key. Hundreds of models. Get your API key →
Try the Playground
Chat with AI models running on real GPUs — no setup required. Three-tab chat UI connected to real Hollers via WebRTC peer-to-peer.
- Browser-based — no install, no API key needed to try
- Connected to real Holler nodes on the mesh
- WebRTC direct — your data never touches our servers
WebRTC Peer-to-Peer
Your data never touches our servers. WebRTC connects you directly to the GPU. The platform handles signaling and billing. Your inference data goes peer-to-peer.
Available Models
The mesh serves what the Hollers run. More Hollers = more models.
| Model | Type | Params | Status |
|---|---|---|---|
| llama3.3:70b | Chat | 70B | Live |
| llama3.2:3b | Chat | 3B | Live |
| mistral-large | Chat | 123B | Live |
| qwen3 | Chat | Various | Live |
| nomic-embed-text | Embedding | 137M | Live |
37+ curated models available on the Holler marketplace. Model availability on the mesh depends on network supply. More Hollers = more models. Run a Holler to add capacity.