Hosting

Become a Hoster

Share your GPU with the world — or just your team. Create a free plan to share privately, or set up paid subscriptions and keep 80% of revenue.

🚧 LLMFinder is currently in beta. Access is invite-only. Request an invite →

Requirements

One-command setup

The fastest way to get started. The setup wizard detects your GPU, picks a model, configures Docker Compose, and registers with LLMFinder:

curl -O https://llmfinder.net/llmfinder-hoster.py && python3 llmfinder-hoster.py

The wizard:

Top-level menu (7 options): 1) Setup wizard, 2) Add/update models, 3) Server verification test, 4) Update server URL, 5) Rotate bearer token, 6) Uninstall, 7) Exit.

💡 The Cloudflare tunnel URL changes on every restart (free tier). The script auto-syncs the URL with LLMFinder on each menu open. For a permanent URL, use a named Cloudflare tunnel.

Manual registration

If you already have a server running, register it directly:

curl -X POST https://api.llmfinder.net/hosters/register \
  -H "Content-Type: application/json" \
  -d '{
    "name": "My GPU Server",
    "email": "[email protected]",
    "endpoint_url": "https://my-server.example.com",
    "api_key": "my-bearer-token",
    "invite_code": "BETA2026",
    "models": [
      {
        "model_id": "llama-3-8b",
        "model_alias": "Llama 3 8B",
        "price_per_input_token": 100,
        "price_per_output_token": 300,
        "context_window": 8192,
        "max_tokens": 2048
      }
    ]
  }'

Supported server software

SoftwareCompatibleNotes
llama.cpp serverRecommended. Supports GGUF models.
vLLMBest for large HuggingFace models.
OllamaNatively OpenAI-compatible. No bridge needed.
Any OpenAI-compatible serverMust expose /health and /v1/chat/completions

Blocked endpoints

The following cannot be registered (ToS violation):

Verification

After registration, LLMFinder runs two checks:

  1. Health checkGET /health must return HTTP 200
  2. Inference test — sends a test prompt, expects a valid response

Once both pass, your server goes live and starts receiving traffic.