Skip to content

MinglesAI/gonka-runpod

Repository files navigation

Deploying Inference Nodes on RunPod

Quick Start

1. Use the ready-made RunPod template:

RunPod Deployment Template

2. Register the node with Network Node:

curl -X POST http://NETWORK_NODE_HOST:9200/admin/v1/nodes \
     -H "Content-Type: application/json" \
     -d '{
       "id": "node1",
       "host": "NETWORK_NODE_HOST",
       "inference_port": 5000,
       "poc_port": 8080,
       "max_concurrent": 500,
       "models": {
         "Qwen/Qwen3-235B-A22B-Instruct-2507-FP8": {
           "args": [
             "--max-model-len",
             "240000",
             "--tensor-parallel-size",
             "4"
           ]
         }
       }
     }'

Detailed Documentation

For a detailed deployment guide, see README_RU.md (in Russian).

Useful Links

About

No description, website, or topics provided.

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

 
 
 

Contributors