KS
Killer-Skills

run-thor — how to use run-thor how to use run-thor, run-thor setup guide, run-thor alternative, run-thor vs llama.cpp, run-thor install, what is run-thor, run-thor deployment, run-thor testing, run-thor GPU lab server

v1.0.0
GitHub

About this Skill

Perfect for AI Agents needing automated model management and dynamic resource allocation for edge deployment on OpenAI-compatible inference servers. run-thor is a cyber-inference tool built on llama.cpp, designed for managing and deploying OpenAI-compatible inference servers with a unique cyberpunk-themed interface.

Features

Provides automatic model management for OpenAI-compatible inference servers
Supports dynamic resource allocation for efficient edge deployment
Offers a cyberpunk-themed interface for user-friendly interaction
Accessible via SSH for secure remote access
Hosts production-like test environment on Thor, the GPU lab server
Utilizes llama.cpp for building and managing inference servers

# Core Topics

RamboRogers RamboRogers
[0]
[0]
Updated: 3/7/2026

Quality Score

Top 5%
39
Excellent
Based on code quality & docs
Installation
SYS Universal Install (Auto-Detect)
Cursor IDE Windsurf IDE VS Code IDE
> npx killer-skills add RamboRogers/cyber-inference/run-thor

Agent Capability Analysis

The run-thor MCP Server by RamboRogers is an open-source Categories.community integration for Claude and other AI agents, enabling seamless task automation and capability expansion. Optimized for how to use run-thor, run-thor setup guide, run-thor alternative.

Ideal Agent Persona

Perfect for AI Agents needing automated model management and dynamic resource allocation for edge deployment on OpenAI-compatible inference servers.

Core Value

Empowers agents to manage OpenAI models with automatic deployment and resource allocation via SSH and HTTP protocols, streamlining integration testing on GPU lab servers like Thor.

Capabilities Granted for run-thor MCP Server

Deploying and testing AI models on edge devices
Automating model updates and rollbacks on Thor
Debugging cyber-inference integration tests

! Prerequisites & Limits

  • Requires SSH access to thor.lab
  • Limited to OpenAI-compatible models
  • Dependent on GPU lab server availability
Project
SKILL.md
2.4 KB
.cursorrules
1.2 KB
package.json
240 B
Ready
UTF-8

# Tags

[No tags]
SKILL.md
Readonly

Deploy & Test on Thor

Thor is the GPU lab server used for integration testing of cyber-inference. It is accessible via SSH and hosts the production-like test environment.

Connection Details

FieldValue
Hostthor.lab
Usermatt
SSHssh matt@thor.lab
Project path/home/matt/Local/cyber-inference
Server URLhttp://thor.lab:8337

Deploy Workflow

Follow these steps in order. Each depends on the previous.

1. Commit & Push (local machine)

bash
1git add -A && git commit -m "<message>" && git push

2. Pull on Thor (remote)

bash
1ssh matt@thor.lab "cd /home/matt/Local/cyber-inference && git pull"

3. Start the Server (remote)

The server runs via start.sh which handles uv sync and auto-restart.

bash
1# Interactive (see logs live) - use for debugging 2ssh -t matt@thor.lab "cd /home/matt/Local/cyber-inference && ./start.sh" 3 4# Background (detached) - use for long-running tests 5ssh matt@thor.lab "cd /home/matt/Local/cyber-inference && nohup ./start.sh > /tmp/cyber-inference.log 2>&1 &"

CUDA PyTorch wheels are verified automatically when NVIDIA hardware is detected.

4. Verify the Server

bash
1# Health check 2curl -s http://thor.lab:8337/health 3 4# List models 5curl -s http://thor.lab:8337/v1/models | python3 -m json.tool 6 7# System status 8curl -s http://thor.lab:8337/admin/status | python3 -m json.tool 9

The web GUI is available at: http://thor.lab:8337

5. Test Inference

bash
1# Chat completion 2curl -s http://thor.lab:8337/v1/chat/completions \ 3 -H "Content-Type: application/json" \ 4 -d '{"model": "<model_name>", "messages": [{"role": "user", "content": "Hello"}]}' \ 5 | python3 -m json.tool 6 7# Embeddings 8curl -s http://thor.lab:8337/v1/embeddings \ 9 -H "Content-Type: application/json" \ 10 -d '{"model": "<model_name>", "input": "test text"}' \ 11 | python3 -m json.tool

Quick One-Liner Deploy

Pull latest and restart in one command:

bash
1ssh -t matt@thor.lab "cd /home/matt/Local/cyber-inference && git pull && ./start.sh"

Troubleshooting

  • Server won't start: Check logs with ssh matt@thor.lab "tail -50 /tmp/cyber-inference.log"
  • Port in use: Kill existing process with ssh matt@thor.lab "pkill -f 'cyber-inference serve'"
  • Check running processes: ssh matt@thor.lab "ps aux | grep cyber-inference"
  • GPU/CUDA issues: ssh matt@thor.lab "nvidia-smi"

Related Skills

Looking for an alternative to run-thor or building a Categories.community AI Agent? Explore these related open-source MCP Servers.

View All

widget-generator

Logo of f
f

widget-generator is an open-source AI agent skill for creating widget plugins that are injected into prompt feeds on prompts.chat. It supports two rendering modes: standard prompt widgets using default PromptCard styling and custom render widgets built as full React components.

149.6k
0
Design

chat-sdk

Logo of lobehub
lobehub

chat-sdk is a unified TypeScript SDK for building chat bots across multiple platforms, providing a single interface for deploying bot logic.

73.0k
0
Communication

zustand

Logo of lobehub
lobehub

The ultimate space for work and life — to find, build, and collaborate with agent teammates that grow with you. We are taking agent harness to the next level — enabling multi-agent collaboration, effortless agent team design, and introducing agents as the unit of work interaction.

72.8k
0
Communication

data-fetching

Logo of lobehub
lobehub

The ultimate space for work and life — to find, build, and collaborate with agent teammates that grow with you. We are taking agent harness to the next level — enabling multi-agent collaboration, effortless agent team design, and introducing agents as the unit of work interaction.

72.8k
0
Communication