Now Open Source v0.1

Infrastructure
Intelligence

Kubernetes was built for web apps. Clawbernetes is AI-native infrastructure management you talk to โ€” deployments, diagnostics, scaling, and secrets, all through conversation. Built in Rust.

Morpheus AI Assistant
๐Ÿ‘ค
You
Deploy Llama 70B inference. Need 4 H100s with low latency for production traffic.
๐Ÿ”ฎ
Morpheus
Analyzing your cluster topology... I found node-01 with 8ร— H100s connected via NVLink โ€” perfect for your latency requirements.
โœ“ Deployed in 2.3s
โœ“ Allocated 4ร— H100 (NVLink topology)
โœ“ Monitoring active โ€” p99: 42ms
๐Ÿ‘ค
You
What's the VRAM usage looking like?
๐Ÿ”ฎ
Morpheus
VRAM utilization is healthy at 68GB/80GB per GPU. I've configured automatic scaling โ€” if traffic spikes above 1000 req/s, I'll spin up 4 more H100s on node-02.
๐Ÿ‘ค
You
Perfect. Alert me if p99 latency goes above 100ms.
๐Ÿ”ฎ
Morpheus
Alert configured. I'll notify you via Slack if p99 exceeds 100ms for more than 60 seconds. Your workload is running smoothly! ๐Ÿš€

What is Clawbernetes?

Clawbernetes is a conversational replacement for Kubernetes โ€” same capabilities, none of the YAML โ€” where you describe what you need in plain language and an AI agent handles scheduling, secrets, networking, scaling, and self-healing across CPUs, GPUs, and heterogeneous clusters.

๐Ÿš€
Deployments
You "Deploy a vLLM server with Llama 3 70B on the node with the most VRAM"
Agent Selects the best node, pulls the image, starts the container with GPU passthrough, and sets up health monitoring โ€” all in one turn.
๐Ÿ”
Diagnostics
You "Why is inference slow on morpheus?"
Agent Checks GPU temps, VRAM, CPU load, and container stats. "GPU 0 at 89ยฐC โ€” thermal throttling. Want me to reduce batch size?"
๐Ÿ“Š
Fleet Overview
You "What GPUs do we have across the cluster?"
Agent Queries every connected node and returns a full inventory โ€” GPU models, VRAM, utilization, temps, and running workloads per node.
๐Ÿ”
Secrets Management
You "Store the HuggingFace token as a secret and rotate it monthly"
Agent Encrypts with AES-256-GCM, stores on the node, and sets up a cron rotation schedule. No Vault needed.
โš–๏ธ
Autoscaling
You "Scale the inference server between 2 and 8 replicas based on queue depth"
Agent Creates an autoscale policy, monitors the metric, and adjusts replicas automatically. Reports scaling events to you in chat.
๐Ÿ’ฐ
MOLT Marketplace
You "I need 4 A100s for 6 hours. Find the cheapest spot on MOLT."
Agent Scans the P2P marketplace, verifies hardware attestation, escrows MOLT tokens, and provisions the GPUs.
66K
Lines of Rust
23
Crates
2.1K
Tests
0
Unsafe in Core
5
GPU Backends

Everything K8s promised. Actually delivered.

๐Ÿ—ฃ๏ธ

Intent-Based Operations

Describe workloads in natural language. The AI agent handles GPU selection, networking, secrets, monitoring, and scaling automatically.

๐Ÿ”

AI-Native Observability

Ask "why is training slow?" and get a diagnosis. Replaces Prometheus, Grafana, Alertmanager, Loki, and Jaeger.

โšก

GPU Topology Awareness

NVLink, PCIe, VRAM-aware scheduling. The agent understands interconnect bandwidth and places workloads optimally.

๐Ÿ”

Zero-Trust Security

AES-GCM encryption, attestation-based access, built-in PKI, automatic certificate and secret rotation.

๐ŸŒ

Flexible Networking

WireGuard for full control, Tailscale for zero-config, or MOLT P2P for decentralized compute marketplace.

๐Ÿ”„

Autonomous Self-Healing

Automatic rollback with root cause analysis. The agent learns from failures and prevents recurrence.

Every GPU. One API.

NVIDIA
CUDA
H100 / A100 / RTX
โœ“ Ready
Apple
Metal
M1โ€“M4 Ultra
โœ“ Tested
AMD
ROCm
MI300X / Radeon
โœ“ Ready
Cross-Platform
Vulkan
Intel / AMD / NV
โœ“ Ready
Fallback
CPU SIMD
AVX / NEON
โœ“ Ready

Kubernetes vs Clawbernetes

Concern Kubernetes Clawbernetes
Configuration 500+ lines of YAML Natural language intent
GPU Scheduling Device plugin hacks NVLink/PCIe/VRAM native
Monitoring Prometheus + Grafana + Loki "What's wrong?" โ†’ Diagnosis
Deployments Helm + ArgoCD + Kustomize Agent-managed intents
Secrets External Secrets + Vault Built-in encrypted rotation
Language Go Rust โ€” 0 unsafe in core

Built for Every Workload

๐Ÿค–

LLM Inference

Deploy and scale large language models with automatic batching and load balancing.

GPU Optimized
๐Ÿง 

Model Training

Distributed training across nodes with checkpoint management and failure recovery.

GPU Optimized
๐Ÿ”„

ETL Pipelines

Data transformation workflows with intelligent scheduling and resource allocation.

CPU + GPU
๐Ÿ“Š

Batch Processing

High-throughput batch jobs with priority queuing and automatic retry logic.

CPU Workload
๐ŸŽจ

Image Generation

Stable Diffusion, DALL-E, and custom diffusion models at scale.

GPU Optimized
๐ŸŽฌ

Video Processing

Transcoding, analysis, and AI-enhanced video pipelines with GPU acceleration.

GPU Optimized
๐Ÿ”ฌ

Scientific Compute

Simulations, molecular dynamics, and research workloads with multi-node support.

HPC Ready
๐ŸŒ

Web Services

Stateless APIs and microservices with auto-scaling and health monitoring.

CPU Workload

Fleet-Scale GPU Orchestration

โ”Œโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ” โ”‚ Control Plane โ”‚ โ”‚ โ”Œโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ” โ”‚ โ”‚ โ”‚ OpenClaw Gateway โ”‚ โ”‚ โ”‚ โ”‚ โ”Œโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ” โ”Œโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ” โ”Œโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ” โ”Œโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ” โ”‚ โ”‚ โ”‚ โ”‚ โ”‚ Fleet โ”‚ โ”‚ Intent โ”‚ โ”‚ Node โ”‚ โ”‚ Workload โ”‚ โ”‚ โ”‚ โ”‚ โ”‚ โ”‚ Agent โ”‚ โ”‚ Parser โ”‚ โ”‚ Registry โ”‚ โ”‚ State โ”‚ โ”‚ โ”‚ โ”‚ โ”‚ โ””โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”˜ โ””โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”˜ โ””โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”˜ โ””โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”˜ โ”‚ โ”‚ โ”‚ โ””โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”˜ โ”‚ โ””โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”ฌโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”˜ โ”‚ WebSocket + Protobuf (TLS) โ”‚ โ”Œโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”ผโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ” โ–ผ โ–ผ โ–ผ โ”Œโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ” โ”Œโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ” โ”Œโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ” โ”‚ clawnode โ”‚ โ”‚ clawnode โ”‚ โ”‚ clawnode โ”‚ โ”‚ 8x H100 โ”‚ โ”‚ 4x A100 โ”‚ โ”‚ M4 Ultra โ”‚ โ”‚ NVLink โ”‚ โ”‚ PCIe โ”‚ โ”‚ Metal โ”‚ โ””โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”˜ โ””โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”˜ โ””โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”˜

Get Started in 5 Minutes

1

Install OpenClaw

The control plane that powers Clawbernetes.

npm install -g openclaw@latest && openclaw onboard --install-daemon
2

Build clawnode

The agent binary that runs on each machine in your fleet.

git clone https://github.com/clawbernetes/clawbernetes && cd clawbernetes && cargo install --path crates/clawnode
3

Connect a node

Generate config, run the node agent, approve from the gateway.

clawnode init-config --gateway ws://gateway:18789 && clawnode run --config ./clawnode-config.json
4

Install skills & plugin

Teach the agent infrastructure ops โ€” deploy, scale, diagnose, heal.

cp -r skills/* ~/.openclaw/workspace/skills/ && openclaw plugins install --link ./plugin/openclaw-clawbernetes
5

Talk to your infrastructure

No YAML. No dashboards. Just tell it what you need.

"What GPUs do we have?" ยท "Deploy nginx on the node with the most free RAM." ยท "Why is inference slow?"

Ready to replace Kubernetes?

Open source. Written in Rust. MIT licensed.