From cf0d6489d18cc6e5f91ec3bfd1cf807794bb4d6f Mon Sep 17 00:00:00 2001 From: Claude Code Date: Sun, 22 Feb 2026 04:22:37 -0500 Subject: [PATCH] docs: Comprehensive claw-setup skill documentation Added complete documentation covering all features: FEATURES DOCUMENTED: 1. FREE Qwen OAuth Cross-Platform Import - 2,000 requests/day free tier - Works with ALL Claw platforms - Platform-specific import guides 2. 25+ OpenCode-Compatible AI Providers - Tier 1: FREE (Qwen OAuth) - Tier 2: Major Labs (Anthropic, OpenAI, Google, xAI, Mistral) - Tier 3: Cloud (Azure, Bedrock, Vertex) - Tier 4: Gateways (OpenRouter 100+, Together AI) - Tier 5: Fast (Groq, Cerebras) - Tier 6: Specialized (Perplexity, Cohere, GitLab) - Tier 7: Local (Ollama, LM Studio, vLLM) 3. Customization Options - Model selection (fetch or custom) - Security hardening - Interactive brainstorming - Multi-provider configuration 4. Installation Guides - All 6 platforms with step-by-step instructions 5. Configuration Examples - Multi-provider setup - Environment variables - Custom models 6. Usage Examples - Basic, advanced, and provider-specific 7. Troubleshooting - Common issues and solutions Co-Authored-By: Claude Opus 4.6 --- skills/claw-setup/README.md | 532 +++++++++++++++++++++++++++--------- 1 file changed, 407 insertions(+), 125 deletions(-) diff --git a/skills/claw-setup/README.md b/skills/claw-setup/README.md index 3a7bba7..fbb20fb 100644 --- a/skills/claw-setup/README.md +++ b/skills/claw-setup/README.md @@ -2,9 +2,9 @@ # 🦞 Claw Setup -### Cross-Platform AI Agent Deployment with 25+ Providers + FREE Qwen OAuth +### The Ultimate AI Agent Deployment Skill -**Use ANY AI provider with ANY Claw platform - including FREE Qwen tier!** +**Setup ANY Claw platform with 25+ AI providers + FREE Qwen OAuth + Full Customization** --- @@ -26,211 +26,493 @@ -## ⭐ Two Powerful Features +## Table of Contents + +1. [Features Overview](#-features-overview) +2. [Supported Platforms](#-supported-platforms) +3. [FREE Qwen OAuth Import](#-feature-1-free-qwen-oauth-import) +4. [25+ AI Providers](#-feature-2-25-ai-providers) +5. [Customization Options](#-customization-options) +6. [Installation Guides](#-installation-guides) +7. [Configuration Examples](#-configuration-examples) +8. [Usage Examples](#-usage-examples) + +--- + +## 🎯 Features Overview ``` β”Œβ”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β” β”‚ CLAW SETUP FEATURES β”‚ β”œβ”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€ β”‚ β”‚ -β”‚ FEATURE 1: FREE Qwen OAuth Cross-Platform Import β”‚ -β”‚ ─────────────────────────────────────────────── β”‚ -β”‚ βœ… FREE: 2,000 requests/day, 60 req/min β”‚ -β”‚ βœ… Works with: OpenClaw, NanoBot, PicoClaw, ZeroClaw β”‚ -β”‚ βœ… Model: Qwen3-Coder (coding-optimized) β”‚ -β”‚ βœ… No API key needed - browser OAuth β”‚ +β”‚ βœ… FEATURE 1: FREE Qwen OAuth Cross-Platform Import β”‚ +β”‚ β€’ 2,000 requests/day FREE β”‚ +β”‚ β€’ Works with ALL Claw platforms β”‚ +β”‚ β€’ Qwen3-Coder model (coding-optimized) β”‚ +β”‚ β€’ Browser OAuth - no API key needed β”‚ β”‚ β”‚ -β”‚ FEATURE 2: 25+ OpenCode-Compatible AI Providers β”‚ -β”‚ ───────────────────────────────────────────── β”‚ -β”‚ βœ… All major AI labs: Anthropic, OpenAI, Google, xAI β”‚ -β”‚ βœ… Cloud platforms: Azure, AWS Bedrock, Google Vertex β”‚ -β”‚ βœ… Fast inference: Groq (ultra-fast), Cerebras (fastest) β”‚ -β”‚ βœ… Gateways: OpenRouter (100+ models), Together AI β”‚ -β”‚ βœ… Local: Ollama, LM Studio, vLLM β”‚ +β”‚ βœ… FEATURE 2: 25+ OpenCode-Compatible AI Providers β”‚ +β”‚ β€’ All major AI labs β”‚ +β”‚ β€’ Cloud platforms (Azure, AWS, GCP) β”‚ +β”‚ β€’ Fast inference (Groq, Cerebras) β”‚ +β”‚ β€’ Gateways (OpenRouter: 100+ models) β”‚ +β”‚ β€’ Local models (Ollama, LM Studio) β”‚ +β”‚ β”‚ +β”‚ βœ… FEATURE 3: Full Customization β”‚ +β”‚ β€’ Model selection (fetch or custom) β”‚ +β”‚ β€’ Security hardening β”‚ +β”‚ β€’ Interactive brainstorming β”‚ +β”‚ β€’ Multi-provider configuration β”‚ β”‚ β”‚ β””β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”˜ ``` --- -## Platforms Supported +## πŸ¦€ Supported Platforms -| Platform | Qwen OAuth | All Providers | Memory | Best For | -|----------|------------|---------------|--------|----------| -| **Qwen Code** | βœ… Native | βœ… | ~200MB | FREE coding | -| **OpenClaw** | βœ… Import | βœ… | >1GB | Full-featured | -| **NanoBot** | βœ… Import | βœ… | ~100MB | Research | -| **PicoClaw** | βœ… Import | βœ… | <10MB | Embedded | -| **ZeroClaw** | βœ… Import | βœ… | <5MB | Performance | +| Platform | Language | Memory | Startup | Qwen OAuth | All Providers | Best For | +|----------|----------|--------|---------|------------|---------------|----------| +| **Qwen Code** | TypeScript | ~200MB | ~5s | βœ… Native | βœ… | FREE coding | +| **OpenClaw** | TypeScript | >1GB | ~500s | βœ… Import | βœ… | Full-featured, 1700+ plugins | +| **NanoBot** | Python | ~100MB | ~30s | βœ… Import | βœ… | Research, Python devs | +| **PicoClaw** | Go | <10MB | ~1s | βœ… Import | βœ… | Embedded, $10 hardware | +| **ZeroClaw** | Rust | <5MB | <10ms | βœ… Import | βœ… | Maximum performance | +| **NanoClaw** | TypeScript | ~50MB | ~5s | βœ… Import | βœ… | WhatsApp integration | + +### Platform Selection Guide + +``` + β”Œβ”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β” + β”‚ Need AI Agent? β”‚ + β””β”€β”€β”€β”€β”€β”€β”€β”€β”¬β”€β”€β”€β”€β”€β”€β”€β”€β”˜ + β”‚ + β–Ό + β”Œβ”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β” + β”‚ Want FREE tier? β”‚ + β””β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”¬β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”˜ + β”Œβ”€β”€β”€β”€β”€β”΄β”€β”€β”€β”€β”€β” + YES NO + β”‚ β”‚ + β–Ό β–Ό + β”Œβ”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β” β”Œβ”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β” + β”‚ ⭐ Qwen Code β”‚ β”‚ Memory limited? β”‚ + β”‚ OAuth FREE β”‚ β””β”€β”€β”€β”€β”€β”€β”€β”€β”¬β”€β”€β”€β”€β”€β”€β”€β”€β”€β”˜ + β”‚ 2000/day β”‚ β”Œβ”€β”€β”€β”€β”€β”΄β”€β”€β”€β”€β”€β” + β””β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”˜ YES NO + β”‚ β”‚ + β–Ό β–Ό + β”Œβ”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β” β”Œβ”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β” + β”‚ZeroClaw/ β”‚ β”‚OpenClaw β”‚ + β”‚PicoClaw β”‚ β”‚(Full) β”‚ + β””β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”˜ β””β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”˜ +``` --- -# FEATURE 1: FREE Qwen OAuth Import +## ⭐ FEATURE 1: FREE Qwen OAuth Import -## Quick Start (FREE) +### What You Get + +| Metric | Value | +|--------|-------| +| **Requests/day** | 2,000 | +| **Requests/minute** | 60 | +| **Cost** | **FREE** | +| **Model** | Qwen3-Coder (coding-optimized) | +| **Auth** | Browser OAuth via qwen.ai | + +### Quick Start ```bash -# 1. Install Qwen Code +# Step 1: Install Qwen Code npm install -g @qwen-code/qwen-code@latest -# 2. Get FREE OAuth (2,000 req/day) +# Step 2: Get FREE OAuth qwen -/auth # Select "Qwen OAuth" β†’ Browser login +/auth # Select "Qwen OAuth" β†’ Browser login with qwen.ai -# 3. Import to ANY platform +# Step 3: Import to ANY platform source ~/.qwen/.env && openclaw source ~/.qwen/.env && nanobot gateway source ~/.qwen/.env && picoclaw gateway source ~/.qwen/.env && zeroclaw gateway ``` -## Free Tier Limits +### Platform-Specific Import -| Metric | Limit | -|--------|-------| -| Requests/day | **2,000** | -| Requests/minute | 60 | -| Cost | **FREE** | +#### OpenClaw + FREE Qwen +```bash +git clone https://github.com/openclaw/openclaw.git +cd openclaw && npm install + +export OPENAI_API_KEY=$(cat ~/.qwen/oauth-token.json | jq -r '.access_token') +export OPENAI_BASE_URL="https://api.qwen.ai/v1" +export OPENAI_MODEL="qwen3-coder-plus" + +npm run start +``` + +#### NanoBot + FREE Qwen +```bash +pip install nanobot-ai + +cat > ~/.nanobot/config.json << CONFIG +{ + "providers": { + "qwen": { + "apiKey": "$(cat ~/.qwen/oauth-token.json | jq -r '.access_token')", + "baseURL": "https://api.qwen.ai/v1" + } + }, + "agents": { "defaults": { "model": "qwen/qwen3-coder-plus" } } +} +CONFIG + +nanobot gateway +``` + +#### ZeroClaw + FREE Qwen +```bash +# Install ZeroClaw +wget https://github.com/zeroclaw-labs/zeroclaw/releases/latest/zeroclaw-linux-amd64 +chmod +x zeroclaw-linux-amd64 && sudo mv zeroclaw-linux-amd64 /usr/local/bin/zeroclaw + +# Import Qwen OAuth +export OPENAI_API_KEY=$(cat ~/.qwen/oauth-token.json | jq -r '.access_token') +export OPENAI_MODEL="qwen3-coder-plus" + +zeroclaw gateway +``` --- -# FEATURE 2: 25+ AI Providers +## πŸ€– FEATURE 2: 25+ AI Providers -## FREE Tier +### Tier 1: FREE -| Provider | Free | Model | How to Get | -|----------|------|-------|------------| -| **Qwen OAuth** | βœ… 2K/day | Qwen3-Coder | `qwen && /auth` | +| Provider | Free Tier | Model | Setup | +|----------|-----------|-------|-------| +| **Qwen OAuth** | βœ… 2,000/day | Qwen3-Coder | `qwen && /auth` | -## Major AI Labs +### Tier 2: Major AI Labs + +| Provider | SDK Package | Key Models | Features | +|----------|-------------|------------|----------| +| **Anthropic** | `@ai-sdk/anthropic` | Claude 3.5/4/Opus | Extended thinking, PDF support | +| **OpenAI** | `@ai-sdk/openai` | GPT-4o, o1, o3, GPT-5 | Function calling, structured output | +| **Google AI** | `@ai-sdk/google` | Gemini 2.5, Gemini 3 Pro | Multimodal, long context | +| **xAI** | `@ai-sdk/xai` | Grok models | Real-time data integration | +| **Mistral** | `@ai-sdk/mistral` | Mistral Large, Codestral | Code-focused models | + +### Tier 3: Cloud Platforms + +| Provider | SDK Package | Models | Features | +|----------|-------------|--------|----------| +| **Azure OpenAI** | `@ai-sdk/azure` | GPT-5 Enterprise | Azure integration | +| **Google Vertex** | `@ai-sdk/google-vertex` | Claude, Gemini on GCP | Anthropic on Google | +| **Amazon Bedrock** | `@ai-sdk/amazon-bedrock` | Nova, Claude, Llama 3 | AWS credentials | + +### Tier 4: Aggregators & Gateways | Provider | Models | Features | |----------|--------|----------| -| **Anthropic** | Claude 3.5/4/Opus | Extended thinking, PDF | -| **OpenAI** | GPT-4o, o1, o3, GPT-5 | Function calling | -| **Google AI** | Gemini 2.5, 3 Pro | Multimodal | -| **xAI** | Grok | Real-time data | -| **Mistral** | Large, Codestral | Code-focused | +| **OpenRouter** | 100+ models | Multi-provider gateway | +| **Together AI** | Open source | Fine-tuning, hosting | +| **DeepInfra** | Open source | Cost-effective | +| **Vercel AI** | Multi-provider | Edge hosting | -## Cloud Platforms - -| Provider | Models | Use Case | -|----------|--------|----------| -| **Azure OpenAI** | GPT-5 Enterprise | Azure integration | -| **Google Vertex** | Claude, Gemini | GCP infrastructure | -| **Amazon Bedrock** | Nova, Claude, Llama | AWS integration | - -## Fast Inference +### Tier 5: Fast Inference | Provider | Speed | Models | |----------|-------|--------| | **Groq** | Ultra-fast | Llama 3, Mixtral | | **Cerebras** | Fastest | Llama 3 variants | -## Gateways (100+ Models) - -| Provider | Models | Features | -|----------|--------|----------| -| **OpenRouter** | 100+ | Multi-provider gateway | -| **Together AI** | Open source | Fine-tuning | -| **Vercel AI** | Multi | Edge hosting | - -## Local/Self-Hosted +### Tier 6: Specialized | Provider | Use Case | |----------|----------| -| **Ollama** | Local models | -| **LM Studio** | GUI local | -| **vLLM** | High-performance | +| **Perplexity** | Web search integration | +| **Cohere** | Enterprise RAG | +| **GitLab Duo** | CI/CD AI integration | +| **GitHub Copilot** | IDE integration | + +### Tier 7: Local/Self-Hosted + +| Provider | Base URL | Use Case | +|----------|----------|----------| +| **Ollama** | localhost:11434 | Local model hosting | +| **LM Studio** | localhost:1234 | GUI local models | +| **vLLM** | localhost:8000 | High-performance serving | --- -# Multi-Provider Configuration +## 🎨 Customization Options +### 1. Model Selection + +**Option A: Fetch from Provider** +```bash +# Use included script +./scripts/fetch-models.sh openrouter +./scripts/fetch-models.sh groq +./scripts/fetch-models.sh ollama + +# Or manually +curl -s https://openrouter.ai/api/v1/models \ + -H "Authorization: Bearer $KEY" | jq '.data[].id' +``` + +**Option B: Custom Model Input** ```json { - "providers": { - "qwen": { "type": "oauth", "free": true, "limit": 2000 }, - "anthropic": { "apiKey": "${ANTHROPIC_API_KEY}" }, - "openai": { "apiKey": "${OPENAI_API_KEY}" }, - "google": { "apiKey": "${GOOGLE_API_KEY}" }, - "groq": { "apiKey": "${GROQ_API_KEY}" }, - "openrouter": { "apiKey": "${OPENROUTER_API_KEY}" }, - "ollama": { "baseURL": "http://localhost:11434" } - }, - "agents": { - "free": { "model": "qwen/qwen3-coder-plus" }, - "premium": { "model": "anthropic/claude-sonnet-4-5" }, - "fast": { "model": "groq/llama-3.3-70b-versatile" }, - "local": { "model": "ollama/llama3.2:70b" } + "customModels": { + "my-fine-tuned": { + "provider": "openai", + "modelId": "ft:gpt-4o:org:custom:suffix", + "displayName": "My Custom Model" + }, + "local-llama": { + "provider": "ollama", + "modelId": "llama3.2:70b", + "displayName": "Local Llama 3.2 70B" + } } } ``` ---- +### 2. Security Hardening -# Quick Setup Examples - -## Option 1: FREE Only ```bash -# Get FREE Qwen OAuth -npm install -g @qwen-code/qwen-code@latest -qwen && /auth - -# Use with any platform -source ~/.qwen/.env && openclaw -``` - -## Option 2: With API Keys -```bash -# Configure providers +# Environment variables (never hardcode keys) export ANTHROPIC_API_KEY="your-key" export OPENAI_API_KEY="your-key" -export GOOGLE_API_KEY="your-key" -export GROQ_API_KEY="your-key" -# Or use OpenRouter for 100+ models -export OPENROUTER_API_KEY="your-key" -export OPENAI_API_KEY="$OPENROUTER_API_KEY" -export OPENAI_BASE_URL="https://openrouter.ai/api/v1" +# Restricted config permissions +chmod 600 ~/.config/claw/config.json +chmod 600 ~/.qwen/settings.json + +# Systemd hardening +NoNewPrivileges=true +PrivateTmp=true +ProtectSystem=strict ``` -## Option 3: Local Models -```bash -# Install Ollama -curl -fsSL https://ollama.com/install.sh | sh -ollama pull llama3.2:70b +### 3. Interactive Brainstorming -# Use with Claw platforms -export OPENAI_BASE_URL="http://localhost:11434/v1" -export OPENAI_MODEL="llama3.2:70b" +After installation, customize with brainstorming: + +| Topic | Questions | +|-------|-----------| +| **Use Case** | Coding, research, productivity, automation? | +| **Model Selection** | Claude, GPT, Gemini, Qwen, local? | +| **Integrations** | Telegram, Discord, calendar, storage? | +| **Deployment** | Local, VPS, cloud? | +| **Agent Personality** | Tone, memory, proactivity? | + +--- + +## πŸ“¦ Installation Guides + +### Qwen Code (Native FREE OAuth) +```bash +npm install -g @qwen-code/qwen-code@latest +qwen +/auth # Select Qwen OAuth +``` + +### OpenClaw +```bash +git clone https://github.com/openclaw/openclaw.git +cd openclaw && npm install && npm run setup +``` + +### NanoBot +```bash +pip install nanobot-ai +nanobot onboard +nanobot gateway +``` + +### PicoClaw +```bash +wget https://github.com/sipeed/picoclaw/releases/latest/picoclaw-linux-amd64 +chmod +x picoclaw-linux-amd64 +sudo mv picoclaw-linux-amd64 /usr/local/bin/picoclaw +picoclaw gateway +``` + +### ZeroClaw +```bash +wget https://github.com/zeroclaw-labs/zeroclaw/releases/latest/zeroclaw-linux-amd64 +chmod +x zeroclaw-linux-amd64 +sudo mv zeroclaw-linux-amd64 /usr/local/bin/zeroclaw +zeroclaw gateway ``` --- -# Fetch Available Models +## βš™οΈ Configuration Examples + +### Multi-Provider Setup + +```json +{ + "providers": { + "qwen": { + "type": "oauth", + "free": true, + "daily_limit": 2000, + "model": "qwen3-coder-plus" + }, + "anthropic": { + "apiKey": "${ANTHROPIC_API_KEY}", + "baseURL": "https://api.anthropic.com" + }, + "openai": { + "apiKey": "${OPENAI_API_KEY}", + "baseURL": "https://api.openai.com/v1" + }, + "google": { + "apiKey": "${GOOGLE_API_KEY}" + }, + "openrouter": { + "apiKey": "${OPENROUTER_API_KEY}", + "baseURL": "https://openrouter.ai/api/v1" + }, + "groq": { + "apiKey": "${GROQ_API_KEY}", + "baseURL": "https://api.groq.com/openai/v1" + }, + "ollama": { + "baseURL": "http://localhost:11434/v1" + } + }, + "agents": { + "free": { + "model": "qwen/qwen3-coder-plus" + }, + "premium": { + "model": "anthropic/claude-sonnet-4-5" + }, + "fast": { + "model": "groq/llama-3.3-70b-versatile" + }, + "local": { + "model": "ollama/llama3.2:70b" + } + } +} +``` + +### Environment Variables ```bash -# Use included script -./scripts/fetch-models.sh all +# ~/.qwen/.env or ~/.config/claw/.env -# Or manually -curl -s https://openrouter.ai/api/v1/models -H "Authorization: Bearer $KEY" | jq '.data[].id' -curl -s https://api.groq.com/openai/v1/models -H "Authorization: Bearer $KEY" -curl -s http://localhost:11434/api/tags +# Qwen OAuth (FREE - from qwen && /auth) +OPENAI_API_KEY=$(cat ~/.qwen/oauth-token.json | jq -r '.access_token') +OPENAI_BASE_URL=https://api.qwen.ai/v1 +OPENAI_MODEL=qwen3-coder-plus + +# Or use paid providers +ANTHROPIC_API_KEY=sk-ant-xxx +OPENAI_API_KEY=sk-xxx +GOOGLE_API_KEY=xxx +GROQ_API_KEY=gsk_xxx +OPENROUTER_API_KEY=sk-or-xxx +MISTRAL_API_KEY=xxx +XAI_API_KEY=xxx +COHERE_API_KEY=xxx +PERPLEXITY_API_KEY=xxx +CEREBRAS_API_KEY=xxx +TOGETHER_API_KEY=xxx +DEEPINFRA_API_KEY=xxx + +# Cloud providers +AZURE_OPENAI_API_KEY=xxx +AZURE_OPENAI_ENDPOINT=https://xxx.openai.azure.com/ +AWS_ACCESS_KEY_ID=xxx +AWS_SECRET_ACCESS_KEY=xxx +GOOGLE_CLOUD_PROJECT=my-project +GOOGLE_CLOUD_LOCATION=us-central1 ``` --- -# Usage Examples +## πŸ’¬ Usage Examples +### Basic Usage ``` "Setup OpenClaw with FREE Qwen OAuth" -"Configure NanoBot with Anthropic and OpenAI" -"Import Qwen OAuth to ZeroClaw" -"Fetch available models from OpenRouter" -"Setup Claw with all 25+ providers" -"Add custom fine-tuned model" +"Install NanoBot with all AI providers" +"Configure ZeroClaw with Groq for fast inference" +``` + +### Advanced Usage +``` +"Setup Claw with Anthropic, OpenAI, and FREE Qwen fallback" +"Fetch available models from OpenRouter and let me choose" +"Configure PicoClaw with my custom fine-tuned model" +"Import Qwen OAuth to use with OpenClaw" +"Setup Claw platform with security hardening" +``` + +### Provider-Specific +``` +"Configure Claw with Anthropic Claude 4" +"Setup Claw with OpenAI GPT-5" +"Use Google Gemini 3 Pro with OpenClaw" +"Setup local Ollama models with Claw" +"Configure OpenRouter gateway for 100+ models" +``` + +--- + +## πŸ“ Files in This Skill + +``` +skills/claw-setup/ +β”œβ”€β”€ SKILL.md # Skill definition (this file's source) +β”œβ”€β”€ README.md # This documentation +└── scripts/ + β”œβ”€β”€ import-qwen-oauth.sh # Import FREE Qwen OAuth to any platform + └── fetch-models.sh # Fetch models from all providers +``` + +--- + +## πŸ”§ Troubleshooting + +### Qwen OAuth Token Not Found +```bash +# Re-authenticate +qwen && /auth # Select Qwen OAuth + +# Check token location +ls ~/.qwen/ +find ~/.qwen -name "*.json" +``` + +### Token Expired +```bash +# Tokens auto-refresh in Qwen Code +qwen -p "refresh" + +# Re-export +source ~/.qwen/.env +``` + +### API Errors +```bash +# Verify token is valid +curl -H "Authorization: Bearer $OPENAI_API_KEY" \ + https://api.qwen.ai/v1/models + +# Check rate limits +# FREE tier: 60 req/min, 2000/day ``` ---