Files
QwenClaw-with-Auth/SPARK_INTEGRATION_GUIDE.md

12 KiB
Raw Permalink Blame History

Spark Intelligence Integration Guide for QwenClaw

🚀 Why Spark Intelligence?

Spark Intelligence transforms QwenClaw from a stateless executor into a learning system that:

  • Remembers what worked and what didn't
  • Warns before repeating mistakes
  • Promotes validated wisdom automatically
  • Adapts to your specific workflows
  • Improves continuously through outcome tracking

📦 Installation

Step 1: Install Spark Intelligence

Windows (PowerShell):

irm https://raw.githubusercontent.com/vibeforge1111/vibeship-spark-intelligence/main/install.ps1 | iex

Mac/Linux:

curl -fsSL https://raw.githubusercontent.com/vibeforge1111/vibeship-spark-intelligence/main/install.sh | bash

Manual Install:

git clone https://github.com/vibeforge1111/vibeship-spark-intelligence
cd vibeship-spark-intelligence
python -m venv .venv
.venv\Scripts\activate  # Windows
# or: source .venv/bin/activate  # Mac/Linux
python -m pip install -e .[services]

Step 2: Verify Installation

python -m spark.cli health
python -m spark.cli up
python -m spark.cli learnings

🔗 Integration with QwenClaw

Architecture

┌─────────────────────────────────────────────────────────────┐
│                     QwenClaw Session                        │
└─────────────────────────────────────────────────────────────┘
                            │
                            ▼
┌─────────────────────────────────────────────────────────────┐
│              Spark Event Capture (Hooks)                    │
│  - PreToolUse   - PostToolUse   - UserPromptSubmit          │
└─────────────────────────────────────────────────────────────┘
                            │
                            ▼
┌─────────────────────────────────────────────────────────────┐
│              Spark Intelligence Pipeline                    │
│  Capture → Distill → Transform → Store → Act → Guard → Learn│
└─────────────────────────────────────────────────────────────┘
                            │
                            ▼
┌─────────────────────────────────────────────────────────────┐
│           Pre-Tool Advisory (Before QwenClaw Acts)          │
│  BLOCK (0.95+) | WARNING (0.80-0.95) | NOTE (0.48-0.80)     │
└─────────────────────────────────────────────────────────────┘

Configuration

Create ~/.spark/config.yaml:

spark:
  enabled: true
  session_id: qwenclaw-${timestamp}
  
  hooks:
    pre_tool_use: true
    post_tool_use: true
    user_prompt: true
    
  advisory:
    enabled: true
    min_score: 0.48
    cooldown_seconds: 300
    authority_levels:
      block: 0.95
      warning: 0.80
      note: 0.48
    
  memory:
    auto_capture: true
    min_importance: 0.55
    importance_boosts:
      causal_language: 0.15
      quantitative_data: 0.30
      technical_specificity: 0.15
    
  observatory:
    enabled: true
    sync_interval_seconds: 120
    vault_path: ~/Documents/Obsidian Vault/Spark-Intelligence-Observatory

Start Spark + QwenClaw

# Terminal 1: Start Spark pipeline
python -m spark.cli up

# Terminal 2: Start QwenClaw
qwenclaw start

# Terminal 3: Send tasks
qwenclaw send "Refactor the authentication module"

🧠 How Spark Improves QwenClaw

1. Pre-Tool Advisory Guidance

Before QwenClaw executes a tool, Spark surfaces relevant lessons:

BLOCK Example (0.95+ score)

⚠️ BLOCKED: Spark advisory

Action: rm -rf ./node_modules
Reason: This command will delete critical dependencies.
        Last 3 executions resulted in 2+ hour rebuild times.
        Alternative: npm clean-install

Confidence: 0.97 | Validated: 12 times

WARNING Example (0.80-0.95 score)

⚠️ WARNING: Spark advisory

Action: Edit file without reading
File: src/config/database.ts
Pattern: This pattern failed 4 times in the last 24 hours.
         Missing context caused incorrect modifications.

Suggestion: Read the file first, then edit.
Reliability: 0.91 | Validated: 8 times

NOTE Example (0.48-0.80 score)

 NOTE: Spark advisory

User Preference: Always use --no-cache flag for Docker builds
Context: Prevents stale layer caching issues
Captured from: Session #4521, 2 days ago

2. Anti-Pattern Detection

Spark identifies and corrects problematic workflows:

Pattern Detection Correction
Edit without Read File modified without prior read Suggests reading first
Recurring Command Failures Same command fails 3+ times Suggests alternatives
Missing Tests Code committed without tests Reminds testing policy
Hardcoded Secrets Secrets detected in code Blocks and warns

3. Memory Capture with Intelligence

Automatic Importance Scoring (0.0-1.0):

Score Action Example
≥0.65 Auto-save "Remember: always use --no-cache for Docker"
0.55-0.65 Suggest "I prefer TypeScript over JavaScript"
<0.55 Ignore Generic statements, noise

Signals that boost importance:

  • Causal language: "because", "leads to" (+0.15-0.30)
  • Quantitative data: "reduced from 4.2s to 1.6s" (+0.30)
  • Technical specificity: real tools, libraries, patterns (+0.15-0.30)

4. Auto-Promotion to Project Files

High-reliability insights automatically promote to:

CLAUDE.md - Wisdom, reasoning, context insights:

## Docker Best Practices
- Always use `--no-cache` flag for production builds
- Validated: 12 times | Reliability: 0.96

AGENTS.md - Meta-learning, self-awareness:

## Project Preferences
- Prefer TypeScript over JavaScript for large projects
- Test-first development required for core modules

SOUL.md - Communication preferences, user understanding:

## User Communication Style
- Prefers concise explanations with code examples
- Values performance metrics and quantitative data

5. EIDOS Episodic Intelligence

Extracts structured rules from experience:

Type Description Example
Heuristics General rules of thumb "Always test before deploying"
Sharp Edges Things to watch out for "API rate limits hit at 100 req/min"
Anti-Patterns What not to do "Don't edit config without backup"
Playbooks Proven approaches "Database migration checklist"
Policies Enforced constraints "Must have tests for core modules"

📊 Obsidian Observatory

Spark auto-generates 465+ markdown pages with live Dataview queries:

Generate Observatory

python scripts/generate_observatory.py --force --verbose

Vault Location: ~/Documents/Obsidian Vault/Spark-Intelligence-Observatory

What's Included

  • Pipeline Health - 12-stage pipeline detail pages with metrics
  • Cognitive Insights - Stored insights with reliability scores
  • EIDOS Episodes - Pattern distillations and heuristics
  • Advisory Decisions - Pre-tool guidance history
  • Explorer Views - Real-time data exploration
  • Canvas View - Spatial pipeline visualization

Auto-Sync

Observatory syncs every 120 seconds when pipeline is running.


📈 Measurable Outcomes

Advisory Source Effectiveness

Source What It Provides Effectiveness
Cognitive Validated session insights ~62% (dominant)
Bank User memory banks ~10%
EIDOS Pattern distillations ~5%
Baseline Static rules ~5%
Trigger Event-specific rules ~5%
Semantic BM25 + embedding retrieval ~3%

Timeline to Value

Time What Happens
Hour 1 Spark starts capturing events
Hour 2-4 Patterns emerge (tool effectiveness, error patterns)
Day 1-2 Insights get promoted to project files
Week 1+ Advisory goes live with pre-tool guidance

🔧 CLI Commands

Spark Commands

# Start pipeline
python -m spark.cli up

# Stop pipeline
python -m spark.cli down

# Check status
python -m spark.cli status

# View learnings
python -m spark.cli learnings

# View advisories
python -m spark.cli advisories

# Promote insight manually
python -m spark.cli promote <insight-id>

# Health check
python -m spark.cli health

QwenClaw Commands

# Start QwenClaw
qwenclaw start

# Send task (Spark captures automatically)
qwenclaw send "Refactor the authentication module"

# Check status
qwenclaw status

🎯 Best Practices

1. Let Spark Learn Naturally

Just use QwenClaw normally. Spark captures and learns in the background.

2. Provide Explicit Feedback

Tell Spark what to remember:

"Remember: always use --force for this legacy package"
"I prefer yarn over npm in this project"
"Test files should be in __tests__ directory"

3. Review Advisories

Pay attention to pre-tool warnings. They're based on validated patterns.

4. Check Observatory

Review the Obsidian vault weekly to understand what Spark has learned.

5. Promote High-Value Insights

Manually promote insights that are immediately valuable:

python -m spark.cli promote <insight-id>

🚨 Troubleshooting

Spark Not Capturing Events

Check:

python -m spark.cli health
python -m spark.cli status

Solution:

  • Ensure Spark pipeline is running: python -m spark.cli up
  • Verify hooks are enabled in config
  • Check QwenClaw session ID matches

Advisories Not Surfacing

Check:

python -m spark.cli advisories

Solution:

  • Verify advisory min_score in config (default: 0.48)
  • Check cooldown period (default: 300 seconds)
  • Ensure insights have been validated (5+ times)

Observatory Not Syncing

Check:

python scripts/generate_observatory.py --verbose

Solution:

  • Verify Obsidian vault path in config
  • Ensure vault exists
  • Check sync interval (default: 120 seconds)

📚 Resources


Summary

Spark Intelligence + QwenClaw = Self-Evolving AI Assistant

Without Spark With Spark
Stateless execution Continuous learning
Repeats mistakes Warns before errors
No memory Captures preferences
Static behavior Evolves over time
No observability Full Obsidian vault

Install Spark today and transform QwenClaw into a learning system! 🧠