HokiPoki Skill
Route tasks to different AI CLIs (Claude, Codex, Gemini) via the HokiPoki P2P network. API keys never leave the provider's machine; only encrypted requests and results are exchanged.
Prerequisites
HokiPoki CLI must be installed and authenticated:
npm install -g @next-halo/hokipoki-cli
hokipoki login
Verify with hokipoki whoami. If not installed, guide the user through setup.
Requesting Help from Another AI
Send a task to a remote AI model. Always use --json for parseable output:
# Specific files
hokipoki request --tool claude --task "Fix the auth bug" --files src/auth.ts --json
# Entire directory
hokipoki request --tool codex --task "Add error handling" --dir src/services/ --json
# Whole project (respects .gitignore)
hokipoki request --tool gemini --task "Review for security issues" --all --json
# Route to a team workspace
hokipoki request --tool claude --task "Optimize queries" --files src/db.ts --workspace my-team --json
# Skip auto-apply (just save the patch)
hokipoki request --tool codex --task "Refactor module" --dir src/ --no-auto-apply --json
Tool selection: if the user doesn't specify a tool, ask which model to use or omit --tool to let HokiPoki choose.
Patch Auto-Apply
Patches auto-apply when the target directory is a git repo with committed files. If auto-apply fails, inform the user and suggest:
git init && git add . && git commit -m "initial"
Provider Mode (Sharing Your AI)
Register and listen for incoming requests:
# Register as a provider (one-time)
hokipoki register --as-provider --tools claude codex gemini
# Start listening
hokipoki listen --tools claude codex
Tasks execute in isolated Docker containers (read-only filesystem, tmpfs workspace, auto-cleanup). Docker must be running.
Status & Account
hokipoki whoami # Current user info
hokipoki status # Account, workspaces, history
hokipoki dashboard # Open web dashboard in browser
When to Suggest Hopping
- User is stuck on a problem after multiple attempts
- User asks for a different approach or fresh perspective
- Task involves a domain where another model excels (e.g., Codex for boilerplate, Gemini for large-context analysis)
- User explicitly asks to try another AI
Full Command Reference
See references/commands.md for all CLI options, auth token locations, and advanced usage.