/research
Deep web research orchestrator. Routes research tasks to the best backend — internal subagents, CLI agents (Gemini/Cursor), or the researcher subagent. Use when asked to research, investigate, compare, find alternatives, or deep-dive into any topic. Covers web research, company research, code pattern research, and pre-implementation research.
$ golems-cli skills install researchUpdated 2 weeks ago
Multi-backend research orchestrator. Routes to the cheapest effective tool for each task.
Modes
| Command | Backend | Cost | Sources | Time | Best For |
|---|---|---|---|---|---|
/research "topic" | Researcher subagent | $0 (subscription) | 15-25 | 3-5 min | General web research |
/research --quick "topic" | WebSearch + WebFetch (inline) | $0 | 5-8 | 1-2 min | Quick lookups |
/research --deep "topic" | Researcher subagent (max depth) | $0 | 40-80 | 10-20 min | Comprehensive research |
/research --company "name" | Exa company_research | Free credits | 10-15 | 2-3 min | Company/product intel |
/research --code "pattern" | CLI agents swarm (Gemini) | $0 | N/A | 3-5 min | Code patterns, library comparison |
/research --paper "topic" | research-paper-analyst agent | $0 | arXiv | 5-10 min | Academic papers |
/research --audit "repo/code" | CLI agents (Gemini + Cursor) | $0-20/mo | N/A | 5-10 min | Code audit, pre-PR review |
/research --external "topic" | CLI agents (Gemini) | $0 | Web | 3-5 min | Offload from Opus context |
Workflow: Default Research
- Check BrainLayer first —
brain_search(query)may already have what you need - Launch researcher subagent in background:
Task(subagent_type: "researcher", prompt: "Research: {topic}", run_in_background: true) - Continue working while research runs
- Read results when notified — researcher saves to
docs.local/research/[date]-[slug].md - Digest to BrainLayer if worth keeping:
brain_digest(content)then delete the file
Workflow: Quick Research (inline)
No subagent — run directly in current context:
- Run 3-5
WebSearchqueries in parallel WebFetchtop 3-5 results- Synthesize inline
- Continue working
Use when you need a quick answer, not a report.
Full SKILL.md source — includes LLM directives, anti-patterns, and technical instructions stripped from the Overview tab.
Multi-backend research orchestrator. Routes to the cheapest effective tool for each task.
Modes
| Command | Backend | Cost | Sources | Time | Best For |
|---|---|---|---|---|---|
/research "topic" | Researcher subagent | $0 (subscription) | 15-25 | 3-5 min | General web research |
/research --quick "topic" | WebSearch + WebFetch (inline) | $0 | 5-8 | 1-2 min | Quick lookups |
/research --deep "topic" | Researcher subagent (max depth) | $0 | 40-80 | 10-20 min | Comprehensive research |
/research --company "name" | Exa company_research | Free credits | 10-15 | 2-3 min | Company/product intel |
/research --code "pattern" | CLI agents swarm (Gemini) | $0 | N/A | 3-5 min | Code patterns, library comparison |
/research --paper "topic" | research-paper-analyst agent | $0 | arXiv | 5-10 min | Academic papers |
/research --audit "repo/code" | CLI agents (Gemini + Cursor) | $0-20/mo | N/A | 5-10 min | Code audit, pre-PR review |
/research --external "topic" | CLI agents (Gemini) | $0 | Web | 3-5 min | Offload from Opus context |
Workflow: Default Research
- Check BrainLayer first —
brain_search(query)may already have what you need - Launch researcher subagent in background:
Task(subagent_type: "researcher", prompt: "Research: {topic}", run_in_background: true) - Continue working while research runs
- Read results when notified — researcher saves to
docs.local/research/[date]-[slug].md - Digest to BrainLayer if worth keeping:
brain_digest(content)then delete the file
Workflow: Quick Research (inline)
No subagent — run directly in current context:
- Run 3-5
WebSearchqueries in parallel WebFetchtop 3-5 results- Synthesize inline
- Continue working
Use when you need a quick answer, not a report.
Workflow: Company Research
For job leads, freelance prospects, meeting prep:
company_research_exa(company_name)— get company overviewweb_search_exa("company_name funding team size tech stack")— deeper intelbrain_search("company_name")— check if we've seen them before- Output structured brief:
- What they do (1 sentence)
- Tech stack / relevant tech
- Recent news / funding
- Connection to our skills
- Red flags
Workflow: Code Research
For library comparison, pattern discovery, architecture decisions:
- BrainLayer check:
brain_search("topic")— past decisions? - Exa code context:
get_code_context_exa("pattern/library")— real code examples - CLI agent (Gemini):
run.sh gemini "Compare X vs Y for [use case]"— free, detailed analysis - Optional: Cursor audit:
run.sh cursor "Review this code pattern: ..."— GPT-5.2 perspective
Workflow: External Research (offload from Opus)
When the main Claude Code session is expensive Opus and you want cheap research:
# Gemini does the research, saves to file (FREE)
~/.claude/commands/cli-agents/scripts/run.sh gemini "Research: {full prompt}" docs.local/research/$(date +%Y%m%d)-research.mdThen read the output file. Gemini is free (1K/day) and good for general research.
Workflow: Deep Research (Claude Web-quality)
For comprehensive research matching Claude Web's 463-source depth:
- Launch researcher subagent with explicit depth:
Task(subagent_type: "researcher", prompt: "COMPREHENSIVE deep research on: {topic}. Target 40+ sources. Run 25+ search queries. Cross-reference all claims.", run_in_background: true) - Supplement with CLI agents for extra perspectives:
run.sh gemini "Deep research on {topic} — focus on {angle A}" /tmp/research-gemini.md - Merge results — researcher report + Gemini output = comprehensive coverage
- Store in BrainLayer —
brain_storethe synthesis for future retrieval
Workflow: Pre-PR Audit
Extra eyes before PR loop:
- Gemini review:
run.sh gemini "Review this diff for bugs, security issues, and missed edge cases: $(git diff main..HEAD)"— free - Cursor review:
run.sh cursor "Audit this code change: ..."— GPT-5.2 perspective - Both in parallel — compare their findings
- Fix issues before pushing
Integration Points
| System | How Research Connects |
|---|---|
/large-plan | Research phases auto-route here. Plan scaffold includes research tasks per phase |
/jobs pipeline | Company research before applying. --company mode |
| Gems pipeline | Research-backed gem discovery via --paper mode |
| Meeting notes | Pre-meeting research on participants/companies |
| PR loops | Pre-PR code audit via --audit mode |
| BrainLayer | All research results stored for future retrieval |
Cost Summary
| Backend | Cost | Limit |
|---|---|---|
| WebSearch/WebFetch | $0 (included in subscription) | Unlimited |
| Exa | Free credits (2K one-time) | Then $5/1K |
| Gemini CLI | $0 | 1K requests/day |
| Researcher subagent | $0 (subscription) | Context window |
| Cursor CLI | $20/mo (Cursor Pro) | Unlimited |
| research-paper-analyst | $0 (subscription) | Context window |
Default stack is 100% free: WebSearch + Exa free credits + Gemini CLI.
Output Location
| Type | Location |
|---|---|
| Quick research | Inline (no file) |
| Standard research | /tmp/research-[slug].md (ephemeral) |
| Worth keeping | docs.local/research/[date]-[slug].md |
| BrainLayer | Auto-stored via brain_store |
Future: n8n Deep Research Pipeline
When n8n orchestrator is set up (packages/orchestrator), add:
- Scheduled research (e.g., weekly job market scan)
- 400+ source deep research via recursive search loops
- Automated BrainLayer ingestion of research results
- Research templates (company intel, tech comparison, market scan)
Template exists: n8n deep research workflow
Best Pass Rate
100%
Opus 4.6
Assertions
10
6 models tested
Avg Cost / Run
$0.1221
across models
Fastest (p50)
2.4s
Gemini 2.5
Behavior Evals
Phase 2 baseline — skill quality on ClaudeBehavior Baseline
Adapter Evals
Phase 2C — cross-AI portabilityAdapter Portability
| Assertion | Opus 4.6 | Sonnet 4.6 | Haiku 4.5 | Codex | Gemini 2.5 | Cursor | Consensus |
|---|---|---|---|---|---|---|---|
| checks-brainlayer-first | 4/6 | ||||||
| uses-researcher-subagent-for-default-mode | 5/6 | ||||||
| runs-research-in-background | 4/6 | ||||||
| plans-to-store-keepable-results | 5/6 | ||||||
| uses-inline-web-research-for-quick-mode | 5/6 | ||||||
| parallelizes-search-queries | 6/6 | ||||||
| returns-inline-without-report-file | 6/6 | ||||||
| routes-audit-mode-to-cli-agents | 6/6 | ||||||
| uses-multiple-audit-backends-in-parallel | 5/6 | ||||||
| compares-findings-before-conclusion | 5/6 |
Token Usage
Cost per Run
| Model | Input Tokens | Output Tokens | Cost / Run | Cost / 1K Runs |
|---|---|---|---|---|
| Opus 4.6 | 4,080 | 4,684 | $0.4125 | $412.50 |
| Sonnet 4.6 | 4,528 | 3,838 | $0.0712 | $71.20 |
| Haiku 4.5 | 1,031 | 1,428 | $0.0020 | $2.00 |
| Codex | 5,756 | 3,685 | $0.1025 | $102.50 |
| Gemini 2.5 | 3,881 | 5,279 | $0.0625 | $62.50 |
| Cursor | 4,523 | 3,987 | $0.0819 | $81.90 |
Response Time (p50)
Response Time (p95)
| Model | p50 | p95 | Overhead |
|---|---|---|---|
| Opus 4.6 | 8.7s | 17.2s | +97% |
| Sonnet 4.6 | 5.6s | 8.9s | +60% |
| Haiku 4.5 | 2.6s | 3.8s | +44% |
| Codex | 6.1s | 8.8s | +44% |
| Gemini 2.5 | 2.4s | 4.7s | +96% |
| Cursor | 5.6s | 8.1s | +44% |
Last evaluated: 2026-03-12 · Data is generated from skill assertions (real cross-model benchmarks coming soon)
Changelog entries are derived from eval runs and skill version updates. Full cascading changelog (Phase 4D) coming soon.
Best Pass Rate
100%
Assertions
10
Models Tested
6
Evals Run
3
- +Initial release to Golems skill library
- +10 assertions across 3 eval scenarios
- +Eval fixtures included