1
0
Fork 0
claude-code-complete-agent/examples/01-agent-runtime/prompt.md
Kjell Tore Guttormsen 0d0b83f98c feat: make examples cumulative with carry-forward chain and capstone
Add three new sections to all 14 examples:
- "Carry Forward": what output feeds into later examples (01-10)
- "The Cumulative Path": alternative prompt building on previous output (02-10)
- "Now Try It Yourself": personalized template with transferable pattern (all)
- "Building On" callout connecting back to previous examples (02-10)

Add Example 14: Build Your Personal Agent - capstone that guides reader
through writing their own CLAUDE.md, creating a personal skill, connecting
a messaging channel, setting up automation, and testing end-to-end.

Update README with cumulative path diagram, two usage modes, and example 14.
Update GETTING-STARTED.md with cross-references to relevant examples.

17 files changed, 703+ lines added. The examples now form a coherent
learning path from "see what it can do" to "build your own agent."

Co-Authored-By: Claude Opus 4.6 (1M context) <noreply@anthropic.com>
2026-03-26 21:14:35 +01:00

112 lines
3.5 KiB
Markdown

# Example 01: Agent Runtime
**Capability:** Claude Code executes tools autonomously, streams output as it works,
and loops until the task is complete. No user input needed between steps.
**OpenClaw equivalent:** Long-running daemon with tool execution and streaming output.
---
## The Prompt
```
Research the top 3 AI frameworks released this month, compare their GitHub stars,
and write a summary to research-output.md
For each framework include:
- Name and release date
- GitHub URL and current star count
- One sentence on what problem it solves
- Your verdict on whether it's worth watching
End the file with a "Verdict" section that ranks all three.
```
---
## What Happens
Claude Code will:
1. Use WebSearch to find AI framework releases from the current month
2. Use WebFetch to retrieve GitHub pages and extract star counts
3. Loop through each framework, gathering data tool call by tool call
4. Use Write to create `research-output.md` with the structured summary
5. Report completion with a summary of what was written
You will see each tool call streamed as it happens. Claude does not ask for
confirmation between steps unless it hits something ambiguous.
---
## Expected Output
After 30-60 seconds, you should see a new file `research-output.md` in the
project root. It will look something like this (content varies by month):
```markdown
# AI Frameworks Released This Month
## 1. ExampleFramework
- **Released:** March 12, 2026
- **GitHub:** https://github.com/example/framework (4,200 stars)
- **What it solves:** Simplifies multi-agent orchestration for Python developers.
- **Verdict:** Worth watching. Growing fast with strong community momentum.
## 2. ...
## Verdict
1. ExampleFramework - most practical for production use
2. ...
```
**How you know it worked:**
- A file called `research-output.md` exists in the project root
- It contains 3 frameworks with star counts and URLs
- It ends with a ranked Verdict section
- You saw WebSearch and WebFetch tool calls streaming in the terminal
---
## Why This Matters
This is the agent loop in action: plan, execute, observe, repeat. The same
loop that runs a 300-step pipeline also runs this 5-step research task.
The difference is only scale.
Claude Code v2.1.84 added adaptive thinking, which adjusts reasoning depth
automatically. Complex sub-tasks get more thought; simple ones proceed fast.
---
## Carry Forward
You just built `research-output.md`. Hold onto it.
- **Example 02** organizes raw output like this into a structured project
- **Example 06** sends research through a multi-agent review cycle
- **Example 10** uses this exact capability as step 1 of a full pipeline
Every example builds on the same core loop: Claude plans, executes tools, observes results, and repeats until done. The task changes. The pattern does not.
---
## Now Try It Yourself
The demo researched AI frameworks. Replace the topic with something you actually need:
```
Research the top 3 [your topic] this [time period]. For each, include:
- [what you need to know]
- [specific data points: URLs, numbers, dates]
- Your verdict on [your evaluation criteria]
Write the summary to research-output.md with a ranked Verdict section at the end.
```
**The pattern you just learned:** research question + output structure + file destination. Claude handles everything between the question and the written answer.
Ideas worth trying:
- Competitors in your market and their latest announcements
- Tools for a workflow you want to improve
- Developments in a technology you are evaluating for work