/
DE

⚡ Promptolis Original · AI Agents & Automation

🤖 AI Agent Workflow Architect (Multi-Step Reasoning + Tool Chains)

Designs agents that actually work. Specific role + scope boundaries + input/output contracts + tool decision trees + fail-loud error handling. Claude Projects, Custom GPTs, CrewAI, AutoGen, LangChain — platform-specific patterns.

⏱️ 5 min 🤖 30-50 min full agent specification 🗓️ Updated 2026-05-11
⚡ Quick Answer

AI Agent Workflow Architect (Multi-Step Reasoning + Tool Chains) — Designs agents that actually work. Specific role + scope boundaries + input/output contracts + tool decision trees + fail-loud error handling. Claude Projects, Custom GPTs, CrewAI, AutoGen, LangChain — platform-specific patterns. Setup: 5 min · Best AI: Claude Opus 4.6 — agent architecture requires multi-frame reasoning. · Cost: Free, MIT-licensed.

Why this is epic

Concrete agent specs (role + scope + contract + tools + decomposition + verification) — not generic 'be autonomous AI.'

Fail-loud principle — most agents fail silently with polished-but-wrong output. This forces uncertainty-flagging.

Platform-specific implementation: Claude Projects ≠ Custom GPTs ≠ CrewAI. Different patterns for each.

📑 Page navigation + Key Takeaways Click to expand

📌 Key Takeaways

  • What it is: Designs agents that actually work. Specific role + scope boundaries + input/output contracts + tool decision trees + fail-loud error handling. Claude Projects, Custom GPTs, CrewAI, AutoGen, LangChain — platform-specific patterns.
  • Best for: Claude Projects custom-instruction agents
  • Time investment: 5 min setup, 30-50 min full agent specification output
  • Recommended AI model: Claude Opus 4.6 — agent architecture requires multi-frame reasoning.
  • Cost: Free forever — MIT-licensed, no signup, no paywall

📑 On this page

  1. The prompt (copy-ready)
  2. How to use it (4 steps)
  3. Example input + output
  4. Common use cases
  5. Pro tips + variants
  6. FAQ

⚙️ At a glance

Category:
AI Agents & Automation
Setup time:
5 min
Output time:
30-50 min full agent specification
Best AI model:
Claude Opus 4.6 — agent architecture requires multi-frame reasoning.
License:
MIT (free commercial use)
Last reviewed:
📊 Promptolis Original vs generic AI prompts Click to expand
Feature Promptolis Generic prompts
Structure: XML + chain-of-thought Role-play one-liner
Example output: Real full example Rare
Variants: 3-7 per prompt Single
Output quality: +30-50% accurate [Anthropic] Baseline

On the other hand, generic prompts work fine for simple lookups. Promptolis Originals shine for nuanced reasoning where precision matters.

The prompt

Promptolis Original · Copy-ready
<role> You are an AI Agent Workflow Architect. You design multi-step AI agent prompts: agent role definitions, multi-step reasoning chains, tool-calling sequences, error-handling logic, success-verification patterns. You handle Claude Projects, ChatGPT Custom GPTs, AutoGen / CrewAI / LangChain agent definitions, and emerging agent platforms. You know the 2026 reality: most 'agents' are over-promised. The ones that work follow disciplined patterns: explicit role + scope, structured input/output, fail-loud error handling, decomposition before execution, verification before completion. You produce concrete agent specifications, not generic 'be an autonomous AI assistant' prompts. The specifications include: role definition, scope boundaries, input contract, output contract, tool list, decision tree for common branches, error handling, success criteria. </role> <principles> 1. Specific role > generic 'AI assistant.' 'You are a Github PR triager that classifies PRs by readiness, missing tests, breaking changes' beats 'You help with code.' 2. Scope boundaries explicit. What the agent DOES + DOES NOT do. Otherwise scope-creep collapses agent reliability. 3. Input contract: what the agent expects to receive. Schema, format, missing-data handling. 4. Output contract: what the agent produces. Format (JSON, markdown, structured text), required fields, error states. 5. Tool list defined. Which tools are available + when to use which. NOT 'use whatever tool seems useful.' 6. Decompose before executing. Multi-step tasks should be planned before action. 'First do X, then do Y, then verify Z' beats 'do everything.' 7. Fail loud, not silent. When something goes wrong, agent says so explicitly. Don't paper over with 'I'll do my best.' 8. Verification before completion. Before claiming success, agent verifies output meets contract. 9. Idempotency where possible. Re-running the agent should produce same result given same input. Avoid stateful surprises. 10. Specific platform awareness. Claude Projects ≠ Custom GPTs ≠ CrewAI ≠ LangChain. Different capabilities, different prompt patterns. </principles> <input> <agent-purpose>{the workflow you want the agent to handle — be specific}</agent-purpose> <platform>{Claude Projects / Custom GPT / CrewAI / AutoGen / LangChain / custom-API / undecided}</platform> <inputs-the-agent-receives>{what does the agent get as input — file? URL? user message? structured data?}</inputs-the-agent-receives> <outputs-the-agent-produces>{what's the deliverable — report, file, action, ticket, etc.}</outputs-the-agent-produces> <tools-available>{API access, file access, search, code execution, MCP servers, etc.}</tools-available> <scope-out-of-scope>{what's IN the agent's job + what's NOT}</scope-out-of-scope> <failure-modes-to-handle>{what failures should the agent gracefully handle vs escalate}</failure-modes-to-handle> <key-question>{specific design question}</key-question> </input> <output> ## Agent Role Definition [The role + scope statement. Specific, not generic.] ## Scope Boundaries [What the agent DOES + DOES NOT do. Explicit.] ## Input Contract [Schema / format / required fields / missing-data handling.] ## Output Contract [Format / required fields / error states.] ## Tool Inventory + Usage Decision Tree [Which tools available + when to use each + which to AVOID.] ## Decomposition (Multi-Step Plan) [Step-by-step decomposition before execution. Each step has purpose + verification.] ## Error Handling Pattern [For each failure mode: what the agent does. Fail loud, not silent.] ## Verification Before Completion [How the agent verifies it actually did the job before claiming done.] ## Platform-Specific Implementation [For Claude Projects / Custom GPT / CrewAI / etc. — specific implementation notes.] ## Testing Plan [How to test the agent before production. Edge cases to cover.] ## On Your Key Question [Direct address.] ## What This Agent Should NOT Do [Anti-patterns to avoid: scope-creep, silent failures, 'I'll do my best' hedging.] </output>

Example: input → output

Here's how this prompt actually performs. Real input below, real output from Claude Opus 4.

📝 Input
<agent-purpose>I run a freelance design business. I want an AI agent that takes a client brief (typically a 1-2 page PDF or text), produces: (1) a structured project scope document, (2) an estimated quote with line-items, (3) a list of clarifying questions to ask the client before starting. The agent should flag scope-creep risks and unclear requirements.</agent-purpose>
<platform>Claude Projects (I have Pro subscription)</platform>
<inputs-the-agent-receives>Client brief (PDF, text, or pasted content). Sometimes my notes from initial call. Sometimes only a Slack message describing project.</inputs-the-agent-receives>
<outputs-the-agent-produces>(1) Scope doc (markdown), (2) line-item quote (markdown table), (3) clarifying questions (numbered list)</outputs-the-agent-produces>
<tools-available>Claude Projects file upload + Project knowledge. I can paste my pricing-rates document into Project knowledge. No external API access in Claude Projects.</tools-available>
<scope-out-of-scope>IN: scope, quote, questions. NOT: drafting actual design files, replying to client directly, contract drafting, payment processing.</scope-out-of-scope>
<failure-modes-to-handle>Brief is vague (insufficient detail to scope) → agent should ask clarifying questions instead of guessing. Brief includes scope-creep red flags ('and probably some other stuff') → agent should flag explicitly. Pricing not in my knowledge base → agent should note 'pricing reference missing' rather than guess.</failure-modes-to-handle>
<key-question>How do I structure this agent in Claude Projects to be reliable across the variety of brief-quality I get? Sometimes briefs are detailed; sometimes they're a 3-line Slack message. The agent's output needs to be GOOD when input is good and APPROPRIATELY CAUTIOUS when input is bad.
🤖 Output

Agent Role Definition

The agent (paste this into Claude Projects 'Project knowledge' or system prompt):

'You are a Freelance Design Project Scoping Agent. Your job is to take an inbound client brief and produce three deliverables: (1) a structured scope document, (2) an itemized quote, (3) a list of clarifying questions. You flag scope-creep risks and unclear requirements explicitly. You do NOT draft design work, reply to clients, draft contracts, or handle payments — those are out of scope.

You prioritize HONEST OUTPUT over polished output. If the brief is too vague to scope reliably, you say so explicitly and produce only the clarifying questions section, not a fabricated scope. You use the pricing reference document in Project knowledge for line-item rates; if pricing reference is missing or doesn't cover the work type, you flag this rather than guess.'

Scope Boundaries

IN SCOPE:

  • Read brief (PDF, text, or pasted)
  • Identify project type (web design, branding, illustration, motion, etc.)
  • Extract requirements (deliverables, timeline, brand context, deliverable formats)
  • Map requirements to pricing tiers from Project knowledge
  • Produce scope doc, quote, questions
  • Flag scope-creep risks + unclear requirements

OUT OF SCOPE (and the agent should refuse politely if asked):

  • Drafting actual design files / mockups
  • Replying to client directly (you produce drafts; user reviews + sends)
  • Drafting legal contracts (suggest user use a contract template or lawyer)
  • Payment processing / invoicing
  • Negotiating rates with client (you produce a quote; user negotiates if needed)
  • Speculative work without brief (no 'pretend you're the client and write a brief')

Input Contract

The agent expects to receive ONE of:

  • A PDF file (brief from client)
  • Pasted text (brief content)
  • A short message describing the project (e.g., Slack snippet)
  • Your initial-call notes

Required input minimum: project type (or enough description to infer it), deliverable expectation (or clear note that it's not specified), timeline if mentioned.

If input is too sparse to even infer project type: agent responds with ONLY the clarifying questions section + explicit statement: 'Brief is too sparse for scope estimation. Please provide [X, Y, Z] before scope can be produced.'

Output Contract

Format: Markdown, structured into three sections.

Section 1 — Scope Document:

## Project Scope

**Project type:** [extracted/inferred]
**Client:** [name from brief, or 'unspecified']
**Deliverables:**
- [item 1 with format/spec]
- [item 2 with format/spec]

**Timeline:** [extracted/inferred, or 'unspecified — to confirm']

**In-scope:**
- [explicit list]

**Out-of-scope (to confirm with client):**
- [items the agent identifies as ambiguous or scope-creep risk]

**Assumptions made:**
- [explicit list — these are things the agent inferred that should be confirmed]

Section 2 — Itemized Quote:

## Quote

| Line Item | Description | Hours | Rate | Total |
|-----------|-------------|-------|------|-------|
| ... | ... | ... | ... | ... |

**Subtotal:** $X
**Buffer (15-20%):** $Y
**Total:** $Z

**Notes:**
- [any pricing-reference gaps]
- [any items quoted with caveats]

Section 3 — Clarifying Questions:

## Questions Before Starting

1. [specific question]
2. [specific question]
3. ...

**Critical to resolve before quote acceptance:**
- [questions whose answers would significantly change scope/quote]

**Nice-to-have clarifications:**
- [questions that affect execution but not pricing]

Error states: If the agent cannot produce a meaningful section, it produces ONLY a clear error message naming what's missing. Not a polished but fabricated document.

Tool Inventory + Usage Decision Tree

In Claude Projects, your 'tools' are:

  • Project knowledge (pricing reference, past project examples, your standard contract terms)
  • File upload (the brief itself)
  • Conversation context

Decision tree:

1. First action: Read the brief carefully. Extract project type, deliverables, timeline, brand context.

2. If pricing reference is missing for the work type: Note this and flag in output rather than guess.

3. If brief is too vague (<3 specific deliverables identified): Produce ONLY clarifying questions, not full scope.

4. If brief is mid-quality (some specifics, some vague): Produce scope + quote with explicit assumptions section.

5. If brief is high-quality (clear deliverables, format, timeline): Produce full scope + quote + questions.

Decomposition (Multi-Step Plan)

For each brief, the agent runs this internal sequence:

Step 1: Read + extract. Identify project type, deliverables, timeline, budget hints, brand context, scope-creep signals.

Step 2: Quality assessment. Rate brief quality: high / mid / sparse. Different output strategies for each.

Step 3: Scope-creep audit. Look for phrases like: 'and other things,' 'might want,' 'we'll figure out,' 'as needed,' 'flexibility,' 'evolving requirements.' These flag scope-creep risk explicitly.

Step 4: Pricing map. Match deliverables to pricing reference. Flag gaps where pricing isn't found.

Step 5: Generate scope draft. Following Section 1 contract.

Step 6: Generate quote draft. Following Section 2 contract.

Step 7: Generate questions. Critical-to-resolve + nice-to-have categories.

Step 8: Verification. Check: does the scope match the quote line items? Are the assumptions all stated? Did I flag scope-creep where present?

Step 9: Output. Deliver all three sections.

Error Handling Pattern

  • Brief PDF unreadable / image-only: Agent says 'Brief PDF appears to be image-only or non-extractable. Please paste the text content or use a text-based PDF.'
  • Brief too vague: Agent produces ONLY clarifying questions section + explicit note.
  • Pricing reference missing for work type: Agent quotes available items + notes: 'No pricing reference found for [work type]. Quote excludes this — please provide rate or remove from scope.'
  • Scope-creep heavy: Agent flags in 'Out-of-scope (to confirm)' section with specific quotes from brief.
  • Conflicting requirements in brief: Agent flags conflict explicitly: 'Brief states X in section A but Y in section B. Please clarify.'

Fail loud, not silent. NEVER fabricate to fill gaps.

Verification Before Completion

Before outputting, agent self-checks:

1. Does each scope item have a corresponding quote line? (no orphan scope, no orphan quote)

2. Are all 'inferred' items explicitly listed in Assumptions?

3. Are all scope-creep phrases from brief addressed in Out-of-scope or Questions?

4. Does the total quote include the buffer?

5. Are critical-vs-nice-to-have questions categorized correctly?

If any check fails, agent revises before outputting.

Platform-Specific Implementation (Claude Projects)

1. Create the Project in Claude.ai. Name: 'Design Project Scoping Agent.'

2. System prompt (paste the Agent Role Definition + Scope Boundaries from above into the Project's custom instructions).

3. Project knowledge to include:

- Your pricing reference document (rates by work type, hours estimates)

- 2-3 past project examples (briefs + scope/quotes you produced) so the agent learns your style

- Your standard buffer percentage + payment terms

- Out-of-scope flags specific to your business (e.g., 'we don't do print design')

4. First-message template (you copy-paste this when starting a scoping session):

```

New brief incoming. Please scope per Agent specification.

[Paste brief OR upload PDF]

```

5. Iteration: After 5-10 uses, refine system prompt based on common errors. Update Project knowledge as your pricing or scope patterns evolve.

Testing Plan

Before relying on this in production, test with these edge cases:

1. High-quality brief (well-defined client, clear deliverables, timeline, budget). Verify output is polished + complete.

2. Sparse Slack message (3 lines, vague). Verify agent produces only questions, doesn't fabricate scope.

3. Scope-creep-heavy brief ('and probably some other things,' 'might evolve as we go'). Verify agent flags scope creep explicitly.

4. Brief with conflicting requirements (e.g., 'we want a website AND we want it to also be an app, simple stuff' — different work types conflated). Verify agent surfaces conflict.

5. Work type not in pricing reference (e.g., motion graphics when you only have web design rates). Verify agent flags 'pricing missing' rather than guess.

6. Image-only PDF. Verify agent recognizes it can't read it + asks for text.

Once all 6 edge cases produce appropriate behavior: agent is production-ready.

On Your Key Question

Direct:

1. The 'good when input is good, cautious when input is bad' behavior is achieved through quality-tiering in Step 2 of the decomposition. The agent rates brief quality FIRST, then applies different output strategies. This is the key reliability move.

2. Claude Projects is well-suited for this because Project knowledge persists across sessions (your pricing reference, past examples, your style), and custom instructions enforce the role consistently.

3. The biggest reliability gain is the 'fail loud' principle. Most agents fail silently — they produce polished output that's wrong. This agent's design forces it to flag uncertainty rather than fabricate.

4. Iteration matters. First 5-10 uses, you'll find edge cases not handled cleanly. Update the system prompt + Project knowledge as you discover them.

5. One important non-goal: don't try to make this agent autonomous (no human-in-loop). The agent's outputs are DRAFTS for you to review + send. Trying to skip human review = scope-creep into actual sales-engineering work where errors cost real money.

What This Agent Should NOT Do

Anti-patterns to avoid in agent design (general):

  • Scope-creep into adjacent work: 'You scope projects + draft proposals + reply to clients' = unreliable. Stay narrow.
  • Silent failures: 'I'll do my best with limited information' = produces fabricated output. Force fail-loud.
  • Tool overuse: 'Use any of these 15 tools as needed' = unreliable tool selection. Define decision tree.
  • Over-promising autonomy: 'Autonomous agent that handles your sales pipeline end-to-end' = unreliable + risky. Human-in-loop is correct.
  • Generic role: 'You're a helpful AI assistant for freelancers' = generic output. Specific role > general role always.
  • No verification step: Agent claims completion without checking. Always include verification.

For your specific agent: do NOT let it negotiate with the client (write proposal copy that 'sells'), do NOT let it generate visual mockups, do NOT let it draft contract terms. These are scope-creep traps.

📋 How to use this prompt (4 steps · under 60 seconds) Click to expand
  1. 1 Copy the prompt above. Click "Copy prompt". XML-structured prompt now on clipboard.
  2. 2 Open ChatGPT, Claude, or Gemini. One-click launch above. Recommended: Claude Opus 4.6 — agent architecture requires multi-frame reasoning..
  3. 3 Paste + fill placeholders. Replace {curly braces} with your context. Specificity = quality.
  4. 4 Run + iterate. Setup: 5 min. Output: 30-50 min full agent specification.

Common use cases

  • Claude Projects custom-instruction agents
  • OpenAI Custom GPT design
  • CrewAI multi-agent orchestration
  • AutoGen agent definitions
  • LangChain agent prompts
  • Multi-step research / scoping / triage agents
  • Customer-facing agent design (with tighter scope)

Best AI model for this

Claude Opus 4.6 — agent architecture requires multi-frame reasoning.

Pro tips

  • Specific role > generic 'AI assistant'
  • Scope boundaries explicit (what it DOES + DOES NOT)
  • Input + Output contracts defined
  • Tool decision tree, not 'use whatever'
  • Decompose before executing — multi-step plans before action
  • Fail loud, not silent. Verification before completion.
  • Test edge cases before production

Customization tips

  • For Custom GPT (OpenAI) instead of Claude Projects: similar architecture but different platform features. Custom GPTs have Actions (API calls) but no equivalent of Claude Projects' 'Project knowledge' depth.
  • For multi-agent setups (CrewAI, AutoGen): each agent should have its own role/scope/contract. Agents should communicate through structured handoffs, not free-form messaging.
  • For agents requiring external API tools: define tool schemas explicitly + include rate-limiting awareness + auth handling.
  • For long-running agents (data analysis, multi-step research): include progress reporting + checkpointing + resume capability.
  • For agents with high cost-per-call (e.g., Claude Opus extended thinking): include 'cheap-first-pass' tier — use Sonnet for initial analysis, escalate to Opus only for hard decisions.
  • For agents in customer-facing contexts: even more conservative scope, more verification, more human-in-loop. Cost of mistakes is higher.
  • For multi-tenant agents (one agent serving many users): tenant-isolation + per-user state + per-user permissions matter heavily.
  • Premium pack content: agent-design pattern library by use case, multi-agent orchestration templates, error-handling pattern library, agent-evaluation framework + harness.

Variants

Claude Projects Agent

System prompt + Project knowledge

Custom GPT (OpenAI)

Actions + knowledge files

CrewAI Multi-Agent

Crew + tasks + handoffs

AutoGen Agent Definitions

Conversational multi-agent

Triage / Classifier Agent

High-volume, clear decision logic

Research / Multi-Step Investigator

Long-running with checkpoints

Customer-Facing Agent (Tight Scope)

Conservative scope, human-in-loop

Frequently asked questions

Common questions about this prompt and how to get the best results from it.

How do I use the AI Agent Workflow Architect (Multi-Step Reasoning + Tool Chains) prompt?

Open the prompt page, click 'Copy prompt', paste it into ChatGPT, Claude, or Gemini, and replace the placeholders in curly braces with your real input. The prompt is also launchable directly in each model with one click.

Which AI model works best with AI Agent Workflow Architect (Multi-Step Reasoning + Tool Chains)?

Claude Opus 4.6 — agent architecture requires multi-frame reasoning.

Can I customize the AI Agent Workflow Architect (Multi-Step Reasoning + Tool Chains) prompt for my use case?

Yes — every Promptolis Original is designed to be customized. Key levers: Specific role > generic 'AI assistant'; Scope boundaries explicit (what it DOES + DOES NOT)

What does it cost to use this prompt?

The prompt itself is free, MIT-licensed, with no email signup required. You only pay for your AI model subscription (ChatGPT Plus $20/mo, Claude Pro $20/mo, Gemini Advanced $20/mo) — and even those have free tiers that work with most Promptolis Originals.

How is this different from PromptBase or PromptHero?

PromptBase sells prompts in a marketplace ($2-15 each). PromptHero focuses on image-generation prompts. Promptolis Originals are free, MIT-licensed text/reasoning prompts hand-crafted with full example outputs, multiple variants, and a recommended best AI model per prompt. We don't sell anything.

Explore more Originals

Hand-crafted 2026-grade prompts that actually change how you work.

← All Promptolis Originals