Source: ai-research (web research, 2026-04-11), raw/instructor-8lsy243ftffjjy1cx9lm3o2bw-public-1773274827-Claude+Certified+Architect+–+Foundations+Certification+Exam+Guide.pdf, raw/guide_en.MD
A practical study guide for the Claude Certified Architect — Foundations (CCA-F) exam. Organized by the 5 exam domains in recommended study order, with the key principles, patterns, and tradeoffs you need to internalize. This guide synthesizes the official exam guide, the Architect’s Playbook, community prep materials, and architectural patterns that recur across exam questions. For the complete task-by-task breakdown, see the Official Exam Guide.
Recommended Study Order
Study the highest-weighted domains first, but do not skip any — the exam draws from all five:
- Agentic Architecture (27%) — largest weight, most complex
- Tool Design & MCP Integration (18%) — tightly coupled with agentic patterns
- Prompt Engineering & Structured Output (20%) — foundational for everything else
- Claude Code Configuration (20%) — practical setup knowledge
- Context & Reliability (15%) — smallest weight but tests subtle production issues
The reason to study Tool Design before Prompt Engineering: agentic systems depend on well-designed tools, so understanding tool patterns first makes the prompt engineering domain easier to absorb.
Domain 1: Agentic Architecture (27%)
This is the make-or-break domain. Anthropic weights it highest because agentic failures are the costliest in production. The official exam guide defines 7 task statements for this domain.
Official Task Statements
- Task 1.1: Design and implement agentic loops (
stop_reason="tool_use"vs"end_turn", tool results in conversation, anti-patterns like NL parsing for termination) - Task 1.2: Orchestrate coordinator-subagent patterns (hub-and-spoke, isolated context, risks of overly narrow decomposition)
- Task 1.3: Configure subagent invocation (
Tasktool,allowedTools,AgentDefinition,fork_session, context must be explicitly passed) - Task 1.4: Implement multi-step workflows (programmatic prerequisites vs prompt guidance, structured handoff summaries)
- Task 1.5: Apply Agent SDK hooks (
PostToolUsefor transformation/compliance, deterministic hooks vs probabilistic prompts) - Task 1.6: Design task decomposition (fixed pipelines vs dynamic adaptive, prompt chaining for predictable tasks, dynamic for open-ended)
- Task 1.7: Manage session state (
--resume,fork_session, starting fresh with summaries vs resuming with stale context)
See Official Exam Guide for complete knowledge/skill items for each task.
What to know
- Subagent spawning patterns — when to use parallel workers vs. sequential chains vs. orchestrator patterns. Understand the tradeoffs: parallel is faster but harder to coordinate; sequential is simpler but slower
- Workflow decomposition — how to break a complex task into subtasks that can be independently completed. Know when decomposition adds value vs. when it introduces unnecessary overhead
- Tool selection and decision complexity — an agent with 50 tools performs worse than one with 5 well-chosen tools. Dynamic scoping (giving an agent only the tools relevant to its current subtask) is a key pattern
- Fallback loop design — what happens when an agent step fails? Design for graceful degradation, not silent failure. Know the difference between retry, fallback to alternate tool, and escalation to human
- Cost management — Batch API for non-urgent work at 50% discount; real-time API for latency-sensitive paths. Know when each is appropriate
- Hooks for compliance —
PostToolUsehooks provide deterministic enforcement; prompts are probabilistic. The Playbook shows that even emphatic prompts (“CRITICAL POLICY: NEVER”) fail ~3% of the time
Key tradeoffs to understand
- Autonomy vs. control: more autonomous agents are more capable but harder to debug
- Parallelism vs. coordination: parallel subagents are faster but may produce conflicting outputs
- Tool breadth vs. precision: more tools increase capability but decrease reliability
- Cost vs. latency: Batch API is cheaper but slower; real-time is immediate but expensive
Hands-on practice
Run these cookbooks: customer_service_agent.ipynb, orchestrator_workers.ipynb, basic_workflows.ipynb, evaluator_optimizer.ipynb
Related wiki articles
- Claude Code Subagents — the stable pattern for parallel work
- Claude Code Agent Teams — experimental multi-instance coordination
- Claude Managed Agents — hosted agent runtime
- Claude Agent Hierarchy
Domain 2: Tool Design & MCP Integration (18%)
Tools are how agents interact with the world. Poorly designed tools are the #1 cause of agent failures after bad prompts. The official exam guide defines 5 task statements for this domain.
Official Task Statements
- Task 2.1: Design effective tool interfaces (descriptions are the primary selection mechanism, ambiguous descriptions cause misrouting, keyword-sensitive system prompts)
- Task 2.2: Implement structured errors (
isErrorflag, transient/validation/business/permission categories,errorCategory+isRetryable+ human description) - Task 2.3: Distribute tools across agents (too many tools degrades reliability — 18 instead of 4-5, scoped access,
tool_choiceauto/any/forced) - Task 2.4: Integrate MCP servers (
.mcp.jsonproject-level vs~/.claude.jsonuser-level, env variable expansion, MCP resources as content catalogs) - Task 2.5: Apply built-in tools effectively (Grep for content search, Glob for file patterns, Read/Write/Edit, Edit fails on non-unique matches → Read+Write fallback)
See Official Exam Guide for complete knowledge/skill items for each task.
What to know
- JSON schema for semantic validation — use schema constraints to prevent invalid inputs before they reach your backend. Enums, pattern matching, and required fields are your first line of defense
- Enum constraints for ambiguous inputs — when a field could be interpreted multiple ways, constrain it to an enum.
status: enum["active", "inactive", "pending"]notstatus: string - Machine-readable identifiers — always use IDs, never natural language names, in tool parameters.
customer_id: "cust_abc123"notcustomer: "Acme Corp" - Lookup/action tool pairs — instead of one tool that takes a name and performs an action, split into: (1) a lookup tool that returns matching IDs, and (2) an action tool that operates on an ID. This eliminates ambiguity
- Pagination — never return all results. Return a page with metadata (total count, next cursor, has_more). Agents need to know they have not seen everything
- Error handling — tools should return structured error objects, not raw error strings. Include error code, message, and suggested recovery action. The Playbook shows the correct pattern: return error in tool result with
isError: true,errorCategory, andisRetryable— never throw exceptions or return empty strings - Tool chaining reliability — when tool B depends on tool A’s output, validate A’s output before passing it to B. Do not assume success
- MCP tool specificity — the Playbook warns that agents default to built-in tools (Grep) over broad custom tools. Split monolithic tools into granular single-purpose tools with detailed descriptions
Key patterns
- Normalize before presenting — if your agent calls 3 different external APIs that return data in different formats, normalize the responses into a consistent schema before the agent sees them
- Dual-field validation — have the model output both the result and a confidence/verification field. If they disagree, trigger review
- Idempotent tools — design tools so calling them twice with the same input produces the same result. This makes retry logic safe
Hands-on practice
Run these cookbooks: calculator_tool.ipynb, tool_choice.ipynb, parallel_tools.ipynb, programmatic_tool_calling_ptc.ipynb
Related wiki articles
Domain 3: Prompt Engineering & Structured Output (20%)
The exam tests prompt engineering as an architectural tool, not a creative exercise. The focus is on reliability and structure. The official exam guide defines 6 task statements for this domain.
Official Task Statements
- Task 4.1: Design prompts with explicit criteria (vague instructions fail, explicit criteria > “be conservative”, categorical > confidence-based filtering)
- Task 4.2: Apply few-shot prompting (most effective for consistent formatting, demonstrate ambiguous cases with reasoning, show varied document structures)
- Task 4.3: Enforce structured output via tool_use (JSON schemas,
tool_choiceauto/any/forced, schemas prevent syntax but not semantic errors, optional/nullable fields prevent fabrication) - Task 4.4: Implement validation/retry loops (append specific errors on retry, retries ineffective for missing info,
detected_patternfields for analysis) - Task 4.5: Design batch processing strategies (Message Batches API 50% savings, 24hr window, no multi-turn tool calls in batch,
custom_idfor correlation) - Task 4.6: Design multi-instance review (self-review limitations, independent instances more effective, per-file + cross-file integration passes)
See Official Exam Guide for complete knowledge/skill items for each task.
What to know
- Structured output schemas — define the exact JSON structure you expect. This is not optional for production systems; it prevents hallucinated fields and missing data
- JSON schema to prevent hallucinations — when the model must choose from known options, use enum constraints in the output schema. The model cannot hallucinate a value that is not in the enum. Add
"other"+ detail string for extensible categories (the Playbook’s “Resilient Catch-All” pattern) - Prompt refinement methodology — the exam tests a specific principle: optimize the prompt before scaling. If your prompt works 80% of the time, fix the prompt before adding retry logic or fallback chains
- Chain-of-thought vs. direct prompting — know when each is appropriate. CoT is better for complex reasoning; direct is better for simple extraction. CoT adds latency and cost
- Self-correction patterns — have the model check its own output against the input constraints. The Playbook defines the schema redundancy pattern: extract
calculated_totalalongsidestated_totalto detect extraction errors - Limits of retry — the Playbook demonstrates that retries work for formatting errors but are useless for missing information. Know when to fail fast
- Multi-instance review — self-review is limited because the model retains its reasoning context. Independent review instances without prior context are more effective
Key principle
The exam prioritizes structural/deterministic solutions over probabilistic ones. If you can enforce a constraint via JSON schema, that answer beats a prompt-based approach every time. Prompts are probabilistic; schemas are deterministic.
Hands-on practice
Run these cookbooks: extracting_structured_json.ipynb, batch_processing.ipynb, building_evals.ipynb, tool_choice.ipynb
Related wiki articles
Domain 4: Claude Code Configuration (20%)
Practical knowledge of how to set up and manage Claude Code for production use. The official exam guide defines 6 task statements for this domain.
Official Task Statements
- Task 3.1: Configure CLAUDE.md hierarchy (user/project/directory levels,
@importfor modularity,.claude/rules/for topic-specific) - Task 3.2: Create custom slash commands and skills (
.claude/commands/project-scoped,~/.claude/commands/personal,SKILL.mdfrontmatter withcontext: fork,allowed-tools,argument-hint) - Task 3.3: Apply path-specific rules (
.claude/rules/with YAMLpathsglob patterns for conditional loading) - Task 3.4: Determine plan mode vs direct execution (plan mode for complex/multi-file/architectural, direct for simple/clear-scope)
- Task 3.5: Apply iterative refinement (concrete I/O examples, test-driven iteration, interview pattern, batching interacting vs sequential independent issues)
- Task 3.6: Integrate into CI/CD (
-pflag for non-interactive,--output-format json+--json-schema, session context isolation for independent review)
See Official Exam Guide for complete knowledge/skill items for each task.
What to know
- CLAUDE.md configuration — how project-level instructions work, what goes in them, how they override defaults. The hierarchy is: user-level (
~/.claude/CLAUDE.md), project-level (.claude/CLAUDE.mdor root), directory-level (subdirectory files) - Permission and environment setup — tool permissions, allowed/denied commands, environment variables
- Hook configuration — pre-commit hooks, automation triggers, quality gates
- Model selection — when to use different Claude models (cost vs. capability tradeoffs)
- Session management — how context persists across sessions, compaction behavior, context window limits. The Playbook covers the scratchpad pattern and dynamic environment resumption
- Cost optimization — prompt caching, Batch API usage, token management
- Path-specific rules —
.claude/rules/files with YAML frontmatterpathsfields are more efficient than multiple directory-level CLAUDE.md files for cross-cutting conventions - CI/CD integration — the
-pflag prevents interactive hangs;--output-format json+--json-schemaproduces machine-parseable output; session context isolation means independent review instances catch more issues
Key principle
Configuration should make the right thing easy and the wrong thing hard. Good CLAUDE.md files encode team conventions so the agent follows them automatically.
Hands-on practice
Examine the cookbooks repo’s own .claude/ directory structure: root CLAUDE.md, .claude/commands/, .claude/skills/, and .claude/agents/ for real-world examples
Related wiki articles
Domain 5: Context & Reliability (15%)
The smallest domain by weight but tests subtle issues that only show up in production. The official exam guide defines 6 task statements for this domain.
Official Task Statements
- Task 5.1: Implement context preservation (progressive summarization risks, “lost in the middle”, tool result accumulation, case facts blocks)
- Task 5.2: Design escalation patterns (customer request/policy gap/no progress triggers, immediate vs investigation-first, sentiment-based unreliable)
- Task 5.3: Implement error propagation (structured error context, access failure vs empty results, suppress errors = anti-pattern)
- Task 5.4: Navigate large codebases (context degradation, scratchpad files, subagent delegation, state exports for crash recovery,
/compact) - Task 5.5: Design human review workflows (aggregate metrics mask field-level issues, stratified random sampling, field-level confidence scores)
- Task 5.6: Maintain information provenance (claim-source mappings, conflicting sources annotated not arbitrarily selected, temporal data needs dates)
See Official Exam Guide for complete knowledge/skill items for each task.
What to know
- “Lost in the middle” effect — models pay more attention to the beginning and end of their context window. Critical information placed in the middle gets lower attention. Know how to structure context to mitigate this
- Context bloat prevention — every token in the context window costs money and reduces quality. Prune aggressively: summarize old messages, remove completed subtask details, never load an entire codebase. The Playbook defines the Tool Context Pruning pattern: application-side filters extract only relevant fields from verbose API responses
- Data provenance — in multi-step agent workflows, track where each piece of information came from. Claim-source mappings are required. Conflicting sources must be annotated, not arbitrarily selected
- Session resumption — when an agent resumes after a pause, the external world may have changed. The Playbook recommends filtering out previous
tool_resultmessages on resumption, keeping only human/assistant turns to force re-fetching current data - Reliability patterns — retries with backoff, circuit breakers for flaky tools, health checks, graceful degradation
- Escalation design — the Playbook shows two paths: immediate escalation for explicit human requests (honor immediately, do not ask for clarification) vs context-gathering for policy issues (call
get_customerbefore escalating) - Human review calibration — aggregate metrics mask field-level issues. A 95% overall accuracy can hide 60% accuracy on a specific field. Analyze by document type AND field before reducing review
Key tradeoffs
- Context size vs. quality: more context is not always better; irrelevant context degrades performance
- Freshness vs. cost: re-fetching data ensures freshness but costs tokens; caching is cheaper but may be stale
- Observability vs. overhead: logging everything aids debugging but adds latency and cost
Hands-on practice
Run these cookbooks: prompt_caching.ipynb, session_memory_compaction.ipynb, automatic-context-compaction.ipynb, using_citations.ipynb
Related wiki articles
- Claude Code Subagents — scoped context for parallel workers
- Claude Code Skills Ecosystem
Cross-Domain Principles
These principles apply across all 5 domains and are the most frequently tested concepts:
- Deterministic over probabilistic — if you can solve it with code/schema/rules, that beats a prompt-based solution
- Optimize before scaling — fix the prompt before adding retries; fix the tool before adding fallbacks
- Machine IDs over natural language — structured identifiers prevent ambiguity
- Paginate everything — never return unbounded result sets
- Normalize inputs — agents should see consistent data formats regardless of source
- Fail loudly — silent failures compound; structured error responses enable recovery
- Scope dynamically — give agents only the tools and context they need for the current step
Out-of-Scope Topics
The following are explicitly NOT on the CCA-F exam (per the study guide source). Use this to narrow your study focus:
- Fine-tuning or training custom models
- API authentication, billing, or account management
- Detailed language/framework implementation (beyond tool/schema config)
- Deploying or hosting MCP servers (infrastructure, networking)
- Claude internals (architecture, training process, model weights)
- Constitutional AI, RLHF, or safety training methodologies
- Embedding models or vector database implementation details
- Computer use, Vision, or Streaming API
- Rate limiting, quotas, OAuth, API key rotation
- Cloud-provider-specific configurations (AWS, GCP, Azure)
- Performance benchmarks or model comparison metrics
- Token counting algorithms or tokenization specifics
Study Tips from the Community
Based on patterns from community prep materials and early test-takers:
- Build a real project — the exam tests architectural judgment that comes from hands-on experience, not memorization
- Practice tradeoff reasoning — for every pattern, know when it is the wrong choice. The exam rarely asks “what is X” — it asks “given these constraints, which approach is best”
- Focus on the ‘why’ — understanding why a pattern exists is more valuable than memorizing how to implement it
- Study the scenario format — the exam presents 4 production scenarios and asks questions within those contexts. Practice reading a system description and identifying architectural decisions
- Time management — 60 questions means roughly 1-2 minutes per question. Scenario-based questions take longer; factor that in
Key Takeaways
- Agentic Architecture (27%) and Tool Design (18%) together account for 45% of the exam — they deserve the most study time
- The exam tests tradeoff thinking, not definitions — prepare by understanding when patterns apply and when they do not
- Deterministic solutions (schemas, enums, machine IDs) are consistently preferred over probabilistic solutions (prompts, retries)
- Build something real to internalize the patterns — reading about architecture is not the same as doing it
- The 4-random-scenario format means you need broad coverage, not deep expertise in one area
- Community resources (GitHub repo, claudecertifications.com) supplement the official Anthropic Academy courses well
Try It
- Complete Anthropic Academy courses — all 13, starting with Agentic Architecture and Tool Design
- Work through practice questions — see certification-practice-questions for domain-by-domain patterns
- Build a multi-agent Claude Code project — include subagents, MCP tools, structured output with JSON schemas, and explicit context management. Touch all 5 domains
- Practice scenario analysis — take any production system you have worked on and identify: what are the agent boundaries, what should be deterministic vs. prompt-based, where are the failure modes
- Review the GitHub exam prep repo —
github.com/avidevelops/claude-architect-exam-prepfor scenario-style questions
Related
- CCA-F Official Exam Guide
- CCA-F Technical Reference — deep technical content with code patterns and implementation details
- CCA-F Practice Exam (60 Questions) — full-length practice test matching exam format
- The Architect’s Playbook
- Anthropic Claude Cookbooks
- Claude Certified Architect — Foundations (CCA-F)
- CCA-F Practice Questions by Domain
- The Complete Guide to Building Skills for Claude
- Claude Code Subagents
- Essential MCP Servers for 2026
- Skill Design Patterns
- Claude Agent Hierarchy
Open Questions
Does Anthropic publish an official study guide or reading list beyond the Academy courses?Answered: Yes. The official exam guide PDF and The Architect’s Playbook PDF are now available. See official-exam-guide and architects-playbook- How closely do the community practice questions match the actual exam difficulty?
- Are the Anthropic Academy courses updated as the exam evolves?
What hands-on experience level is assumed?Answered: The official guide specifies 6+ months of practical experience building with Claude APIs, Agent SDK, Claude Code, and MCP