Quiz · 50 questions · self-graded

The quiz.

Fifty questions. Mostly short-answer. Some marked hard. You write your answer, reveal the correct one, rate yourself honestly. Your progress is saved in your browser. No one is watching.

50 questions 12 modules 12 hard ★
Answered 0 / 50 Score 0.0 / 50

Module 1 — Foundations

01 Module 1 · Foundations standard
True or false: A skill is just a markdown file. Justify your answer in one sentence.
Answer

False. A skill is a folder. SKILL.md is the entrypoint, but the entire file system is intentional context engineering and progressive disclosure.

Source: Shihipar (LinkedIn) · code.claude.com/skills

How did you do?
02Module 1 · Foundationsstandard
Name the four install scopes for skills in Claude Code, in override-priority order (highest priority first).
Answer

Enterprise > Personal > Project > Plugin. Plugin skills use a plugin-name:skill-name namespace so they can't conflict with other levels.

Source: code.claude.com/skills §"Where skills live"

How did you do?
03Module 1 · Foundationshard ★
Anthropic frames skills primarily as A) expertise import, B) procedural knowledge, or C) capability uplift? Cite where you got this.
Answer

B — Procedural knowledge. Verbatim from support.claude.com/12512176: "Skills provide procedural knowledge—instructions for how to complete specific tasks or workflows." Not framed by Anthropic as expertise import.

Source: support.claude.com/12512176-what-are-skills

How did you do?

Module 2 — Open standard

04Module 2 · Open standardstandard
Approximately how many tools currently support the agentskills.io open standard? Within 5 is fine.
Answer

36+ tools — Junie, Gemini CLI, GitHub Copilot, VS Code, Cursor, OpenAI Codex, Claude Code, Roo Code, Mistral Vibe, Databricks, Snowflake, Block (Goose), JetBrains, ByteDance, and more.

Source: agentskills.io client showcase

How did you do?
05Module 2 · Open standardstandard
What two YAML frontmatter fields are REQUIRED by the open standard?
Answer

name and description. Everything else (license, compatibility, metadata, allowed-tools) is optional.

Source: agentskills.io/specification

How did you do?
06Module 2 · Open standardstandard
Name three Anthropic Claude Code frontmatter fields that are NOT in the open standard.
Answer

Any three of: when_to_use, argument-hint, arguments, disable-model-invocation, user-invocable, model, effort, context: fork, agent, hooks, paths, shell.

Source: code.claude.com/skills frontmatter table vs agentskills.io/specification

How did you do?
07Module 2 · Open standardhard ★
What constraint does the open standard impose on the relationship between the name field and the directory?
Answer

The name field MUST match the parent directory name. Verbatim from agentskills.io/specification: "Must match the parent directory name." Not documented anywhere else.

Source: agentskills.io/specification

How did you do?

Module 3 — Architecture

08Module 3 · Architecturestandard
List the three tiers of progressive disclosure, in load order.
Answer

Tier 1 — Metadata (name + description, ~100 tokens, always loaded). Tier 2 — Instructions (full SKILL.md body, <5,000 tokens, loaded on activation). Tier 3 — Resources (scripts, references, assets — loaded only when needed).

Source: agentskills.io/specification + Cookbook 01

How did you do?
09Module 3 · Architecturehard ★
Walk through the four steps of skill triggering, from prompt to file load. (Per Anthropic engineering blog.)
Answer

Step 1: Description sits in system prompt. Step 2: Claude DECIDES whether to use the skill (description-driven gate). Step 3: Claude INVOKES BASH TOOL to read SKILL.md (an actual file read, not context injection). Step 4: Claude chooses which BUNDLED FILES to load based on what SKILL.md says.

Source: Anthropic engineering blog — "Equipping agents for the real world with Skills"

How did you do?
10Module 3 · Architecturestandard
Roughly how many tokens does an unused skill cost in the system prompt? How many when invoked?
Answer

Unused: ~30 tokens (metadata only). Invoked: ~5,000 additional tokens (full SKILL.md body). Total per file-generating request: 6,000–10,000 tokens typical.

Source: Cookbook 01

How did you do?
11Module 3 · Architecturestandard
When auto-compaction fires in Claude Code, what's the per-skill token budget for re-attached skills, and what's the combined budget?
Answer

Per-skill: 5,000 tokens kept (the first 5,000 of each re-attached skill). Combined: 25,000 tokens total budget for re-attached skills, fills from most recent. Older skills can be dropped after compaction.

Source: code.claude.com/skills §"Skill content lifecycle"

How did you do?
12Module 3 · Architecturestandard
True or false: Once a skill is invoked in a Claude Code session, it gets re-read from disk on every subsequent turn. Justify.
Answer

False. Verbatim: "the rendered SKILL.md content enters the conversation as a single message and stays there for the rest of the session. Claude Code does not re-read the skill file on later turns." Frozen at invocation.

Source: code.claude.com/skills

How did you do?

Module 4 — Authoring mechanics

13Module 4 · Authoringstandard
What's the SKILL.md body line limit per Anthropic's official checklist?
Answer

500 lines. Hard limit, called out in the official Anthropic checklist. (The 150-line figure that floats around is Vercel's empirical median, not the spec.)

Source: Anthropic best-practices §"Token budgets"

How did you do?
14Module 4 · Authoringstandard
Are all .md files in a skill folder loaded automatically when the skill activates, or only SKILL.md and REFERENCE.md? Cite the source.
Answer

All .md files in the root directory load. Verbatim from Cookbook 03: "All .md files in the root directory will be available to Claude when the skill is loaded." Subfolder .md files load on-demand.

Source: Cookbook 03

How did you do?
15Module 4 · Authoringstandard
Reference files should be at most ___ level(s) deep from SKILL.md. Why?
Answer

One. Reason: Claude may partially read deeply-nested files (e.g., via head -100), missing complete information.

Source: Anthropic best-practices §"Avoid deeply nested references"

How did you do?
16Module 4 · Authoringhard ★
Name the substitution variable that solves the "data wiped on skill upgrade" problem, and cite which Anthropic-team author surfaced it.
Answer

${CLAUDE_PLUGIN_DATA}, surfaced by Thariq Shihipar (Anthropic Claude Code team) on LinkedIn. Skill directory data is wiped on upgrade; this variable points to the stable per-plugin folder. Undocumented in official Anthropic public docs.

Source: Shihipar — "Lessons from Building Claude Code" (LinkedIn)

How did you do?

Module 5 — Description & activation

18Module 5 · Descriptionstandard
What's the description-field character cap (the field alone, not the listing)?
Answer

1024 characters.

Source: Anthropic best-practices §"YAML Frontmatter requirements"

How did you do?
19Module 5 · Descriptionstandard
What's the COMBINED description + when_to_use character cap in the skill listing?
Answer

1,536 characters. The combined description + when_to_use text is capped at 1,536 chars in the skill listing for context-management.

Source: code.claude.com/skills

How did you do?
20Module 5 · Descriptionhard ★
Recite the Variant C directive template VERBATIM (the structure, not a specific instance).
Answer

<Domain> expert. ALWAYS invoke this skill when the user asks about <trigger topics>. Do not <alternative action> directly — use this skill first.

Source: Seleznov 650-trial study

How did you do?
21Module 5 · Descriptionstandard
In Seleznov's 650-trial study, what's the catastrophic failure cell — which variant + condition combo collapses to 37% activation?
Answer

Variant A (Passive) + C3 (Hook condition): activation collapses to 37.0%. The single catastrophic failure cell. CLAUDE.md "rescues" Variant A back to 100% in C4 (both hook AND CLAUDE.md).

Source: Seleznov

How did you do?
22Module 5 · Descriptionhard ★
The CMH odds ratio for Variant C vs Variant A is approximately ___. The Cohen's h is ___. The p-value is ___.
Answer

OR ≈ 20.6, Cohen's h = 1.83 ("huge effect" by Cohen's conventions), p < 0.0001. From Cochran-Mantel-Haenszel test, Variant C vs Variant A.

Source: Seleznov

How did you do?
23Module 5 · Descriptionstandard
What's the "directive saturation risk" Seleznov flagged as a future failure mode that hasn't been tested yet?
Answer

If ALL skills use "ALWAYS invoke" language with overlapping triggers, the directive may lose force through dilution. Multiple skills claiming the same keywords could confuse Claude about which to invoke. No controlled test exists yet — open opportunity for someone to publish first.

Source: Seleznov §"Limitations"

How did you do?

Module 6 — Two reliability problems

24Module 6 · Reliabilitystandard
Marc Bara identifies TWO reliability problems. Name them and give a one-sentence description of each.
Answer

(1) Activation failure: Claude does not invoke the skill at all and defaults to its own approach. (2) Execution failure: Claude loads the skill but skips internal procedural steps that delay output without producing visible content.

Source: Marc Bara — "Claude Skills Have Two Reliability Problems, Not One"

How did you do?
25Module 6 · Reliabilityhard ★
What three factors does Bara cite as the mechanism for execution failure?
Answer

(1) User request sits at end of context window — recency = strongest attention weight. (2) Skills are "further back, procedural, and meta-level." (3) RLHF reinforced "produce the requested output directly is the pattern that got rewarded most consistently." Result: verification steps lose because they delay output, add no visible content, and sit at maximum distance from initial prompt.

Source: Bara

How did you do?
26Module 6 · Reliabilitystandard
Recite the verbatim "before / after" rewrite pattern Bara recommends for execution-skip prevention.
Answer

Before (skippable): "Before delivering, verify that every milestone aligns with the stated scope."

After (visible-output required): "Do not deliver the final charter until you have output a verification block listing each milestone and the in-scope deliverable it maps to. Flag any milestone that references an out-of-scope item."

Principle: force visible output from verification steps.

Source: Bara

How did you do?
27Module 6 · Reliabilitystandard
True or false: A controlled experiment for execution-fidelity (analogous to Seleznov's 650-trial activation study) has been published. Justify.
Answer

False. Bara verbatim: "No one has run a controlled experiment on step-level execution the way Seleznov did for activation. The 650-trial methodology exists. The step-execution version of that experiment does not, yet." Open research opportunity.

Source: Bara

How did you do?
28Module 6 · Reliabilitystandard
Name three of Bara's six anti-patterns.
Answer

Any three of: (1) Passive/suggestion wording in descriptions; (2) Late-stage procedural steps without visible output; (3) Assuming output completeness implies process compliance; (4) Relying on reminder-style hooks; (5) Missing the negative-constraint component; (6) Treating activation and execution as one problem.

Source: Bara

How did you do?

Module 7 — Skills API

29Module 7 · Skills APIstandard
Name the THREE beta headers required for the full skills + files-creation workflow.
Answer

code-execution-2025-08-25 (cookbook version; API ref says 2025-05-22 — discrepancy unresolved), skills-2025-10-02, files-api-2025-04-14.

Source: Cookbook 01 + skills-guide

How did you do?
30Module 7 · Skills APIstandard
What's the MAX number of skills per container.skills array in a single API request?
Answer

8 skills max per request.

Source: skills-guide

How did you do?
31Module 7 · Skills APIstandard
What's the upload size cap for custom skill creation?
Answer

30 MB total upload size. Hard cap; includes all bundled files.

Source: skills-guide §"Creating a Skill"

How did you do?
32Module 7 · Skills APIhard ★
True or false: You can call the Skills API to retrieve the SKILL.md body of an uploaded skill. Justify.
Answer

False. Even versions/retrieve returns only metadata (id, created_at, description, directory, name, skill_id, type, version). There is NO API endpoint that returns the SKILL.md body content. Once uploaded, the file content cannot be retrieved via API. Implication: callers must keep their own local source-of-truth copy.

Source: API beta skills/versions/retrieve

How did you do?
33Module 7 · Skills APIstandard
What's the difference between the version format used by Anthropic-managed vs custom skills?
Answer

Anthropic-managed: date-based (e.g., 20251013). Custom: Unix epoch timestamp (e.g., 1759178010641129). Both support "latest" as a version string.

Source: skills-guide

How did you do?
17Module 7 · Skills APIstandard
Three pieces of metadata get extracted automatically from SKILL.md when uploaded via the Skills API. Name them.
Answer

name, description, and directory (top-level directory name from upload). Caller doesn't pass them as separate API parameters — they're extracted from the uploaded SKILL.md and folder structure.

Source: API beta skills/versions/create

How did you do?

Module 8 — Cookbooks & Anthropic-managed

34Module 8 · Cookbooksstandard
Name the four Anthropic-managed skills (the only four shipped with type: "anthropic").
Answer

pptx, xlsx, docx, pdf. All accessible via container.skills with type: "anthropic".

Source: Cookbook 01 + skills-guide

How did you do?
35Module 8 · Cookbooksstandard
Cookbook 02 makes a verbatim claim about the relationship between these skills and Claude Creates Files. What is it?
Answer

"These are the same Skills that power Claude Creates Files." Verbatim. So one underlying skills tech, three surfaces: API (developer), Claude.ai (consumer Creates Files), Cookbook (educational).

Source: Cookbook 02

How did you do?
36Module 8 · Cookbooksstandard
Cookbook 02 cites a reliability ceiling. What's the maximum number of sheets per workbook that "works reliably"?
Answer

2-3 sheets per workbook. "Works reliably and generates quickly." Beyond that, segment into multiple files using a pipeline pattern.

Source: Cookbook 02

How did you do?
37Module 8 · Cookbooksstandard
Why must you use client.beta.messages.create() instead of client.messages.create() when working with skills? What happens if you use the non-beta call?
Answer

The container parameter is not recognized by client.messages.create() — only client.beta.messages.create() supports it. If you use the non-beta call with skills, the call fails (often silently or with an opaque error). This is Issue #2 in Cookbook 01's troubleshooting.

Source: Cookbook 01 + skills-guide

How did you do?

Module 9 — Plugins & distribution

38Module 9 · Pluginsstandard
A Cowork plugin BUNDLES three things. Name them.
Answer

Skills + Connectors + Sub-agents. Verbatim from support 13837440: "Each plugin bundles together skills, connectors, and sub-agents into a single package."

Source: support.claude.com/13837440

How did you do?
39Module 9 · Pluginshard ★
Anthropic's documentation explicitly admits a major governance gap in org-wide skill provisioning. Quote the verbatim claim or summarize the gap accurately.
Answer

Verbatim: "There's no approval workflow for org-wide sharing. If you enable Share with organization, any member can publish a skill to the directory without review." Plus: audit log captures share events but NOT skill content; no admin dashboard for content inspection. Major governance gap.

Source: support.claude.com/13119606

How did you do?
40Module 9 · Pluginsstandard
On third-party platform deployments (Bedrock, Vertex AI, Azure), is the Cowork plugin marketplace available? What's the alternative distribution mechanism?
Answer

No, the Cowork plugin marketplace is NOT available on third-party platforms. Verbatim: "The skills and plugin marketplace available in Claude Enterprise isn't available with third-party platforms." Alternative: local filesystem mounts via MDM — macOS: /Library/Application Support/Claude/org-plugins/; Windows: C:\ProgramData\Claude\org-plugins\.

Source: support.claude.com/14680753

How did you do?
41Module 9 · Pluginsstandard
Name three of the five Anthropic-shipped financial-services plugins.
Answer

Any three of: Financial Analysis (core; required first install), Investment Banking, Equity Research, Private Equity, Wealth Management.

Source: support.claude.com/13851150

How did you do?

Module 10 — Claude Design

42Module 10 · Claude Designstandard
What date did Claude Design launch, and what model powers it?
Answer

April 17, 2026. Powered by Claude Opus 4.7 ("most capable vision model"). Anthropic Labs product. Pro/Max/Team/Enterprise (Enterprise default-off, admin enables).

Source: Claude Design announcement

How did you do?
43Module 10 · Claude Designhard ★
Claude Design has TWO paths to brand-aware output. Name them and identify which is available to which plan tiers.
Answer

Path 1 — DESIGN.md upload (Hassid's individual workflow): Cowork-extract → upload to Claude Design → persistent context. Available to all paid tiers. Path 2 — Built-in org-level design system integration: verbatim "Organizations configure brand colors, typography, and component patterns once; all projects automatically inherit these assets without manual uploads." Available to Team and Enterprise tiers only. Implication: DESIGN.md upload is Enact's strongest play for individual Pro/Max + small teams.

Source: Get Started article

How did you do?
44Module 10 · Claude Designstandard
Recite the verbatim Cowork extraction prompt Hassid uses to generate a DESIGN.md from a brand-asset folder.
Answer

"Analyze this folder and produce a full design system write-up. Fonts, colors, graphical styles, component patterns, tone, layout conventions. Flag anything that's missing. Save it as DESIGN.md in my folder."

Source: Hassid Substack

How did you do?
45Module 10 · Claude Designstandard
Name the 9 sections of the Stitch DESIGN.md format.
Answer

(1) Visual Theme & Atmosphere · (2) Color Palette & Roles · (3) Typography Rules · (4) Component Stylings · (5) Layout Principles · (6) Depth & Elevation · (7) Do's and Don'ts · (8) Responsive Behavior · (9) Agent Prompt Guide.

Source: VoltAgent/awesome-design-md collection / Stitch spec

How did you do?
46Module 10 · Claude Designstandard
Which section of DESIGN.md is the natural home for Seleznov's directive-template research, and why?
Answer

Section 9 — Agent Prompt Guide. It's the directive layer of DESIGN.md, analogous to SKILL.md's description field. Imperative phrasing, ALWAYS/NEVER constraints, explicit triggers for when the agent should reach for which token. Same Seleznov-style directive discipline applies.

Source: Stitch spec + synthesis

How did you do?

Module 11 — Empirical sources

47Module 11 · Empiricalstandard
Per Shcheglov's 200+ skill audit, what percentage of public skills perform BELOW baseline?
Answer

~80% of public skills perform below baseline. Specifically: 40 of 47 skills from a viral list scored below baseline; ~80% of 200+ tested.

Source: Shcheglov — "The Ultimate Guide to Claude Code Skills"

How did you do?
48Module 11 · Empiricalhard ★
State the encoded-preference-vs-capability-uplift distinction in your own words and give one concrete example of each.
Answer

Capability uplift = teaching Claude techniques it will eventually master through model improvements (e.g., "how to write a good email"). Expires with model upgrades. Encoded preference = documenting org-specific business logic that doesn't change with model improvements (e.g., "your team routes Severity 1 tickets to #critical-ops within 15 minutes"). Gains value as models improve, because the business knowledge is durable.

Source: Beehiiv aggregator (original framing surfaced via Marc Bara)

How did you do?
49Module 11 · Empiricalstandard
Per Shihipar, approximately how many skills are in active use at Anthropic itself?
Answer

"Hundreds of skills in active use at Anthropic" — Shihipar's verbatim quantifier. Not a more precise number.

Source: Shihipar (LinkedIn)

How did you do?

Module 12 — Strategic synthesis

50Module 12 · Synthesishard ★
Enact's eval moat vs Anthropic's skill-creator is NOT "we have eval, they don't." State the actual differentiation in 2-3 sentences, naming at least 4 specific dimensions where Enact's methodology or scope differs.
Answer

Enact's eval moat is methodology + scope, not "we have it, they don't." Anthropic ships skill-creator at TWO surfaces (Claude Code plugin + Claude.ai conversational meta-skill) with a real eval framework. Enact differentiates on:

  1. Source-content-driven generation (video/PDF/URL/Figma) vs skill-creator's Q&A interview
  2. Automated Refusal Gate vs skill-creator's human-in-the-loop "iterate until satisfied"
  3. Multi-model cross-testing (Haiku + Sonnet + Opus on every run) vs skill-creator's single-model-per-run
  4. Hook-robustness testing (Seleznov-derived) — not in skill-creator
  5. Execution-fidelity testing (Bara-derived; new dimension) — not in skill-creator or anywhere else
  6. Spec-drift monitoring (Sync background service) — not in skill-creator
  7. Post-install monitoring (Monitor background service) — not in skill-creator
  8. 86-item visible Checklist UX during generation — skill-creator runs eval AFTER generation

Honest framing: skill-creator is real and capable. Enact wins on the dimensions above, not by pretending Anthropic has nothing.

Source: SOURCE_NOTES.md F3 corrected · LEGAL_CLAIMS_FRAMEWORK.md

How did you do?