Game Dev Teams

team-qa

Donchitos/Claude-Code-Game-Studios · updated Apr 16, 2026

$npx skills add https://github.com/Donchitos/Claude-Code-Game-Studios --skill team-qa
summary

### Team Qa

  • description: "Orchestrate the QA team through a full testing cycle. Coordinates qa-lead (strategy + test plan) and qa-tester (test case writing + bug reporting) to produce a complete QA package for a
  • argument-hint: "[sprint | feature: system-name]"
  • allowed-tools: Read, Glob, Grep, Write, Task, AskUserQuestion
skill.md

When this skill is invoked, orchestrate the QA team through a structured testing cycle.

Decision Points: At each phase transition, use AskUserQuestion to present the user with the subagent's proposals as selectable options. Write the agent's full analysis in conversation, then capture the decision with concise labels. The user must approve before moving to the next phase.

Team Composition

  • qa-lead — QA strategy, test plan generation, story classification, sign-off report
  • qa-tester — Test case writing, bug report writing, manual QA documentation

How to Delegate

Use the Task tool to spawn each team member as a subagent:

  • subagent_type: qa-lead — Strategy, planning, classification, sign-off
  • subagent_type: qa-tester — Test case writing and bug report writing

Always provide full context in each agent's prompt (story file paths, QA plan path, scope constraints). Launch independent qa-tester tasks in parallel where possible (e.g., multiple stories in Phase 5 can be scaffolded simultaneously).

Pipeline

Phase 1: Load Context

Before doing anything else, gather the full scope:

  1. Detect the current sprint or feature scope from the argument:

    • If argument is a sprint identifier (e.g., sprint-03): read all story files in production/sprints/[sprint]/
    • If argument is feature: [system-name]: glob story files tagged for that system
    • If no argument: read production/session-state/active.md and production/sprint-status.yaml (if present) to infer the active sprint
  2. Read production/stage.txt to confirm the current project phase.

  3. Count stories found and report to the user:

    "QA cycle starting for [sprint/feature]. Found [N] stories. Current stage: [stage]. Ready to begin QA strategy?"

Phase 2: QA Strategy (qa-lead)

Spawn qa-lead via Task to review all in-scope stories and produce a QA strategy.

Prompt the qa-lead to:

  • Read each story file

  • Classify each story by type: Logic / Integration / Visual/Feel / UI / Config/Data

  • Identify which stories require automated test evidence vs. manual QA

  • Flag any stories with missing acceptance criteria or missing test evidence that would block QA

  • Estimate manual QA effort (number of test sessions needed)

  • Check tests/smoke/ for smoke test scenarios; for each, assess whether it can be verified given the current build. Produce a smoke check verdict: PASS / PASS WITH WARNINGS [list] / FAIL [list of failures]

  • Produce a strategy summary table and smoke check result:

    StoryTypeAutomated RequiredManual RequiredBlocker?

    Smoke Check: [PASS / PASS WITH WARNINGS / FAIL] — [details if not PASS]

If the smoke check result is FAIL, the qa-lead must list the failures prominently. QA cannot proceed past the strategy phase with a failed smoke check.

Present the qa-lead's full strategy to the user, then use AskUserQuestion:

question: "QA Strategy Review"
options:
  - "Looks good — proceed to test plan"
  - "Adjust story types before proceeding"
  - "Skip blocked stories and proceed with the rest"
  - "Smoke check failed — fix issues and re-run /team-qa"
  - "Cancel — resolve blockers first"

If smoke check FAIL: do not proceed to Phase 3. Surface the failures and stop. The user must fix them and re-run /team-qa. If smoke check PASS WITH WARNINGS: note the warnings for the sign-off report and continue. If blockers are present: list them explicitly. The user may choose to skip blocked stories or cancel the cycle.

Phase 3: Test Plan Generation

Using the strategy from Phase 2, produce a structured test plan document.

The test plan should cover:

  • Scope: sprint/feature name, story count, dates
  • Story Classification Table: from Phase 2 strategy
  • Automated Test Requirements: which stories need test files, expected paths in tests/
  • Manual QA Scope: which stories need manual walkthrough and what to validate
  • Out of Scope: what is explicitly not being tested this cycle and why
  • Entry Criteria: what must be true before QA can begin (smoke check pass, build stable)
  • Exit Criteria: what constitutes a completed QA cycle (all stories PASS or FAIL with bugs filed)

Ask: "May I write the QA plan to production/qa/qa-plan-[sprint]-[date].md?"

Write only after receiving approval.

Phase 4: Test Case Writing (qa-tester)

Smoke check is performed as part of Phase 2 (QA Strategy). If the smoke check returned FAIL in Phase 2, the cycle was stopped there. This phase only runs when the Phase 2 smoke check was PASS or PASS WITH WARNINGS.

For each story requiring manual QA (Visual/Feel, UI, Integration without automated tests):

Spawn qa-tester via Task for each story (run in parallel where possible), providing:

  • The story file path
  • The relevant section of the QA plan for that story
  • The GDD acceptance criteria for the system being tested (if available)
  • Instructions to write detailed test cases covering all acceptance criteria

Each test case set should include:

  • Preconditions: game state required before testing begins
  • Steps: numbered, unambiguous actions
  • Expected Result: what should happen
  • Actual Result: field left blank for the tester to fill in
  • Pass/Fail: field left blank

Present the test cases to the user for review before execution. Group by story.

Use AskUserQuestion per story group (batched 3-4 at a time):

question: "Test cases ready for [Story Group]. Review before manual QA begins?"
options:
  - "Approved — begin manual QA for these stories"
  - "Revise test cases for [story name]"
  - "Skip manual QA for [story name] — not ready"

Phase 6: Manual QA Execution

Walk through each story in the approved manual QA list.

Batch stories into groups of 3-4 and use AskUserQuestion for each:

question: "Manual QA — [Story Title]\n[brief description of what to test]"
options:
  - "PASS — all acceptance criteria verified"
  - "PASS WITH NOTES — minor issues found (describe after)"
  - "FAIL — criteria not met (describe after)"
  - "BLOCKED — cannot test yet (reason)"

After each FAIL result: use AskUserQuestion to collect the failure description, then spawn qa-tester via Task to write a formal bug report in production/qa/bugs/.

Bug report naming: BUG-[NNN]-[short-slug].md (increment NNN from existing bugs in the directory).

After collecting all results, summarize:

  • Stories PASS: [count]
  • Stories PASS WITH NOTES: [count]
  • Stories FAIL: [count] — bugs filed: [IDs]
  • Stories BLOCKED: [count]

Phase 7: QA Sign-Off Report

Spawn qa-lead via Task to produce the sign-off report using all results from Phases 4–6.

The sign-off report format:

## QA Sign-Off Report: [Sprint/Feature]
**Date**: [date]
**QA Lead sign-off**: [pending]

### Test Coverage Summary
| Story | Type | Auto Test | Manual QA | Result |
|-------|------|-----------|-----------|--------|
| [title] | Logic | PASS | — | PASS |
| [title] | Visual | — | PASS | PASS |

### Bugs Found
| ID | Story | Severity | Status |
|----|-------|----------|--------|
| BUG-001 | [story] | S2 | Open |

### Verdict: APPROVED / APPROVED WITH CONDITIONS / NOT APPROVED

**Conditions** (if any): [list what must be fixed before the build advances]

### Next Step
[guidance based on verdict]

Verdict rules:

  • APPROVED: All stories PASS or PASS WITH NOTES; no S1/S2 bugs open
  • APPROVED WITH CONDITIONS: S3/S4 bugs open, or PASS WITH NOTES issues documented; no S1/S2 bugs
  • NOT APPROVED: Any S1/S2 bugs open; or stories FAIL without documented workaround

Next step guidance by verdict:

  • APPROVED: "Build is ready for the next phase. Run /gate-check to validate advancement."
  • APPROVED WITH CONDITIONS: "Resolve conditions before advancing. S3/S4 bugs may be deferred to polish."
  • NOT APPROVED: "Resolve S1/S2 bugs and re-run /team-qa or targeted manual QA before advancing."

Ask: "May I write this QA sign-off report to production/qa/qa-signoff-[sprint]-[date].md?"

Write only after receiving approval.

Error Recovery Protocol

If any spawned agent (via Task) returns BLOCKED, errors, or cannot complete:

  1. Surface immediately: Report "[AgentName]: BLOCKED — [reason]" to the user before continuing to dependent phases
  2. Assess dependencies: Check whether the blocked agent's output is required by subsequent phases. If yes, do not proceed past that dependency point without user input.
  3. Offer options via AskUserQuestion with choices:
    • Skip this agent and note the gap in the final report
    • Retry with narrower scope
    • Stop here and resolve the blocker first
  4. Always produce a partial report — output whatever was completed. Never discard work because one agent blocked.

Common blockers:

  • Input file missing (story not found, GDD absent) → redirect to the skill that creates it
  • ADR status is Proposed → do not implement; run /architecture-decision first
  • Scope too large → split into two stories via /create-stories
  • Conflicting instructions between ADR and story → surface the conflict, do not guess

Output

A summary covering: stories in scope, smoke check result, manual QA results, bugs filed (with IDs and severities), and the final APPROVED / APPROVED WITH CONDITIONS / NOT APPROVED verdict.

Verdict: COMPLETE — QA cycle finished. Verdict: BLOCKED — smoke check failed or critical blocker prevented cycle completion; partial report produced.

general reviews

Ratings

4.852 reviews
  • Chaitanya Patil· Dec 28, 2024

    team-qa fits our agent workflows well — practical, well scoped, and easy to wire into existing repos.

  • Jin Robinson· Dec 24, 2024

    We added team-qa from the explainx registry; install was straightforward and the SKILL.md answered most questions upfront.

  • Ren Wang· Dec 24, 2024

    Useful defaults in team-qa — fewer surprises than typical one-off scripts, and it plays nicely with `npx skills` flows.

  • Diego Rao· Dec 12, 2024

    Solid pick for teams standardizing on skills: team-qa is focused, and the summary matches what you get after install.

  • Charlotte Johnson· Dec 4, 2024

    I recommend team-qa for anyone iterating fast on agent tooling; clear intent and a small, reviewable surface area.

  • Piyush G· Nov 19, 2024

    Registry listing for team-qa matched our evaluation — installs cleanly and behaves as described in the markdown.

  • Jin White· Nov 15, 2024

    team-qa reduced setup friction for our internal harness; good balance of opinion and flexibility.

  • Diego Gill· Nov 3, 2024

    team-qa is among the better-maintained entries we tried; worth keeping pinned for repeat workflows.

  • Valentina Srinivasan· Oct 22, 2024

    Keeps context tight: team-qa is the kind of skill you can hand to a new teammate without a long onboarding doc.

  • Shikha Mishra· Oct 10, 2024

    team-qa reduced setup friction for our internal harness; good balance of opinion and flexibility.

showing 1-10 of 52

1 / 6