flow-review — claude-code flow-review, flowai, community, claude-code, ide skills, cursor-ide, opencode, Claude Code, Cursor, Windsurf

v1.0.0
GitHub

About this Skill

Perfect for Code Review Agents needing simultaneous task completion and code quality verification capabilities. Review current changes as QA engineer and lead engineer simultaneously: verify task completion, code quality, architecture, and cleanup.

# Core Topics

korchasa korchasa
[3]
[0]
Updated: 3/21/2026

Agent Capability Analysis

The flow-review skill by korchasa is an open-source community AI agent skill for Claude Code and other IDE workflows, helping agents execute tasks with better context, repeatability, and domain-specific guidance. Optimized for claude-code, cursor-ide, opencode.

Ideal Agent Persona

Perfect for Code Review Agents needing simultaneous task completion and code quality verification capabilities.

Core Value

Empowers agents to review changes made during coding tasks, verifying task completion and code quality in a single pass, utilizing diff analysis from task branches or unstaged/staged changes, and supporting AI coding assistants like Claude Code and Cursor.

Capabilities Granted for flow-review

Reviewing changes for task completion and code quality
Analyzing diff from task branches for code review
Verifying unstaged/staged changes against original requests

! Prerequisites & Limits

  • Requires access to coding task branches or unstaged/staged changes
  • Designed for use with AI coding assistants like Claude Code and Cursor
  • Only reviews current changes, not the entire project
Labs Demo

Browser Sandbox Environment

⚡️ Ready to unleash?

Experience this Agent in a zero-setup browser environment powered by WebContainers. No installation required.

Boot Container Sandbox
SKILL.md
Readonly

Task: Review Changes

Overview

Act as QA engineer + lead engineer simultaneously. Review only the current changes (diff from the task branch or unstaged/staged changes) against the original request and plan. Verify task completion AND code quality in a single pass.

Context

<context> The user has completed (or nearly completed) a coding task and needs a combined review before committing. You review ONLY the changes made during this task, NOT the entire project. Your two hats: 1. **QA**: Did the changes satisfy the request? Is anything missing, broken, or left in a dirty state? 2. **Lead Engineer**: Are the changes well-designed, readable, safe, and consistent with the project's conventions?

Input sources:

  • Git diff (git diff, git diff --cached, git diff <base>..HEAD).
  • The original User Request (from chat history).
  • The Plan (task management tool or documents/whiteboard.md).
  • Project conventions (AGENTS.md, linter/formatter configs). </context>

Rules & Constraints

<rules> 1. **Scope**: Review ONLY changed/added files. Do NOT audit the whole project (that is `flow-maintenance`'s job). 2. **Diff-first**: Start from `git diff`. Every finding must reference a specific file and line in the diff. 3. **Two roles, one pass**: Produce findings under two categories (QA, Code Review) but run them in parallel, not sequentially. 4. **Verification**: Do not assume it works — read files, run project checks (linter, tests, type-checker) if available. 5. **Mandatory**: Use a task management tool (e.g., `todo_write`, `todowrite`) to track the execution steps of this review. 6. **Severity levels**: Tag each finding as `[critical]`, `[warning]`, or `[nit]`. Critical = blocks merge. Warning = should fix. Nit = optional improvement. 7. **Output**: Final verdict is **Approve**, **Request Changes**, or **Needs Discussion** with actionable items. </rules>

Instructions

<step_by_step>

  1. Empty Diff Guard

    • Run git diff --stat, git diff --cached --stat, and git status --short.
    • If there are NO changes (no diff, no staged files, no untracked files), report "No changes to review" and STOP.
  2. Gather Context

    • Create a review plan in the task management tool.
    • Collect the diff: git diff (unstaged), git diff --cached (staged), or git log --oneline <base>..HEAD + git diff <base>..HEAD for branch-based changes.
    • Read the original user request and the plan (whiteboard / task list).
    • Look for project conventions in AGENTS.md and config files. If these files do not exist, rely on conventions visible in the diff and surrounding code.
  3. QA: Task Completion

    • Map each requirement/plan item to concrete changes in the diff.
    • Flag requirements with no corresponding changes as [critical] Missing.
    • Flag plan items marked "done" but not present in diff as [critical] Phantom completion.
    • Check for regressions: do changed files break existing functionality?
  4. QA: Hygiene

    • Temp artifacts: New temp_*, *.tmp, *.bak, debug console.log/ print statements, hardcoded secrets or localhost URLs.
    • Unfinished markers: New TODO, FIXME, HACK, XXX introduced in this diff (distinguish from pre-existing ones).
    • Dead code: Commented-out blocks, unused imports/variables/functions added in this diff.
  5. Code Review: Design & Architecture

    • Responsibility: Does each changed file/module stay within its stated responsibility? Flag scope creep.
    • Coupling: Are new dependencies (imports, API calls) justified? Flag tight coupling or circular dependencies.
    • Abstraction: Is the level of abstraction appropriate? Flag over-engineering (unnecessary interfaces, premature generalization) and under-engineering (god-functions, duplicated logic).
  6. Code Review: Implementation Quality

    • Naming: Are new identifiers (vars, funcs, types) clear and consistent with project conventions?
    • Error handling: Are errors handled explicitly? Flag swallowed exceptions, missing error paths, generic catch-all handlers.
    • Edge cases: Are boundary conditions (null, empty, overflow, concurrent access) handled?
    • Types & contracts: Are type signatures precise? Flag any, untyped parameters, missing return types (where project conventions require them).
    • Tests: Do new/changed behaviors have corresponding tests? Are existing tests updated for changed behavior?
  7. Code Review: Readability & Style

    • Consistency: Do changes follow the project's established patterns (file structure, naming, formatting)?
    • Comments: Are non-obvious decisions explained? Flag misleading or stale comments.
    • Complexity: Flag functions > 40 lines or cyclomatic complexity spikes introduced in this diff.
    • Clarity: Flag clarity sacrificed for brevity — nested ternaries, dense one-liners, overly compact expressions. Explicit code is preferred over clever short forms.
  8. Run Automated Checks

    • If the project has a check command (deno task check, npm run lint, make check, etc.), run it and include results.
    • If no check command is found, explicitly note "No automated checks configured" in the report — do not silently skip.
    • If tests exist, run them and report failures.
  9. Final Report Output a structured report with the verdict on the FIRST line:

    ## Review: [Approve | Request Changes | Needs Discussion]
    
    ### QA Findings
    - [severity] file:line — description
    
    ### Code Review Findings
    - [severity] file:line — description
    
    ### Automated Checks
    - [pass|fail|skipped] command — summary
    
    ### Summary
    - Requirements covered: X/Y
    - Critical issues: N
    - Warnings: N
    - Nits: N
    

    If no issues: short confirmation "Changes look good. All requirements covered, no issues found."

</step_by_step>

Verification

<verification> [ ] Empty diff guard checked before starting. [ ] Diff collected and reviewed (not the whole project). [ ] Each requirement/plan item mapped to changes. [ ] Hygiene check: no temp files, debug output, unfinished markers in diff. [ ] Design review: responsibility, coupling, abstraction checked. [ ] Implementation review: naming, errors, edge cases, types, tests checked. [ ] Readability: consistency, comments, complexity checked. [ ] Automated checks executed (or explicitly noted as missing). [ ] Structured report produced with severity-tagged findings. [ ] Verdict on the first line of the report. </verification>

FAQ & Installation Steps

These questions and steps mirror the structured data on this page for better search understanding.

? Frequently Asked Questions

What is flow-review?

Perfect for Code Review Agents needing simultaneous task completion and code quality verification capabilities. Review current changes as QA engineer and lead engineer simultaneously: verify task completion, code quality, architecture, and cleanup.

How do I install flow-review?

Run the command: npx killer-skills add korchasa/flowai/flow-review. It works with Cursor, Windsurf, VS Code, Claude Code, and 19+ other IDEs.

What are the use cases for flow-review?

Key use cases include: Reviewing changes for task completion and code quality, Analyzing diff from task branches for code review, Verifying unstaged/staged changes against original requests.

Which IDEs are compatible with flow-review?

This skill is compatible with Cursor, Windsurf, VS Code, Trae, Claude Code, OpenClaw, Aider, Codex, OpenCode, Goose, Cline, Roo Code, Kiro, Augment Code, Continue, GitHub Copilot, Sourcegraph Cody, and Amazon Q Developer. Use the Killer-Skills CLI for universal one-command installation.

Are there any limitations for flow-review?

Requires access to coding task branches or unstaged/staged changes. Designed for use with AI coding assistants like Claude Code and Cursor. Only reviews current changes, not the entire project.

How To Install

  1. 1. Open your terminal

    Open the terminal or command line in your project directory.

  2. 2. Run the install command

    Run: npx killer-skills add korchasa/flowai/flow-review. The CLI will automatically detect your IDE or AI agent and configure the skill.

  3. 3. Start using the skill

    The skill is now active. Your AI agent can use flow-review immediately in the current project.

Related Skills

Looking for an alternative to flow-review or another community skill for your workflow? Explore these related open-source skills.

View All

openclaw-release-maintainer

Logo of openclaw
openclaw

Your own personal AI assistant. Any OS. Any Platform. The lobster way. 🦞

333.8k
0
AI

widget-generator

Logo of f
f

Generate customizable widget plugins for the prompts.chat feed system

149.6k
0
AI

flags

Logo of vercel
vercel

The React Framework

138.4k
0
Browser

pr-review

Logo of pytorch
pytorch

Tensors and Dynamic neural networks in Python with strong GPU acceleration

98.6k
0
Developer