flow-review — claude-code flow-review, flowai, community, claude-code, ide skills, cursor-ide, opencode

v1.0.0

이 스킬 정보

코드 리뷰 에이전트가 동시에 작업을 완료하고 코드 품질을 검증해야 하는 경우에 적합합니다. Review current changes as QA engineer and lead engineer simultaneously: verify task completion, code quality, architecture, and cleanup.

# Core Topics

korchasa korchasa
[3]
[0]
Updated: 3/21/2026

Killer-Skills Review

Decision support comes first. Repository text comes second.

Reference-Only Page Review Score: 9/11

This page remains useful for operators, but Killer-Skills treats it as reference material instead of a primary organic landing page.

Original recommendation layer Concrete use-case guidance Explicit limitations and caution Quality floor passed for review
Review Score
9/11
Quality Score
60
Canonical Locale
en
Detected Body Locale
en

코드 리뷰 에이전트가 동시에 작업을 완료하고 코드 품질을 검증해야 하는 경우에 적합합니다. Review current changes as QA engineer and lead engineer simultaneously: verify task completion, code quality, architecture, and cleanup.

이 스킬을 사용하는 이유

에이전트가 코딩 작업 중에 수행 된 변경 사항을 검토하고 작업 완료 및 코드 품질을 단일 패스로 검증 할 수있는 기능을 제공하며 작업 분기 또는暂存/暂存 변경 사항의 diff 분석을 사용하고 Claude Code 및 Cursor와 같은 AI 코딩 도우미를 지원합니다.

최적의 용도

코드 리뷰 에이전트가 동시에 작업을 완료하고 코드 품질을 검증해야 하는 경우에 적합합니다.

실행 가능한 사용 사례 for flow-review

작업 완료 및 코드 품질의 변경 사항을 검토
코드 리뷰를 위해 작업 분기의 diff를 분석
暂存/暂存 변경 사항이 원래 요청과 일치하는지 확인

! 보안 및 제한 사항

  • 코딩 작업 분기 또는暂存/暂存 변경 사항에 대한 액세스가 필요
  • Claude Code 및 Cursor와 같은 AI 코딩 도우미와 함께 사용하도록 설계되었습니다.
  • 현재 변경 사항만 검토하고 전체 프로젝트는 검토하지 않음

Why this page is reference-only

  • - Current locale does not satisfy the locale-governance contract.

Source Boundary

The section below is imported from the upstream repository and should be treated as secondary evidence. Use the Killer-Skills review above as the primary layer for fit, risk, and installation decisions.

After The Review

Decide The Next Action Before You Keep Reading Repository Material

Killer-Skills should not stop at opening repository instructions. It should help you decide whether to install this skill, when to cross-check against trusted collections, and when to move into workflow rollout.

Labs Demo

Browser Sandbox Environment

⚡️ Ready to unleash?

Experience this Agent in a zero-setup browser environment powered by WebContainers. No installation required.

Boot Container Sandbox

FAQ & Installation Steps

These questions and steps mirror the structured data on this page for better search understanding.

? Frequently Asked Questions

What is flow-review?

코드 리뷰 에이전트가 동시에 작업을 완료하고 코드 품질을 검증해야 하는 경우에 적합합니다. Review current changes as QA engineer and lead engineer simultaneously: verify task completion, code quality, architecture, and cleanup.

How do I install flow-review?

Run the command: npx killer-skills add korchasa/flowai/flow-review. It works with Cursor, Windsurf, VS Code, Claude Code, and 19+ other IDEs.

What are the use cases for flow-review?

Key use cases include: 작업 완료 및 코드 품질의 변경 사항을 검토, 코드 리뷰를 위해 작업 분기의 diff를 분석, 暂存/暂存 변경 사항이 원래 요청과 일치하는지 확인.

Which IDEs are compatible with flow-review?

This skill is compatible with Cursor, Windsurf, VS Code, Trae, Claude Code, OpenClaw, Aider, Codex, OpenCode, Goose, Cline, Roo Code, Kiro, Augment Code, Continue, GitHub Copilot, Sourcegraph Cody, and Amazon Q Developer. Use the Killer-Skills CLI for universal one-command installation.

Are there any limitations for flow-review?

코딩 작업 분기 또는暂存/暂存 변경 사항에 대한 액세스가 필요. Claude Code 및 Cursor와 같은 AI 코딩 도우미와 함께 사용하도록 설계되었습니다.. 현재 변경 사항만 검토하고 전체 프로젝트는 검토하지 않음.

How To Install

  1. 1. Open your terminal

    Open the terminal or command line in your project directory.

  2. 2. Run the install command

    Run: npx killer-skills add korchasa/flowai/flow-review. The CLI will automatically detect your IDE or AI agent and configure the skill.

  3. 3. Start using the skill

    The skill is now active. Your AI agent can use flow-review immediately in the current project.

! Reference-Only Mode

This page remains useful for installation and reference, but Killer-Skills no longer treats it as a primary indexable landing page. Read the review above before relying on the upstream repository instructions.

Upstream Repository Material

The section below is imported from the upstream repository and should be treated as secondary evidence. Use the Killer-Skills review above as the primary layer for fit, risk, and installation decisions.

Upstream Source

flow-review

Install flow-review, an AI agent skill for AI agent workflows and automation. Review the use cases, limitations, and setup path before rollout.

SKILL.md
Readonly
Upstream Repository Material
The section below is imported from the upstream repository and should be treated as secondary evidence. Use the Killer-Skills review above as the primary layer for fit, risk, and installation decisions.
Supporting Evidence

Task: Review Changes

Overview

Act as QA engineer + lead engineer simultaneously. Review only the current changes (diff from the task branch or unstaged/staged changes) against the original request and plan. Verify task completion AND code quality in a single pass.

Context

<context> The user has completed (or nearly completed) a coding task and needs a combined review before committing. You review ONLY the changes made during this task, NOT the entire project. Your two hats: 1. **QA**: Did the changes satisfy the request? Is anything missing, broken, or left in a dirty state? 2. **Lead Engineer**: Are the changes well-designed, readable, safe, and consistent with the project's conventions?

Input sources:

  • Git diff (git diff, git diff --cached, git diff <base>..HEAD).
  • The original User Request (from chat history).
  • The Plan (task management tool or documents/whiteboard.md).
  • Project conventions (AGENTS.md, linter/formatter configs). </context>

Rules & Constraints

<rules> 1. **Scope**: Review ONLY changed/added files. Do NOT audit the whole project (that is `flow-maintenance`'s job). 2. **Diff-first**: Start from `git diff`. Every finding must reference a specific file and line in the diff. 3. **Two roles, one pass**: Produce findings under two categories (QA, Code Review) but run them in parallel, not sequentially. 4. **Verification**: Do not assume it works — read files, run project checks (linter, tests, type-checker) if available. 5. **Mandatory**: Use a task management tool (e.g., `todo_write`, `todowrite`) to track the execution steps of this review. 6. **Severity levels**: Tag each finding as `[critical]`, `[warning]`, or `[nit]`. Critical = blocks merge. Warning = should fix. Nit = optional improvement. 7. **Output**: Final verdict is **Approve**, **Request Changes**, or **Needs Discussion** with actionable items. </rules>

Instructions

<step_by_step>

  1. Empty Diff Guard

    • Run git diff --stat, git diff --cached --stat, and git status --short.
    • If there are NO changes (no diff, no staged files, no untracked files), report "No changes to review" and STOP.
  2. Gather Context

    • Create a review plan in the task management tool.
    • Collect the diff: git diff (unstaged), git diff --cached (staged), or git log --oneline <base>..HEAD + git diff <base>..HEAD for branch-based changes.
    • Read the original user request and the plan (whiteboard / task list).
    • Look for project conventions in AGENTS.md and config files. If these files do not exist, rely on conventions visible in the diff and surrounding code.
  3. QA: Task Completion

    • Map each requirement/plan item to concrete changes in the diff.
    • Flag requirements with no corresponding changes as [critical] Missing.
    • Flag plan items marked "done" but not present in diff as [critical] Phantom completion.
    • Check for regressions: do changed files break existing functionality?
  4. QA: Hygiene

    • Temp artifacts: New temp_*, *.tmp, *.bak, debug console.log/ print statements, hardcoded secrets or localhost URLs.
    • Unfinished markers: New TODO, FIXME, HACK, XXX introduced in this diff (distinguish from pre-existing ones).
    • Dead code: Commented-out blocks, unused imports/variables/functions added in this diff.
  5. Code Review: Design & Architecture

    • Responsibility: Does each changed file/module stay within its stated responsibility? Flag scope creep.
    • Coupling: Are new dependencies (imports, API calls) justified? Flag tight coupling or circular dependencies.
    • Abstraction: Is the level of abstraction appropriate? Flag over-engineering (unnecessary interfaces, premature generalization) and under-engineering (god-functions, duplicated logic).
  6. Code Review: Implementation Quality

    • Naming: Are new identifiers (vars, funcs, types) clear and consistent with project conventions?
    • Error handling: Are errors handled explicitly? Flag swallowed exceptions, missing error paths, generic catch-all handlers.
    • Edge cases: Are boundary conditions (null, empty, overflow, concurrent access) handled?
    • Types & contracts: Are type signatures precise? Flag any, untyped parameters, missing return types (where project conventions require them).
    • Tests: Do new/changed behaviors have corresponding tests? Are existing tests updated for changed behavior?
  7. Code Review: Readability & Style

    • Consistency: Do changes follow the project's established patterns (file structure, naming, formatting)?
    • Comments: Are non-obvious decisions explained? Flag misleading or stale comments.
    • Complexity: Flag functions > 40 lines or cyclomatic complexity spikes introduced in this diff.
    • Clarity: Flag clarity sacrificed for brevity — nested ternaries, dense one-liners, overly compact expressions. Explicit code is preferred over clever short forms.
  8. Run Automated Checks

    • If the project has a check command (deno task check, npm run lint, make check, etc.), run it and include results.
    • If no check command is found, explicitly note "No automated checks configured" in the report — do not silently skip.
    • If tests exist, run them and report failures.
  9. Final Report Output a structured report with the verdict on the FIRST line:

    ## Review: [Approve | Request Changes | Needs Discussion]
    
    ### QA Findings
    - [severity] file:line — description
    
    ### Code Review Findings
    - [severity] file:line — description
    
    ### Automated Checks
    - [pass|fail|skipped] command — summary
    
    ### Summary
    - Requirements covered: X/Y
    - Critical issues: N
    - Warnings: N
    - Nits: N
    

    If no issues: short confirmation "Changes look good. All requirements covered, no issues found."

</step_by_step>

Verification

<verification> [ ] Empty diff guard checked before starting. [ ] Diff collected and reviewed (not the whole project). [ ] Each requirement/plan item mapped to changes. [ ] Hygiene check: no temp files, debug output, unfinished markers in diff. [ ] Design review: responsibility, coupling, abstraction checked. [ ] Implementation review: naming, errors, edge cases, types, tests checked. [ ] Readability: consistency, comments, complexity checked. [ ] Automated checks executed (or explicitly noted as missing). [ ] Structured report produced with severity-tagged findings. [ ] Verdict on the first line of the report. </verification>

관련 스킬

Looking for an alternative to flow-review or another community skill for your workflow? Explore these related open-source skills.

모두 보기

openclaw-release-maintainer

Logo of openclaw
openclaw

Your own personal AI assistant. Any OS. Any Platform. The lobster way. 🦞

333.8k
0
인공지능

widget-generator

Logo of f
f

prompts.chat 피드 시스템을 위한 사용자 지정 가능한 위젯 플러그인을 생성합니다

149.6k
0
인공지능

flags

Logo of vercel
vercel

리액트 프레임워크

138.4k
0
브라우저

pr-review

Logo of pytorch
pytorch

파이썬에서 텐서와 동적 신경망 구현 및 강력한 GPU 가속 지원

98.6k
0
개발자