eval-audit-review — agent-framework eval-audit-review, soulmap-ai, community, agent-framework, ide skills, ai-agent, code-analysis, developer-tools, devtools, document-processing, Claude Code

v1.0.0

À propos de ce Skill

SoulMap AI: a content-first reflective companion with a curated Markdown knowledge base, Python detectors, and tooling to validate and bundle agent-ready skills.

# Core Topics

tuanductran tuanductran
[1]
[0]
Updated: 4/3/2026

Killer-Skills Review

Decision support comes first. Repository text comes second.

Reference-Only Page Review Score: 1/11

This page remains useful for operators, but Killer-Skills treats it as reference material instead of a primary organic landing page.

Review Score
1/11
Quality Score
29
Canonical Locale
en
Detected Body Locale
en

SoulMap AI: a content-first reflective companion with a curated Markdown knowledge base, Python detectors, and tooling to validate and bundle agent-ready skills.

Pourquoi utiliser cette compétence

SoulMap AI: a content-first reflective companion with a curated Markdown knowledge base, Python detectors, and tooling to validate and bundle agent-ready skills.

Meilleur pour

Suitable for operator workflows that need explicit guardrails before installation and execution.

Cas d'utilisation exploitables for eval-audit-review

! Sécurité et Limitations

Why this page is reference-only

  • - Current locale does not satisfy the locale-governance contract.
  • - The page lacks a strong recommendation layer.
  • - The page lacks concrete use-case guidance.
  • - The page lacks explicit limitations or caution signals.
  • - The underlying skill quality score is below the review floor.

Source Boundary

The section below is supporting source material from the upstream repository. Use the Killer-Skills review above as the primary decision layer.

Labs Demo

Browser Sandbox Environment

⚡️ Ready to unleash?

Experience this Agent in a zero-setup browser environment powered by WebContainers. No installation required.

Boot Container Sandbox

FAQ & Installation Steps

These questions and steps mirror the structured data on this page for better search understanding.

? Frequently Asked Questions

What is eval-audit-review?

SoulMap AI: a content-first reflective companion with a curated Markdown knowledge base, Python detectors, and tooling to validate and bundle agent-ready skills.

How do I install eval-audit-review?

Run the command: npx killer-skills add tuanductran/soulmap-ai/eval-audit-review. It works with Cursor, Windsurf, VS Code, Claude Code, and 19+ other IDEs.

Which IDEs are compatible with eval-audit-review?

This skill is compatible with Cursor, Windsurf, VS Code, Trae, Claude Code, OpenClaw, Aider, Codex, OpenCode, Goose, Cline, Roo Code, Kiro, Augment Code, Continue, GitHub Copilot, Sourcegraph Cody, and Amazon Q Developer. Use the Killer-Skills CLI for universal one-command installation.

How To Install

  1. 1. Open your terminal

    Open the terminal or command line in your project directory.

  2. 2. Run the install command

    Run: npx killer-skills add tuanductran/soulmap-ai/eval-audit-review. The CLI will automatically detect your IDE or AI agent and configure the skill.

  3. 3. Start using the skill

    The skill is now active. Your AI agent can use eval-audit-review immediately in the current project.

! Reference-Only Mode

This page remains useful for installation and reference, but Killer-Skills no longer treats it as a primary indexable landing page. Read the review above before relying on the upstream repository instructions.

Imported Repository Instructions

The section below is supporting source material from the upstream repository. Use the Killer-Skills review above as the primary decision layer.

Supporting Evidence

eval-audit-review

Install eval-audit-review, an AI agent skill for AI agent workflows and automation. Works with Claude Code, Cursor, and Windsurf with one-command setup.

SKILL.md
Readonly
Imported Repository Instructions
The section below is supporting source material from the upstream repository. Use the Killer-Skills review above as the primary decision layer.
Supporting Evidence

Eval audit review

Use this skill when the task is to inspect or improve the trustworthiness of SoulMap's eval system rather than only adding one more case.

Do not use this skill for

Mission

Keep evals honest, source-backed, and useful against real failure modes instead of optimizing for easy green runs.

Sources to check first

  • evals/README.md
  • evals/datasets/
  • tests/contract/
  • tests/eval_regression/
  • src/soulmap/devtools/evals/
  • the source Markdown or Python files each eval claims to protect

What to look for

  • evals that pass because assertions are too loose
  • cases with no clear source backing in AGENTS.md, skills/, or templates/
  • wording checks that drift from runtime examples
  • evaluator logic that is brittle, fuzzy, or easy to satisfy accidentally
  • important failure modes that appear in code or docs but are not represented in datasets

Workflow

  1. Identify the failure mode or product contract the eval is supposed to protect.
  2. Check whether the current dataset, harness, and source files all describe the same thing.
  3. Tighten assertions only where the behavior is actually important.
  4. Prefer a few sharp cases over many noisy ones.
  5. Add or update source_markers when confidence needs to be explicit.
  6. Run the matching eval commands, then the closest pytest contracts.

Expected output

Findings

List the eval weaknesses first, especially loose assertions, stale source links, or blind spots.

Fixes

Summarize the dataset, harness, or contract changes that improved audit quality.

Validation

State which eval and pytest commands were run.

Definition of done

The audited eval surface should be:

  • harder to game accidentally
  • clearly tied back to real source files or runtime behavior
  • focused on meaningful failure modes
  • validated with the exact commands maintainers actually use

Compétences associées

Looking for an alternative to eval-audit-review or another community skill for your workflow? Explore these related open-source skills.

Voir tout

openclaw-release-maintainer

Logo of openclaw
openclaw

Your own personal AI assistant. Any OS. Any Platform. The lobster way. 🦞

widget-generator

Logo of f
f

Générez des plugins de widgets personnalisables pour le système de flux prompts.chat

flags

Logo of vercel
vercel

Le Cadre de Réaction

138.4k
0
Navigateur

pr-review

Logo of pytorch
pytorch

Tenseurs et réseaux neuronaux dynamiques en Python avec une forte accélération GPU

98.6k
0
Développeur