add-model — for Claude Code add-model, convex-evals, community, for Claude Code, ide skills, Claude 4.6 Opus, claude-opus-4.5, claude-opus-4.6, apiKind, "responses"

v1.0.0

Acerca de este Skill

Escenario recomendado: Ideal for AI agents that need add a new model to the eval runner. Resumen localizado: # Add a New Model to the Eval Runner Follow these steps whenever the user asks to add a new AI model to the eval suite. This AI agent skill supports Claude Code, Cursor, and Windsurf workflows.

Características

Add a New Model to the Eval Runner
Follow these steps whenever the user asks to add a new AI model to the eval suite.
Step 0: Gather Information
Determine the following (ask the user if not provided):
Formatted name - human-readable, e.g. Claude 4.6 Opus.

# Core Topics

get-convex get-convex
[109]
[9]
Updated: 4/17/2026

Killer-Skills Review

Decision support comes first. Repository text comes second.

Reference-Only Page Review Score: 10/11

This page remains useful for operators, but Killer-Skills treats it as reference material instead of a primary organic landing page.

Original recommendation layer Concrete use-case guidance Explicit limitations and caution Quality floor passed for review
Review Score
10/11
Quality Score
59
Canonical Locale
en
Detected Body Locale
en

Escenario recomendado: Ideal for AI agents that need add a new model to the eval runner. Resumen localizado: # Add a New Model to the Eval Runner Follow these steps whenever the user asks to add a new AI model to the eval suite. This AI agent skill supports Claude Code, Cursor, and Windsurf workflows.

¿Por qué usar esta habilidad?

Recomendacion: add-model helps agents add a new model to the eval runner. Add a New Model to the Eval Runner Follow these steps whenever the user asks to add a new AI model to the eval suite. This AI agent skill

Mejor para

Escenario recomendado: Ideal for AI agents that need add a new model to the eval runner.

Casos de uso accionables for add-model

Caso de uso: Applying Add a New Model to the Eval Runner
Caso de uso: Applying Follow these steps whenever the user asks to add a new AI model to the eval suite
Caso de uso: Applying Step 0: Gather Information

! Seguridad y limitaciones

  • Limitacion: apiKind - only needed for OpenAI Codex/Responses-API models; set to "responses". Omit for all other models.
  • Limitacion: // apiKind: "responses", // only for OpenAI Codex / Responses-API models
  • Limitacion: Only proceed to the next step once at least one eval completes successfully.

Why this page is reference-only

  • - Current locale does not satisfy the locale-governance contract.

Source Boundary

The section below is imported from the upstream repository and should be treated as secondary evidence. Use the Killer-Skills review above as the primary layer for fit, risk, and installation decisions.

After The Review

Decide The Next Action Before You Keep Reading Repository Material

Killer-Skills should not stop at opening repository instructions. It should help you decide whether to install this skill, when to cross-check against trusted collections, and when to move into workflow rollout.

Labs Demo

Browser Sandbox Environment

⚡️ Ready to unleash?

Experience this Agent in a zero-setup browser environment powered by WebContainers. No installation required.

Boot Container Sandbox

FAQ & Installation Steps

These questions and steps mirror the structured data on this page for better search understanding.

? Frequently Asked Questions

What is add-model?

Escenario recomendado: Ideal for AI agents that need add a new model to the eval runner. Resumen localizado: # Add a New Model to the Eval Runner Follow these steps whenever the user asks to add a new AI model to the eval suite. This AI agent skill supports Claude Code, Cursor, and Windsurf workflows.

How do I install add-model?

Run the command: npx killer-skills add get-convex/convex-evals/add-model. It works with Cursor, Windsurf, VS Code, Claude Code, and 19+ other IDEs.

What are the use cases for add-model?

Key use cases include: Caso de uso: Applying Add a New Model to the Eval Runner, Caso de uso: Applying Follow these steps whenever the user asks to add a new AI model to the eval suite, Caso de uso: Applying Step 0: Gather Information.

Which IDEs are compatible with add-model?

This skill is compatible with Cursor, Windsurf, VS Code, Trae, Claude Code, OpenClaw, Aider, Codex, OpenCode, Goose, Cline, Roo Code, Kiro, Augment Code, Continue, GitHub Copilot, Sourcegraph Cody, and Amazon Q Developer. Use the Killer-Skills CLI for universal one-command installation.

Are there any limitations for add-model?

Limitacion: apiKind - only needed for OpenAI Codex/Responses-API models; set to "responses". Omit for all other models.. Limitacion: // apiKind: "responses", // only for OpenAI Codex / Responses-API models. Limitacion: Only proceed to the next step once at least one eval completes successfully..

How To Install

  1. 1. Open your terminal

    Open the terminal or command line in your project directory.

  2. 2. Run the install command

    Run: npx killer-skills add get-convex/convex-evals/add-model. The CLI will automatically detect your IDE or AI agent and configure the skill.

  3. 3. Start using the skill

    The skill is now active. Your AI agent can use add-model immediately in the current project.

! Reference-Only Mode

This page remains useful for installation and reference, but Killer-Skills no longer treats it as a primary indexable landing page. Read the review above before relying on the upstream repository instructions.

Upstream Repository Material

The section below is imported from the upstream repository and should be treated as secondary evidence. Use the Killer-Skills review above as the primary layer for fit, risk, and installation decisions.

Upstream Source

add-model

# Add a New Model to the Eval Runner Follow these steps whenever the user asks to add a new AI model to the eval suite. This AI agent skill supports Claude

SKILL.md
Readonly
Upstream Repository Material
The section below is imported from the upstream repository and should be treated as secondary evidence. Use the Killer-Skills review above as the primary layer for fit, risk, and installation decisions.
Supporting Evidence

Add a New Model to the Eval Runner

Follow these steps whenever the user asks to add a new AI model to the eval suite.

Step 0: Gather Information

Determine the following (ask the user if not provided):

  1. Model identifier - the OpenRouter-style name, e.g. anthropic/claude-opus-4.6. If the user gives a marketing name or URL, look up the OpenRouter model id.
  2. Formatted name - human-readable, e.g. Claude 4.6 Opus.
  3. Provider family & version - needed to find older siblings (e.g. claude-opus-4.5 is the predecessor of claude-opus-4.6).
  4. apiKind - only needed for OpenAI Codex/Responses-API models; set to "responses". Omit for all other models.

If you're unsure, check how the closest existing model in the same family is configured in runner/models/index.ts and match it.

Step 1: Add the Model to runner/models/index.ts

Open runner/models/index.ts and add a new entry to the ALL_MODELS array. Place it next to its family siblings, respecting the existing grouping comments.

Template:

typescript
1{ 2 name: "<provider>/<model-id>", 3 formattedName: "<Human Name>", 4 // apiKind: "responses", // only for OpenAI Codex / Responses-API models 5},

Step 2: Update the Manual Evals Workflow

Open .github/workflows/manual_evals.yml and replace the entire matrix.model list with only the new model. This workflow exists solely to collect baseline data for newly added models, so it should only ever contain the latest addition.

yaml
1matrix: 2 model: 3 - "<provider>/<model-id>"

Step 3: Typecheck

Run bun run typecheck to verify no type errors were introduced.

Step 4: Smoke Test the New Model Locally

Before committing, run a quick local sanity check with one or two simple evals to confirm the model ID is valid, the API key works, and results are being produced. Use the simplest fundamentals evals:

bash
1MODELS=<new-model-name> TEST_FILTER="000-fundamentals/000" bun run local:run

If that passes, optionally run one more:

bash
1MODELS=<new-model-name> TEST_FILTER="000-fundamentals/001" bun run local:run

What to look for:

  • No authentication or "model not found" errors
  • The run completes and produces a score (even a low score is fine - we just want to confirm it runs)
  • If it fails with an API/auth error, stop and fix the model ID or check the .env file before proceeding

Only proceed to the next step once at least one eval completes successfully.

Step 5: Commit and Push

Create a descriptive commit message and push to main:

git add runner/models/index.ts .github/workflows/manual_evals.yml
git commit -m "add <model-name>; demote older <family> versions"
git push origin main

Step 6: Trigger Manual Eval Runs for Baseline Data

Use the GitHub CLI to dispatch the manual eval workflow 3 times (to get a statistically meaningful baseline):

bash
1gh workflow run manual_evals.yml --ref main

Run this command 3 times, waiting ~5 seconds between dispatches to avoid collisions.

Step 7: Monitor the Runs Until Completion

You MUST poll until all 3 runs reach a terminal state (completed/failed/cancelled). Do not stop monitoring early or hand back to the user while runs are still in progress.

Poll every ~2 minutes using:

bash
1gh run list --workflow=manual_evals.yml --limit=6

Runs typically take 20-30 minutes. Keep checking until all show completed. If a run fails, immediately investigate:

bash
1gh run view <run-id> --log-failed

Report the final pass/fail status for each run to the user once all 3 are done.

Summary Checklist

  • Model added to ALL_MODELS in runner/models/index.ts
  • .github/workflows/manual_evals.yml matrix replaced with only the new model
  • bun run typecheck passes
  • Smoke test: at least one eval completes successfully locally
  • Changes committed and pushed to main
  • Manual eval workflow dispatched 3 times
  • All 3 runs monitored to completion

Habilidades relacionadas

Looking for an alternative to add-model or another community skill for your workflow? Explore these related open-source skills.

Ver todo

openclaw-release-maintainer

Logo of openclaw
openclaw

Your own personal AI assistant. Any OS. Any Platform. The lobster way. 🦞

333.8k
0
Inteligencia Artificial

widget-generator

Logo of f
f

Generar complementos de widgets personalizables para el sistema de feeds de prompts.chat

149.6k
0
Inteligencia Artificial

flags

Logo of vercel
vercel

El Marco de React

138.4k
0
Navegador

pr-review

Logo of pytorch
pytorch

Tensores y redes neuronales dinámicas en Python con fuerte aceleración de GPU

98.6k
0
Desarrollador