Add Model

Claude 4.6 Opus

v1.0.0

Sobre este Skill

Install Add Model, an AI agent skill for AI agent workflows and automation. Explore features, use cases, limitations, and setup guidance.

Recursos

Add a New Model to the Eval Runner
Follow these steps whenever the user asks to add a new AI model to the eval suite.
Step 0: Gather Information
Determine the following (ask the user if not provided):
Formatted name - human-readable, e.g. Claude 4.6 Opus.

# Tópicos principais

get-convex get-convex
[109]
[9]
Atualizado: 4/17/2026

Skill Overview

Start with fit, limitations, and setup before diving into the repository.

Install Add Model, an AI agent skill for AI agent workflows and automation. Explore features, use cases, limitations, and setup guidance.

Por que usar essa habilidade

Recomendacao: add-model helps agents add a new model to the eval runner. Add a New Model to the Eval Runner Follow these steps whenever the user asks to add a new AI model to the eval suite. This AI agent skill

Melhor para

Cenario recomendado: add a new model to the eval runner.

Casos de Uso Práticos for Add Model

Caso de uso: Add a New Model to the Eval Runner
Caso de uso: Follow these steps whenever the user asks to add a new AI model to the eval suite
Caso de uso: Step 0: Gather Information

! Segurança e Limitações

  • Limitacao: apiKind - only needed for OpenAI Codex/Responses-API models; set to "responses". Omit for all other models.
  • Limitacao: // apiKind: "responses", // only for OpenAI Codex / Responses-API models
  • Limitacao: Only proceed to the next step once at least one eval completes successfully.

About The Source

The section below is adapted from the upstream repository. Use it as supporting material alongside the fit, use-case, and installation summary on this page.

Demo Labs

Browser Sandbox Environment

⚡️ Ready to unleash?

Experience this Agent in a zero-setup browser environment powered by WebContainers. No installation required.

Boot Container Sandbox

FAQ e etapas de instalação

These questions and steps mirror the structured data on this page for better search understanding.

? Perguntas frequentes

O que é Add Model?

Install Add Model, an AI agent skill for AI agent workflows and automation. Explore features, use cases, limitations, and setup guidance.

Como instalar Add Model?

Execute o comando: npx killer-skills add get-convex/convex-evals. Ele funciona com Cursor, Windsurf, VS Code, Claude Code e mais de 19 outros IDEs.

Quais são os casos de uso de Add Model?

Os principais casos de uso incluem: Caso de uso: Add a New Model to the Eval Runner, Caso de uso: Follow these steps whenever the user asks to add a new AI model to the eval suite, Caso de uso: Step 0: Gather Information.

Quais IDEs são compatíveis com Add Model?

Esta skill é compatível com Cursor, Windsurf, VS Code, Trae, Claude Code, OpenClaw, Aider, Codex, OpenCode, Goose, Cline, Roo Code, Kiro, Augment Code, Continue, GitHub Copilot, Sourcegraph Cody, and Amazon Q Developer. Use a CLI do Killer-Skills para uma instalação unificada.

Add Model tem limitações?

Limitacao: apiKind - only needed for OpenAI Codex/Responses-API models; set to "responses". Omit for all other models.. Limitacao: // apiKind: "responses", // only for OpenAI Codex / Responses-API models. Limitacao: Only proceed to the next step once at least one eval completes successfully..

Como instalar este skill

  1. 1. Abra o terminal

    Abra o terminal ou linha de comando no diretório do projeto.

  2. 2. Execute o comando de instalação

    Execute: npx killer-skills add get-convex/convex-evals. A CLI detectará sua IDE ou agente automaticamente e configurará a skill.

  3. 3. Comece a usar o skill

    O skill já está ativo. Seu agente de IA pode usar Add Model imediatamente no projeto atual.

! Source Notes

This page is still useful for installation and source reference. Before using it, compare the fit, limitations, and upstream repository notes above.

Upstream Repository Material

The section below is adapted from the upstream repository. Use it as supporting material alongside the fit, use-case, and installation summary on this page.

Upstream Source

Add Model

Install Add Model, an AI agent skill for AI agent workflows and automation. Explore features, use cases, limitations, and setup guidance.

SKILL.md
Readonly
Upstream Repository Material
The section below is adapted from the upstream repository. Use it as supporting material alongside the fit, use-case, and installation summary on this page.
Upstream Source

Add a New Model to the Eval Runner

Follow these steps whenever the user asks to add a new AI model to the eval suite.

Step 0: Gather Information

Determine the following (ask the user if not provided):

  1. Model identifier - the OpenRouter-style name, e.g. anthropic/claude-opus-4.6. If the user gives a marketing name or URL, look up the OpenRouter model id.
  2. Formatted name - human-readable, e.g. Claude 4.6 Opus.
  3. Provider family & version - needed to find older siblings (e.g. claude-opus-4.5 is the predecessor of claude-opus-4.6).
  4. apiKind - only needed for OpenAI Codex/Responses-API models; set to "responses". Omit for all other models.

If you're unsure, check how the closest existing model in the same family is configured in runner/models/index.ts and match it.

Step 1: Add the Model to runner/models/index.ts

Open runner/models/index.ts and add a new entry to the ALL_MODELS array. Place it next to its family siblings, respecting the existing grouping comments.

Template:

typescript
1{ 2 name: "<provider>/<model-id>", 3 formattedName: "<Human Name>", 4 // apiKind: "responses", // only for OpenAI Codex / Responses-API models 5},

Step 2: Update the Manual Evals Workflow

Open .github/workflows/manual_evals.yml and replace the entire matrix.model list with only the new model. This workflow exists solely to collect baseline data for newly added models, so it should only ever contain the latest addition.

yaml
1matrix: 2 model: 3 - "<provider>/<model-id>"

Step 3: Typecheck

Run bun run typecheck to verify no type errors were introduced.

Step 4: Smoke Test the New Model Locally

Before committing, run a quick local sanity check with one or two simple evals to confirm the model ID is valid, the API key works, and results are being produced. Use the simplest fundamentals evals:

bash
1MODELS=<new-model-name> TEST_FILTER="000-fundamentals/000" bun run local:run

If that passes, optionally run one more:

bash
1MODELS=<new-model-name> TEST_FILTER="000-fundamentals/001" bun run local:run

What to look for:

  • No authentication or "model not found" errors
  • The run completes and produces a score (even a low score is fine - we just want to confirm it runs)
  • If it fails with an API/auth error, stop and fix the model ID or check the .env file before proceeding

Only proceed to the next step once at least one eval completes successfully.

Step 5: Commit and Push

Create a descriptive commit message and push to main:

git add runner/models/index.ts .github/workflows/manual_evals.yml
git commit -m "add <model-name>; demote older <family> versions"
git push origin main

Step 6: Trigger Manual Eval Runs for Baseline Data

Use the GitHub CLI to dispatch the manual eval workflow 3 times (to get a statistically meaningful baseline):

bash
1gh workflow run manual_evals.yml --ref main

Run this command 3 times, waiting ~5 seconds between dispatches to avoid collisions.

Step 7: Monitor the Runs Until Completion

You MUST poll until all 3 runs reach a terminal state (completed/failed/cancelled). Do not stop monitoring early or hand back to the user while runs are still in progress.

Poll every ~2 minutes using:

bash
1gh run list --workflow=manual_evals.yml --limit=6

Runs typically take 20-30 minutes. Keep checking until all show completed. If a run fails, immediately investigate:

bash
1gh run view <run-id> --log-failed

Report the final pass/fail status for each run to the user once all 3 are done.

Summary Checklist

  • Model added to ALL_MODELS in runner/models/index.ts
  • .github/workflows/manual_evals.yml matrix replaced with only the new model
  • bun run typecheck passes
  • Smoke test: at least one eval completes successfully locally
  • Changes committed and pushed to main
  • Manual eval workflow dispatched 3 times
  • All 3 runs monitored to completion

Habilidades Relacionadas

Looking for an alternative to Add Model or another community skill for your workflow? Explore these related open-source skills.

Ver tudo

openclaw-release-maintainer

Logo of openclaw
openclaw

Resumo localizado: 🦞 # OpenClaw Release Maintainer Use this skill for release and publish-time workflow. It covers ai, assistant, crustacean workflows. Claude Code, Cursor, and Windsurf workflows.

widget-generator

Logo of f
f

Resumo localizado: Generate customizable widget plugins for the prompts.chat feed system # Widget Generator Skill This skill guides creation of widget plugins for prompts.chat. It covers ai, artificial-intelligence, awesome-list workflows. Claude Code, Cursor, and

flags

Logo of vercel
vercel

Resumo localizado: The React Framework # Feature Flags Use this skill when adding or changing framework feature flags in Next.js internals. It covers blog, browser, compiler workflows. Claude Code, Cursor, and Windsurf workflows.

138.4k
0
Navegador

pr-review

Logo of pytorch
pytorch

Resumo localizado: Usage Modes No Argument If the user invokes /pr-review with no arguments, do not perform a review. It covers autograd, deep-learning, gpu workflows. Claude Code, Cursor, and Windsurf workflows.

98.6k
0
Desenvolvedor