bulk-inference — for Claude Code bulk-inference, counterfactual-chart-reasoning, community, for Claude Code, ide skills, vllm_local, vllm-serve, openai, OPENAI_API_KEY, gemini

v1.0.0

이 스킬 정보

적합한 상황: Ideal for AI agents that need input jsonl file with at minimum: an image path field, a question/prompt field, and one or more id. 현지화된 요약: Handles JSONL input/output, resume from interruption, and concurrent async requests. This AI agent skill supports Claude Code, Cursor, and Windsurf workflows.

기능

Input JSONL file with at minimum: an image path field, a question/prompt field, and one or more ID
For vllm local: running vLLM server(s) — use /vllm-serve first.
For openai: OPENAI API KEY env var set.
For gemini: GOOGLE API KEY env var set.
Gather parameters from user :

# Core Topics

pminhyung pminhyung
[0]
[0]
Updated: 3/14/2026

Killer-Skills Review

Decision support comes first. Repository text comes second.

Reference-Only Page Review Score: 10/11

This page remains useful for teams, but Killer-Skills treats it as reference material instead of a primary organic landing page.

Original recommendation layer Concrete use-case guidance Explicit limitations and caution Quality floor passed for review
Review Score
10/11
Quality Score
64
Canonical Locale
en
Detected Body Locale
en

적합한 상황: Ideal for AI agents that need input jsonl file with at minimum: an image path field, a question/prompt field, and one or more id. 현지화된 요약: Handles JSONL input/output, resume from interruption, and concurrent async requests. This AI agent skill supports Claude Code, Cursor, and Windsurf workflows.

이 스킬을 사용하는 이유

추천 설명: bulk-inference helps agents input jsonl file with at minimum: an image path field, a question/prompt field, and one or more id. Handles JSONL input/output, resume from interruption, and concurrent async

최적의 용도

적합한 상황: Ideal for AI agents that need input jsonl file with at minimum: an image path field, a question/prompt field, and one or more id.

실행 가능한 사용 사례 for bulk-inference

사용 사례: Applying Input JSONL file with at minimum: an image path field, a question/prompt field, and one or more ID
사용 사례: Applying For vllm local: running vLLM server(s) — use /vllm-serve first
사용 사례: Applying For openai: OPENAI API KEY env var set

! 보안 및 제한 사항

  • 제한 사항: Requires repository-specific context from the skill documentation
  • 제한 사항: Works best when the underlying tools and dependencies are already configured

Why this page is reference-only

  • - Current locale does not satisfy the locale-governance contract.

Source Boundary

The section below is imported from the upstream repository and should be treated as secondary evidence. Use the Killer-Skills review above as the primary layer for fit, risk, and installation decisions.

After The Review

Decide The Next Action Before You Keep Reading Repository Material

Killer-Skills should not stop at opening repository instructions. It should help you decide whether to install this skill, when to cross-check against trusted collections, and when to move into workflow rollout.

Labs Demo

Browser Sandbox Environment

⚡️ Ready to unleash?

Experience this Agent in a zero-setup browser environment powered by WebContainers. No installation required.

Boot Container Sandbox

FAQ & Installation Steps

These questions and steps mirror the structured data on this page for better search understanding.

? Frequently Asked Questions

What is bulk-inference?

적합한 상황: Ideal for AI agents that need input jsonl file with at minimum: an image path field, a question/prompt field, and one or more id. 현지화된 요약: Handles JSONL input/output, resume from interruption, and concurrent async requests. This AI agent skill supports Claude Code, Cursor, and Windsurf workflows.

How do I install bulk-inference?

Run the command: npx killer-skills add pminhyung/counterfactual-chart-reasoning/bulk-inference. It works with Cursor, Windsurf, VS Code, Claude Code, and 19+ other IDEs.

What are the use cases for bulk-inference?

Key use cases include: 사용 사례: Applying Input JSONL file with at minimum: an image path field, a question/prompt field, and one or more ID, 사용 사례: Applying For vllm local: running vLLM server(s) — use /vllm-serve first, 사용 사례: Applying For openai: OPENAI API KEY env var set.

Which IDEs are compatible with bulk-inference?

This skill is compatible with Cursor, Windsurf, VS Code, Trae, Claude Code, OpenClaw, Aider, Codex, OpenCode, Goose, Cline, Roo Code, Kiro, Augment Code, Continue, GitHub Copilot, Sourcegraph Cody, and Amazon Q Developer. Use the Killer-Skills CLI for universal one-command installation.

Are there any limitations for bulk-inference?

제한 사항: Requires repository-specific context from the skill documentation. 제한 사항: Works best when the underlying tools and dependencies are already configured.

How To Install

  1. 1. Open your terminal

    Open the terminal or command line in your project directory.

  2. 2. Run the install command

    Run: npx killer-skills add pminhyung/counterfactual-chart-reasoning/bulk-inference. The CLI will automatically detect your IDE or AI agent and configure the skill.

  3. 3. Start using the skill

    The skill is now active. Your AI agent can use bulk-inference immediately in the current project.

! Reference-Only Mode

This page remains useful for installation and reference, but Killer-Skills no longer treats it as a primary indexable landing page. Read the review above before relying on the upstream repository instructions.

Upstream Repository Material

The section below is imported from the upstream repository and should be treated as secondary evidence. Use the Killer-Skills review above as the primary layer for fit, risk, and installation decisions.

Upstream Source

bulk-inference

Handles JSONL input/output, resume from interruption, and concurrent async requests. This AI agent skill supports Claude Code, Cursor, and Windsurf workflows.

SKILL.md
Readonly
Upstream Repository Material
The section below is imported from the upstream repository and should be treated as secondary evidence. Use the Killer-Skills review above as the primary layer for fit, risk, and installation decisions.
Supporting Evidence

Bulk Inference

Purpose

Execute bulk VLM inference across multiple providers (vLLM local, OpenAI, Gemini) using scripts/inference_runner.py. Handles JSONL input/output, resume from interruption, and concurrent async requests.

Prerequisites

  • Input JSONL file with at minimum: an image path field, a question/prompt field, and one or more ID fields.
  • For vllm_local: running vLLM server(s) — use /vllm-serve first.
  • For openai: OPENAI_API_KEY env var set.
  • For gemini: GOOGLE_API_KEY env var set.

Process

  1. Gather parameters from user:

    • --provider: vllm_local, openai, or gemini
    • --endpoints: server URLs (vllm_local) or API base URL
    • --model-id: HF model name or API model ID
    • --input: path to input JSONL
    • --output: path for output JSONL
    • --n-concurrent: requests per endpoint (vllm) or total (API), default 6
    • --max-tokens: default 100
    • --temperature: default 0.0
    • Optional: --api-key-env, --reasoning-effort, --thinking-budget, --rate-limit-delay
    • Optional: --image-field, --question-field, --id-fields, --prompt-template
  2. Validate inputs — Confirm input JSONL exists and is readable. Check provider-specific requirements (API keys, server health).

  3. Run inference:

    bash
    1python scripts/inference_runner.py \ 2 --provider {provider} \ 3 --endpoints {urls} \ 4 --model-id {model_id} \ 5 --input {input_jsonl} \ 6 --output {output_jsonl} \ 7 --n-concurrent {n} \ 8 --max-tokens {max_tokens} \ 9 --temperature {temp} \ 10 [--api-key-env {env_var}] \ 11 [--reasoning-effort {effort}] \ 12 [--thinking-budget {budget}] \ 13 [--rate-limit-delay {delay}] \ 14 [--no-resume] \ 15 [--image-field {field}] \ 16 [--question-field {field}] \ 17 [--id-fields {f1},{f2}] \ 18 [--prompt-template "Answer the question..."]
  4. Monitor output — The script prints a tqdm progress bar and final summary with total, success, errors, and throughput.

  5. Report results — After completion, report: output file path, total processed, success rate, error count.

Input JSONL Format

Each line is a JSON object. Required fields are configurable via --image-field, --question-field, --id-fields. Defaults:

  • image_path — path to image file
  • question_string — prompt/question text
  • triplet_id, condition — composite ID for resume

Output JSONL Format

Each output line preserves ALL original input fields plus:

json
1{"...original fields...", "model": "...", "raw_response": "...", "parsed_answer": "...", "error": null}

Rules

  • Resume is ON by default — interrupted runs continue from where they stopped.
  • Never modify the input JSONL file.
  • Append mode: output JSONL is opened in append mode, one line per completed item.
  • All errors are captured per-item; the runner does not abort on individual failures.

관련 스킬

Looking for an alternative to bulk-inference or another community skill for your workflow? Explore these related open-source skills.

모두 보기

openclaw-release-maintainer

Logo of openclaw
openclaw

Your own personal AI assistant. Any OS. Any Platform. The lobster way. 🦞

333.8k
0
인공지능

widget-generator

Logo of f
f

prompts.chat 피드 시스템을 위한 사용자 지정 가능한 위젯 플러그인을 생성합니다

149.6k
0
인공지능

flags

Logo of vercel
vercel

리액트 프레임워크

138.4k
0
브라우저

pr-review

Logo of pytorch
pytorch

파이썬에서 텐서와 동적 신경망 구현 및 강력한 GPU 가속 지원

98.6k
0
개발자