·langgraph-testing-evaluation
</>

langgraph-testing-evaluation

Use this skill when you need to test or evaluate LangGraph/LangChain agents: writing unit or integration tests, generating test scaffolds, mocking LLM/tool behavior, running trajectory evaluation (match or LLM-as-judge), running LangSmith dataset evaluations, and comparing two agent versions with A/B-style offline analysis. Use it for Python and JavaScript/TypeScript workflows, evaluator design, experiment setup, regression gates, and debugging flaky/incorrect evaluation results.

16Installs·3Trend·@lubu-labs

Installation

$npx skills add https://github.com/lubu-labs/langchain-agent-skills --skill langgraph-testing-evaluation

How to Install langgraph-testing-evaluation

Quickly install langgraph-testing-evaluation AI skill to your development environment via command line

  1. Open Terminal: Open your terminal or command line tool (Terminal, iTerm, Windows Terminal, etc.)
  2. Run Installation Command: Copy and run this command: npx skills add https://github.com/lubu-labs/langchain-agent-skills --skill langgraph-testing-evaluation
  3. Verify Installation: Once installed, the skill will be automatically configured in your AI coding environment and ready to use in Claude Code, Cursor, or OpenClaw

Source: lubu-labs/langchain-agent-skills.

SKILL.md

View raw

Use this file for high-level flow. Load references/ for detailed implementation.

| Validate node logic quickly | Unit tests with mocks | references/unit-testing-patterns.md | | Validate multi-step agent behavior | Trajectory evaluation | references/trajectory-evaluation.md | | Track quality over datasets over time | LangSmith evaluation | references/langsmith-evaluation.md |

| Compare old vs new agent versions | A/B comparison | references/ab-testing.md |

Use this skill when you need to test or evaluate LangGraph/LangChain agents: writing unit or integration tests, generating test scaffolds, mocking LLM/tool behavior, running trajectory evaluation (match or LLM-as-judge), running LangSmith dataset evaluations, and comparing two agent versions with A/B-style offline analysis. Use it for Python and JavaScript/TypeScript workflows, evaluator design, experiment setup, regression gates, and debugging flaky/incorrect evaluation results. Source: lubu-labs/langchain-agent-skills.

Facts (cite-ready)

Stable fields and commands for AI/search citations.

Install command
npx skills add https://github.com/lubu-labs/langchain-agent-skills --skill langgraph-testing-evaluation
Category
</>Dev Tools
Verified
First Seen
2026-02-22
Updated
2026-03-10

Browse more skills from lubu-labs/langchain-agent-skills

Quick answers

What is langgraph-testing-evaluation?

Use this skill when you need to test or evaluate LangGraph/LangChain agents: writing unit or integration tests, generating test scaffolds, mocking LLM/tool behavior, running trajectory evaluation (match or LLM-as-judge), running LangSmith dataset evaluations, and comparing two agent versions with A/B-style offline analysis. Use it for Python and JavaScript/TypeScript workflows, evaluator design, experiment setup, regression gates, and debugging flaky/incorrect evaluation results. Source: lubu-labs/langchain-agent-skills.

How do I install langgraph-testing-evaluation?

Open your terminal or command line tool (Terminal, iTerm, Windows Terminal, etc.) Copy and run this command: npx skills add https://github.com/lubu-labs/langchain-agent-skills --skill langgraph-testing-evaluation Once installed, the skill will be automatically configured in your AI coding environment and ready to use in Claude Code, Cursor, or OpenClaw

Where is the source repository?

https://github.com/lubu-labs/langchain-agent-skills