Skip to main content
The following sections help you create datasets, run evaluations, and analyze results:

Evaluation concepts

Review core terminology and concepts to understand how evaluations work in LangSmith.

Manage datasets

Create and manage datasets for evaluation through the UI or SDK.

Run evaluations

Evaluate your applications with different evaluators and techniques to measure quality.

Analyze results

View and analyze evaluation results, compare experiments, filter data, and export findings.

Collect feedback

Gather human feedback through annotation queues and inline annotation on outputs.

Follow tutorials

Learn by following step-by-step tutorials, from simple chatbots to complex agent evaluations.

Connect these docs programmatically to Claude, VSCode, and more via MCP for real-time answers.