Claude Skill

InternLM/WildClawBench

WildClawBench is an in-the-wild benchmark for evaluating AI agents in the OpenClaw environment, supporting agentic AI research and evaluation.

Overview

Stars367
Forks25
LanguagePython
Last pushed2026-05-15
Last synced2026-05-15
View on GitHub

Repository

OwnerInternLM
RepositoryWildClawBench
Full nameInternLM/WildClawBench
Repo ID1,189,335,371

🚀 Install this Skill

openclaw install InternLM/WildClawBench

Summary

WildClawBench is an in-the-wild benchmark designed to evaluate AI agents operating within the OpenClaw environment, providing a realistic and challenging testbed for agentic AI systems.

Chinese description

OpenClaw环境中AI代理的野外基准测试。

Key features

  • In-the-wild benchmark for AI agents
  • Built on the OpenClaw environment
  • Focuses on agentic AI evaluation
  • Realistic and challenging test scenarios

Use cases

  • Evaluating AI agent performance in open environments
  • Benchmarking agentic AI models
  • Research on agentic evaluation methodologies

Topics

Explore more

Data from GitHub. Synced on 2026-05-15