eee_test / README.md
deepmage121's picture
Update dataset card
9d2a6e1 verified
---
configs:
- config_name: default
data_files:
- split: hfopenllm_v2
path: data/hfopenllm_v2.parquet
- split: livecodebenchpro
path: data/livecodebenchpro.parquet
---
# Every Eval Ever Dataset
Evaluation results from various AI model leaderboards.
## Usage
```python
from datasets import load_dataset
# Load specific leaderboard
dataset = load_dataset("deepmage121/eee_test", split="hfopenllm_v2")
# Load all
dataset = load_dataset("deepmage121/eee_test")
```
## Available Leaderboards (Splits)
- `hfopenllm_v2`
- `livecodebenchpro`
## Schema
- `model_name`, `model_id`, `model_developer`: Model information
- `evaluation_source_name`: Leaderboard name
- `evaluation_results`: JSON string with all metrics
- Additional metadata for reproducibility
Auto-updated via GitHub Actions.