Dataset Viewer
Auto-converted to Parquet Duplicate
The dataset viewer is not available for this split.
Cannot load the dataset split (in streaming mode) to extract the first rows.
Error code:   StreamingRowsError
Exception:    CastError
Message:      Couldn't cast
suite: string
model: string
git_sha: string
timestamp: string
trigger: string
schema_version: string
pr_number: int64
skipped: bool
os: string
chip: string
memory_gb: int64
vllm_mlx_version: string
runner: string
metric_name: string
metric_metric: string
metric_value: double
metric_unit: string
tags_json: string
errors_json: string
to
{'schema_version': Value('string'), 'timestamp': Value('string'), 'git_sha': Value('string'), 'trigger': Value('string'), 'suite': Value('string'), 'model': Value('string'), 'os': Value('string'), 'chip': Value('string'), 'memory_gb': Value('int32'), 'name': Value('string'), 'metric': Value('string'), 'value': Value('float64'), 'unit': Value('string'), 'tag_filter': Value('string'), 'tag_caveats': Value('string')}
because column names don't match
Traceback:    Traceback (most recent call last):
                File "/src/services/worker/src/worker/utils.py", line 99, in get_rows_or_raise
                  return get_rows(
                         ^^^^^^^^^
                File "/src/libs/libcommon/src/libcommon/utils.py", line 272, in decorator
                  return func(*args, **kwargs)
                         ^^^^^^^^^^^^^^^^^^^^^
                File "/src/services/worker/src/worker/utils.py", line 77, in get_rows
                  rows_plus_one = list(itertools.islice(ds, rows_max_number + 1))
                                  ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
                File "/usr/local/lib/python3.12/site-packages/datasets/iterable_dataset.py", line 2690, in __iter__
                  for key, example in ex_iterable:
                                      ^^^^^^^^^^^
                File "/usr/local/lib/python3.12/site-packages/datasets/iterable_dataset.py", line 2227, in __iter__
                  for key, pa_table in self._iter_arrow():
                                       ^^^^^^^^^^^^^^^^^^
                File "/usr/local/lib/python3.12/site-packages/datasets/iterable_dataset.py", line 2251, in _iter_arrow
                  for key, pa_table in self.ex_iterable._iter_arrow():
                                       ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
                File "/usr/local/lib/python3.12/site-packages/datasets/iterable_dataset.py", line 494, in _iter_arrow
                  for key, pa_table in iterator:
                                       ^^^^^^^^
                File "/usr/local/lib/python3.12/site-packages/datasets/iterable_dataset.py", line 384, in _iter_arrow
                  for key, pa_table in self.generate_tables_fn(**gen_kwags):
                                       ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
                File "/usr/local/lib/python3.12/site-packages/datasets/packaged_modules/parquet/parquet.py", line 209, in _generate_tables
                  yield Key(file_idx, batch_idx), self._cast_table(pa_table)
                                                  ^^^^^^^^^^^^^^^^^^^^^^^^^^
                File "/usr/local/lib/python3.12/site-packages/datasets/packaged_modules/parquet/parquet.py", line 147, in _cast_table
                  pa_table = table_cast(pa_table, self.info.features.arrow_schema)
                             ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
                File "/usr/local/lib/python3.12/site-packages/datasets/table.py", line 2281, in table_cast
                  return cast_table_to_schema(table, schema)
                         ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
                File "/usr/local/lib/python3.12/site-packages/datasets/table.py", line 2227, in cast_table_to_schema
                  raise CastError(
              datasets.table.CastError: Couldn't cast
              suite: string
              model: string
              git_sha: string
              timestamp: string
              trigger: string
              schema_version: string
              pr_number: int64
              skipped: bool
              os: string
              chip: string
              memory_gb: int64
              vllm_mlx_version: string
              runner: string
              metric_name: string
              metric_metric: string
              metric_value: double
              metric_unit: string
              tags_json: string
              errors_json: string
              to
              {'schema_version': Value('string'), 'timestamp': Value('string'), 'git_sha': Value('string'), 'trigger': Value('string'), 'suite': Value('string'), 'model': Value('string'), 'os': Value('string'), 'chip': Value('string'), 'memory_gb': Value('int32'), 'name': Value('string'), 'metric': Value('string'), 'value': Value('float64'), 'unit': Value('string'), 'tag_filter': Value('string'), 'tag_caveats': Value('string')}
              because column names don't match

Need help to make the dataset viewer work? Make sure to review how to configure the dataset viewer, and open a discussion for direct support.

MLX Benchmarks

Structured benchmark results for MLX-quantized and other locally-hosted LLMs on Apple Silicon. Covers throughput, time-to-first-token, tool-calling, code generation, reasoning, knowledge, and math suites.

Results are produced by a sweep harness that wires upstream evaluation tools against a local vllm-mlx inference server:

All data here is generated on Apple Silicon hardware (MINISFORUM MS-A2 / M4 Max class), stored in flat columnar Parquet for easy querying, and appended to via unique-filename commits so historical shards are never overwritten.

Quickstart

from datasets import load_dataset

ds = load_dataset("JacobPEvans/mlx-benchmarks")
print(ds)

# Example: average throughput per model
import pandas as pd
df = ds["train"].to_pandas()
throughput_rows = df[df.suite == "throughput"]
print(
    throughput_rows.groupby("model")["metric_value"]
    .mean()
    .sort_values(ascending=False)
)

Raw Parquet fetch (token-optimal for agents):

curl -sSL \
  https://huggingface.co/datasets/JacobPEvans/mlx-benchmarks/resolve/main/data/train-00000-of-00001.parquet \
  -o run.parquet

Schema

Each input benchmark run produces a JSON envelope (see schema.json in this repo for the authoritative v1 spec). The envelope is exploded row-wise into flat scalar columns — one row per entry in the envelope's results[] array. Skipped runs become a single sentinel row with null metric columns and skipped=true. This mirrors the columnar layout used by the Open LLM Leaderboard contents dataset.

Column Type Notes
suite string One of: throughput, ttft, tool-calling, code-accuracy, framework-eval, capability-comparison, coding, reasoning, knowledge, evalplus, math-hard
model string Full model identifier
git_sha string Commit SHA of the generator at run time
timestamp string ISO-8601 UTC start of the run
trigger string schedule, pr, workflow_dispatch, or local
schema_version string Envelope schema version (currently "1")
pr_number int64 PR number if triggered by a pull request, else null
skipped bool True for sentinel rows where the suite was skipped
os string Operating system at run time
chip string CPU/chip identifier
memory_gb int64 Total system RAM
vllm_mlx_version string Backend version if captured
runner string Runner label or local
metric_name string Individual test/measurement name
metric_metric string Metric family (e.g. throughput, latency, score)
metric_value float64 Numeric value
metric_unit string Unit (tok/s, seconds, ratio, ...)
tags_json string JSON-serialized tag dict (per-suite custom metadata)
errors_json string JSON-serialized list of non-fatal errors from the run

Nested fields from the envelope (tags, errors) are preserved as JSON-serialized strings so no information is lost — rehydrate with json.loads(row["tags_json"]).

Update cadence

New rows are appended on every sweep via a unique-filename commit pattern (data/run-{timestamp}-{sha}-{suite}-{model}.parquet). Historical shards are never overwritten. load_dataset() concatenates all data/*.parquet files into a single train split at load time.

License

Apache 2.0 — same as the underlying upstream evaluation tools.

Downloads last month
125