ROOK-CLF-9M Chess Analysis

Strategic reasoning through language models • 49% action accuracy (ChessBench)57% checkmate-in-one (BIG-bench) • 9M parameter LLaMA • Trained on ChessBench (40M)

Jonathan Rahn • Jenia Jitsev • Qi Sun • LAION Research Project

Loading ROOK-CLF-9M Model

Initializing...

Preparing to load model files...

rnbqkbnr/pppppppp/8/8/8/8/PPPPPPPP/RNBQKBNR w KQkq - 0 1

Analysis Results

Top Predicted Moves (click to play)

Loading model...

Model Information

  • Architecture: LlamaForSequenceClassification
  • Parameters: 9M
  • Classes: 1968 chess moves
  • Input: Customized FEN notation
  • Runtime: ONNX WebAssembly

Research Benchmark Evaluation

Evaluate ROOK-CLF-9M performance across established research benchmarks using authentic datasets and methodologies

Accuracy Over Time

Performance
0.0%
Throughput
0.0
Progress
0 / 0
ETA
--:--

Overall Progress

Recent Results

Start a benchmark to see results

ChessBench Puzzles Benchmark

Target: 49% 1,000 positions

Original Source

Tactical puzzle solve rate evaluation from "Grandmaster-level chess without search" paper. Measures model's ability to find correct moves in tactical sequences, not isolated best-move accuracy.

Evaluation Methodology

Original Research
CSV Format:
PuzzleId,FEN,Moves,Solution
00MTG,4r1k1/2p1q...,h4f2 f1f2 e2f2,Bf2+ Rxf2 ...

Evaluation: Puzzle solve rate: model tested on every other move in tactical sequences (when model's turn to play)

ROOK-CLF Adaptation
JSON Format:
{"fen": "4r1k1/2p1q...", "correct_move": "f1f2"}

Evaluation: Position-based tactical move prediction at model decision points, evaluating puzzle-solving capability

Citation

Ruoss et al. 2024. Grandmaster-level chess without search. arXiv:2402.04494

Attention Rollout

Compute square relevance from exported attentions and explore how focus evolves across layers.

all

Attention Rollout

Residual‑corrected attention aggregated across layers. Shows token‑to‑decision relevance per layer. Use the slider to scrub layers.

Early Logit Lens

For each layer’s decision-token state, projects with the final classifier to show layer-wise move beliefs. Hover a move to highlight its squares.

Awaiting run…

Model Diagram & Heads

Interactive overview of layers and heads. Click a head to view that head’s attention at the selected layer; click “Mean over heads” to return.

Layers: 1..8 • Heads: 1..8

Top‑k Attention Paths

Strongest paths to the decision token. Opacity scales with path strength.

Occlusion Heatmap

Forward‑only masking. Shows Δ for the selected move; click a move in the Logit Lens.

Enter a FEN and click Run. Use the slider to explore layers. Hover moves in the Logit Lens to see overlays.