Bird backdrop
CleverBirds: A Multiple-Choice Benchmark for
Fine-grained Human Knowledge Tracing
Fine-grained visual recognition skills are vital to many expert domains, yet understanding how humans acquire such expertise remains an open challenge. We introduce CleverBirds, a large-scale benchmark for knowledge tracing in fine-grained visual recognition. The dataset contains 17.9 million multiple-choice questions from 40,144 participants across 10,779 bird species, with an average of 444 questions per participant. This dataset was introduced in CleverBirds: A Multiple-Choice Benchmark for Fine-grained Human Knowledge Tracing, to appear at NeurIPS 2025 (Datasets and Benchmarks track).
CleverBirds enables us to study how individuals learn to recognize fine-grained visual distinctions over time. We evaluate state-of-the-art knowledge tracing methods on this benchmark and find that tracking learner knowledge across participant subgroups and question types is challenging, with different forms of contextual information providing varying degrees of predictive benefit.
We collected CleverBirds from the Photo and Sound Quiz of the eBird citizen science platform. In this quiz, participants are shown a bird image and asked to identify the species from a list of options. They receive immediate feedback on the correct answer after each response. Quiz responses were collected from March 2018 to October 2024.
Quiz Examples
Figure 1: Three examples of quiz questions from CleverBirds. Each question has four species options plus a "None of the above" option. The correct answer is indicated in green. All five options are valid answers, and the candidate species differ for each question.
CleverBirds is one of the largest benchmarks for visual knowledge tracing, with substantially more learnable concepts than existing datasets. The dataset captures learning patterns across a diverse participant population, enabling us to study how expertise develops over time.
40,144
Participants
17,859,392
Total Interactions
10,779
Bird Species
444
Avg Questions/User
Participants can choose which location they want to be quizzed on, and quiz locations are distributed globally.
World Map
Figure 2: World map showing quiz locations using Hex 3 polygonal bins, where color intensity encodes the number of interactions per location cell.
The knowledge tracing task requires predicting a participant's response given the current question and the correct answer. Models are provided with additional context such as the participant's interaction history and species information and have to infer the learner's evolving knowledge state to predict their guess.
Task Schema
Figure 3: (Left) Human Learning. Participants learn from CleverBirds quiz questions through repeated interactions. For each question, participants see a bird image and a list of possible species names, which may include the correct answer. After making a guess, they receive feedback with the correct answer. (Right) Knowledge Tracing. The prediction task: given a participant's interaction history, the current question's image, options, and correct answer, predict the participant's guess.
We evaluated a range of machine learning models and state-of-the-art knowledge tracing methods on CleverBirds. We found that tracking learner knowledge is challenging, especially when predicting incorrect choices, with different forms of contextual information providing varying degrees of predictive benefit.
Results
Figure 4: Performance on the multiple-choice and binary tasks. Top-left: accuracy on the full multiple-choice dataset. Top-right: accuracy on the subset of questions answered incorrectly. Bottom-left: macro-averaged accuracy on the binary task. Bottom-right: average precision (AP) for predicting user errors. Models are grouped by color into simple classifiers (RF U, RF S, RF U+S), MLPs (MLP U+S+Img, MLP Img), KT models (LM MCC, LM Seq2seq, AKT1, ATKT2, and simpleKT3), and simple heuristics (Always Correct, Random binary, Random multiple-choice, Conf Prior, Conf Prior Inc).
We release this dataset to support the development and evaluation of new methods for visual knowledge tracing. CleverBirds is among the largest benchmarks of its kind, offering a substantially higher number of learnable concepts. With it, we hope to enable new avenues for studying the development of visual expertise over time and across individuals.
If you found CleverBirds useful, please consider citing our work:
@inproceedings{bossemeyercleverbirds,
  title={CleverBirds: A Multiple-Choice Benchmark for Fine-grained Human Knowledge Tracing},
  author={Bossemeyer, Leonie and Heinrich, Samuel and Van Horn, Grant and Mac Aodha, Oisin},
  booktitle={The Thirty-ninth Annual Conference on Neural Information Processing Systems Datasets and Benchmarks Track}
}