Home

Home

OCPSG Benchmarking LLMs is a project website for benchmark development, documentation, and reproducible evaluation workflows for large language models and fine-tuned models applied to multilingual policy agenda annotation in parliamentary speeches.

The project is developed within the Oxford Computational Political Science Group (OCPSG), a research initiative supported by Oxford’s Department of Politics and International Relations.

What this project does

This website serves as the public entry point for the benchmark and its supporting infrastructure. It is designed to make benchmark design, model evaluation, and submission workflows more transparent, standardised, and easier to reproduce across research teams.

Pipeline overview

The benchmark workflow moves from multilingual parliamentary speech data to standardised evaluation outputs and documentation.

Pipeline diagram of the OCPSG Benchmarking LLMs project

What this site provides

  • multilingual benchmark datasets for parliamentary speech classification;
  • standardised prediction and evaluation contracts;
  • model wrappers and submission templates;
  • reproducible scoring workflows and reporting conventions;
  • documentation for benchmark use, outputs, and submissions.

Current focus

The current development phase is centred on benchmarking workflows for policy agenda annotation using multilingual parliamentary text. The goal is to support systematic comparison across LLMs and fine-tuned classifiers under shared evaluation conditions.

  • silver-standard benchmark construction;
  • multilingual parliamentary speech data;
  • standardised evaluation pipelines;
  • reproducible model submissions and outputs.

Explore the site

You can use this site to:

  • learn about the benchmark and its scope;
  • access documentation for workflows and submissions;
  • browse project news and updates;
  • navigate to related repositories and supporting materials.