Reviewers do not assign scores to papers in the same way. Some tend to give
lower scores and some, higher. Inconsistent scoring can affect which papers
are discussed in program committee meetings and which papers are advocated for,
or how strongly.
This tool is intended to help program committee chairs compensate for
scoring bias. It suggests a compensation to normalize review scores, and
plots average score versus normalized score for each paper.
All processing is done locally in the browser: your data will not be transmitted elsewhere.
Upload the CSV file containing the review scores. The first line should read: