eSourcingData — UK Government Procurement Software
HomeBlogTraining
Training
9 min read

Evaluation management best practice: how to run a legally defensible scoring process

Poor evaluation is the most common cause of procurement challenges. Here is how to structure your evaluation process to be consistent, auditable and challenge-proof.

D

David Harding

Founder, eSourcingData · 5 May 2025

Procurement challenges most commonly arise at the evaluation stage. A process that was technically compliant at every other stage can be undermined by an inconsistent, undocumented or apparently biased evaluation. A defensible evaluation process requires structure, documentation and the right tools.

The four fundamentals of a defensible evaluation

  • Fixed criteria — evaluation criteria, methodology and weightings published before submissions opened and not changed afterwards
  • Independent scoring — panel members score without seeing each other's scores until moderation
  • Documented rationale — every score, for every criterion, for every supplier, with written justification
  • Consistent application — the same standard applied to all suppliers throughout

Before evaluation begins

  • Confirm evaluation criteria and weightings match the published tender documents exactly
  • Collect conflict of interest declarations from all panel members before any submissions are opened
  • Brief the panel on the methodology, scoring scale and what each score level means
  • Confirm each panel member has capacity to complete evaluation before the deadline

During individual scoring

  • Each evaluator scores independently — no discussion, no visibility of other scores
  • Every score must have written rationale — what in the response justifies this score?
  • Evaluators note where responses are unclear or where assumptions have been made

Moderation

  • Moderated scores must be justified — "I feel it should be higher" is not acceptable rationale
  • Document the moderation discussion and reasoning for any score changes
  • Where panel scores diverge by more than 20% of the scale, the divergence must be explicitly addressed and documented

The evaluation report

The evaluation report must include: final moderated scores for all suppliers across all criteria; scoring rationale summary; ranking of all suppliers; award recommendation with justification; confirmation of compliance with the published methodology; and confirmation of conflict of interest status.

Common mistakes that create challenge risk

  • Panel members discussing scores before individual scoring is complete
  • Changing evaluation criteria after submissions are received
  • Scoring without written rationale — "good response" is not a rationale
  • Applying different standards to different suppliers for the same criterion
  • Failing to document that conflict of interest declarations were obtained

More from the blog

All articles
Buyers

PA23 compliance checklist for councils

Buyers

eSourcing software guide UK