AlignLLM: alignment-based evaluation using ensemble of LLMs-as-judges for Q &A.
(2025)
Presentation / Conference Contribution
ABEYRATNE, R., WIRATUNGA, N., MARTIN, K., NKISI-ORJ, I. and JAYAWARDENA, L. 2025. AlignLLM: alignment-based evaluation using ensemble of LLMs-as judges for Q&A. In Bichindaritz, I. and López, B. (eds.) Case-based reasoning research and development: proceedings of the 33rd International conference on case-based reasoning 2025 (ICCBR 2025), 30 June - 3 July 2025, Biarritz, France. Lecture notes in computer science (LNCS), 15662. Cham: Springer [online], pages 21-36. Available from: https://doi.org/10.1007/978-3-031-96559-3_2
Evaluating responses generated by large language models (LLMs) is challenging in the absence of ground-truth knowledge, particularly in specialised domains such as law. Increasingly, LLMs themselves are used to evaluate the responses they generate; h... Read More about AlignLLM: alignment-based evaluation using ensemble of LLMs-as-judges for Q &A..