JudgeMeNot:个性化大语言模型以模拟希伯来语司法推理 / JudgeMeNot: Personalizing Large Language Models to Emulate Judicial Reasoning in Hebrew
1️⃣ 一句话总结
这篇论文提出了一种结合合成与有机监督的方法,能够高效地利用少量数据,将大语言模型个性化定制成能模仿特定法官在希伯来语环境下的司法推理风格,其生成结果与真实法官的推理难以区分。
Despite significant advances in large language models, personalizing them for individual decision-makers remains an open problem. Here, we introduce a synthetic-organic supervision pipeline that transforms raw judicial decisions into instruction-tuning data, enabling parameter-efficient fine-tuning of personalized models for individual judges in low-resource settings. We compare our approach to state-of-the-art personalization techniques across three different tasks and settings. The results show that Causal Language Modeling followed by synthetically generated instruction-tuning significantly outperforms all other baselines, providing significant improvements across lexical, stylistic, and semantic similarity. Notably, our model-generated outputs are indistinguishable from the reasoning of human judges, highlighting the viability of efficient personalization, even in low-resource settings.
JudgeMeNot:个性化大语言模型以模拟希伯来语司法推理 / JudgeMeNot: Personalizing Large Language Models to Emulate Judicial Reasoning in Hebrew
这篇论文提出了一种结合合成与有机监督的方法,能够高效地利用少量数据,将大语言模型个性化定制成能模仿特定法官在希伯来语环境下的司法推理风格,其生成结果与真实法官的推理难以区分。
源自 arXiv: 2604.18041