Selecting Better Samples from Pre-trained LLMs: A Case Study on Question Generation - Inria - Institut national de recherche en sciences et technologies du numérique Accéder directement au contenu
Pré-Publication, Document De Travail (Preprint/Prepublication) Année : 2022

Selecting Better Samples from Pre-trained LLMs: A Case Study on Question Generation

Résumé

Large Language Models (LLMs) have in recent years demonstrated impressive prowess in natural language generation. A common practice to improve generation diversity is to sample multiple outputs from the model. However, there lacks a simple and robust way of selecting the best output from these stochastic samples. As a case study framed in the context of question generation, we propose two prompt-based approaches to selecting high-quality questions from a set of LLM-generated candidates. Our method works under the constraints of 1) a black-box (non-modifiable) question generation model and 2) lack of access to human-annotated references -- both of which are realistic limitations for real-world deployment of LLMs. With automatic as well as human evaluations, we empirically demonstrate that our approach can effectively select questions of higher qualities than greedy generation.
Fichier principal
Vignette du fichier
2209.11000.pdf (507.4 Ko) Télécharger le fichier
Origine : Fichiers produits par l'(les) auteur(s)

Dates et versions

hal-03897371 , version 1 (13-12-2022)

Identifiants

  • HAL Id : hal-03897371 , version 1

Citer

Xingdi Yuan, Tong Wang, Yen-Hsiang Wang, Emery Fine, Rania Abdelghani, et al.. Selecting Better Samples from Pre-trained LLMs: A Case Study on Question Generation. 2022. ⟨hal-03897371⟩

Collections

INRIA INRIA2
22 Consultations
29 Téléchargements

Partager

Gmail Facebook X LinkedIn More