Abilities of Contrastive Soft Prompting for Open Domain Rhetorical Question Detection

Authors

  • Josef Baloun Dept. of Computer Science & Engineering, University of West Bohemia, Plzeň, Czech Republic
  • Jiří Martínek Dept. of Computer Science & Engineering, University of West Bohemia, Plzeň, Czech Republic
  • Christophe Cerisara CNRS LORIA, Université de Lorraine, Nancy, France
  • Pavel Král Dept. of Computer Science & Engineering, University of West Bohemia, Plzeň, Czech Republic

Keywords:

Soft prompts, prompt-tuning, rhetorical question, contrastive learning, triplet loss, pre-trained language models

Abstract

In this work, we start by demonstrating experimentally that rhetorical question detection is still a challenging task, even for state-of-the-art Large Language Models (LLMs). We then propose an approach that boosts the performances of such LLMs by training a soft prompt in a way that enables building a joint embedding space from multiple loosely related corpora. The advantages of using a soft-prompt compared to finetuning is to limit the training costs and combat overfitting and forgetting. Soft prompting is often viewed as a way to guide the model towards a specific known task, or to introduce new knowledge into the model through in-context learning. We further show that soft prompting may also be used to modify the geometry of the embedding space, so that the distance between embeddings becomes semantically relevant for a target task, similarly to what is commonly achieved with contrastive finetuning. We exploit this property to combat data scarcity for the task of rhetorical question detection by merging several datasets into a joint semantic embedding space. We finally show on the standard Switchboard dataset that the resulting BERT-based model nearly divides by 2 the number of errors as compared to Flan-T5-XXL with only 5 few-shot labeled samples, thanks to this joint embedding space. We have chosen in our experiments a BERT model because it has already been shown with S-BERT that contrastive finetuning of BERT leads to semantically meaningful representations. Therefore, we also show that this property of BERT nicely transfers to the soft-prompting paradigm. Finally, we qualitatively analyze the resulting embedding space and propose a few heuristic criteria to select appropriate related tasks for inclusion into the pool of training datasets.

Downloads

Download data is not yet available.

Published

2025-06-30

How to Cite

Baloun, J., Martínek, J., Cerisara, C., & Král, P. (2025). Abilities of Contrastive Soft Prompting for Open Domain Rhetorical Question Detection. Computing and Informatics, 44(3). Retrieved from http://147.213.75.17/ojs/index.php/cai/article/view/7269

Most read articles by the same author(s)