Making Task-Oriented Dialogue Datasets More Natural by Synthetically Generating Indirect User Requests
Indirect User Requests (IURs), such as "It's cold in here" instead of "Could you please increase the temperature?" are common in human-human task-oriented dialogue and require world knowledge and pragmatic reasoning from the listener. While large language models (LLMs) can h...
Saved in:
Main Authors | , , , , , |
---|---|
Format | Journal Article |
Language | English |
Published |
11.06.2024
|
Subjects | |
Online Access | Get full text |
Cover
Loading…
Summary: | Indirect User Requests (IURs), such as "It's cold in here" instead of "Could
you please increase the temperature?" are common in human-human task-oriented
dialogue and require world knowledge and pragmatic reasoning from the listener.
While large language models (LLMs) can handle these requests effectively,
smaller models deployed on virtual assistants often struggle due to resource
constraints. Moreover, existing task-oriented dialogue benchmarks lack
sufficient examples of complex discourse phenomena such as indirectness. To
address this, we propose a set of linguistic criteria along with an LLM-based
pipeline for generating realistic IURs to test natural language understanding
(NLU) and dialogue state tracking (DST) models before deployment in a new
domain. We also release IndirectRequests, a dataset of IURs based on the Schema
Guided Dialog (SGD) corpus, as a comparative testbed for evaluating the
performance of smaller models in handling indirect requests. |
---|---|
DOI: | 10.48550/arxiv.2406.07794 |