3260 papers • 126 benchmarks • 313 datasets
Simulation of user interacting with a dialog system
(Image credit: Papersgraph)
These leaderboards are used to track progress in goal-oriented-dialog
No benchmarks available.
Use these libraries to find goal-oriented-dialog models and implementations
No datasets available.
No subtasks available.
A fully-observed dataset from Kuaishou's online environment, where almost all 1,411 users have been exposed to all 3,327 items, and a preliminary analysis of how the two factors - data density and exposure bias - affect the evaluation results of multi-round conversational recommendation.
A new, publicly available simulation framework, where the simulator, designed for the movie-booking domain, leverages both rules and collected data, and several agents are demonstrated and the procedure to add and test your own agent is detailed.
The large size and rich annotation of CrossWOZ make it suitable to investigate a variety of tasks in cross-domain dialogue modeling, such as dialogue state tracking, policy learning, user simulation, etc.
A novel application of large language models in user simulation for task-oriented dialog systems, specifically focusing on an in-context learning approach, which eliminates the need for labor-intensive rule definition or extensive annotated data, making it more efficient and accessible.
A user satisfaction annotation dataset, USS, is proposed that includes 6,800 dialogues sampled from multiple domains, spanning real-world e-commerce dialogues, task-oriented dialogues constructed through Wizard-of-Oz experiments, and movie recommendation dialogues and three baseline methods for user satisfaction prediction and action prediction tasks.
The user simulator aims to generate responses that a real human would give by considering both individual preferences and the general flow of interaction with the system, and shows that preference modeling and task-specific interaction models both contribute to more realistic simulations.
This work proposed a reactive agent model which can ensure safety without comprising the original purposes, by learning only high-level decisions from expert data and a low level decentralized controller guided by the jointly learned decentralized barrier certificates.
This paper proposes a conversational User Simulator, called USi, for automatic evaluation of conversational search systems, capable of automatically answering clarifying questions about the topic throughout the search session, and shows that responses generated by USi are both inline with the underlying information need and comparable to human-generated answers.
This work proposes the Variational Hierarchical Dialog Autoencoder (VHDA) for modeling complete aspects of goal-oriented dialogs, including linguistic features and underlying structured annotations, namely dialog acts and goals, and proposes two training policies to mitigate issues that arise with training VAE-based models.
A method of standardizing user simulator building is proposed that can be used by the community to compare dialog system quality using the same set of user simulators fairly and asks human users to assess the simulators directly and indirectly by rating the simulated dialogs and interacting with the trained systems.
Adding a benchmark result helps the community track progress.