showSidebars ==
showTitleBreadcrumbs == 1
node.field_disable_title_breadcrumbs.value ==

Research Seminar by Grace Hui Yang | Sequencing Matters: A Generate-Retrieve-Generate Model for Building Conversational Agents

Please click here if you are unable to view this page.

 

Sequencing Matters: A Generate-Retrieve-Generate Model
for Building Conversational Agents

Speaker (s):



Grace Hui Yang
Associate Professor, 
Georgetown University, 
Department of Computer Science

Date:

Time:

Venue:

 

20 August 2024, Tuesday

10:30am – 11:30am

School of Computing & 
Information Systems 2 (SCIS 2)
Level 2, Seminar Room 2-7
Singapore Management University
90 Stamford Road
Singapore 178903

Please register by 19 August 2024.

We look forward to seeing you at this research seminar.

About the Talk

Conversational information retrieval systems transform the traditional query-response model into an interactive dialogue, allowing users to engage with information retrieval systems through natural language conversations. Traditionally, conversational information retrieval systems employ a Retrieve-Then-Generate pipeline that initially performs document or passage retrieval, followed by extracting or generating answers from the retrieved results. In this paper, we propose an alternative approach, "Generate-Retrieve-Generate" (GRG), which significantly outperforms the Retrieve-Then-Generate method in the recent TREC (Text REtrieval Conference) 2023 Interactive Knowledge Assistant Track (iKAT). We found that even when the GRG approach uses simpler retrieval methods like BM25, it can still outscore Retrieve-Then-Generate methods that employ more complex dense or sparse neural retrievers. This phenomenon raises intriguing questions about the impact of the order in which the retrieval engine and the Large Language Model (LLM)-supported generation are invoked. Initiating the entire process with generation may provide a more contextually enriched query for the retrieval process, which in turn, might improve the specificity and relevance of the information retrieved, even when using simpler algorithms. In this paper, we investigate more details of how various parameters and settings impact the GRG pipeline in conversational information retrieval. We also investigate, via reinforcement learning, the optimal number of interplays between generation and retrieval steps to find the optimal balance between efficiency and quality, preventing unnecessary iterations that may not contribute to further improvement or even decrease the results.

ABOUT THE SPEAKER

Dr. Grace Hui Yang is an Associate Professor at Georgetown University, leading the InfoSense group. She earned her Ph.D. from Carnegie Mellon University in 2011. Her research focuses on AI, deep reinforcement learning, conversational agents, search engines, and privacy-preserving information retrieval, with prior work in question answering, ontology construction etc. Supported by DARPA and NSF, she led the TREC Dynamic Domain Tracks (2015-2017) and SIGIR privacy-preserving workshops (2014-2016). Dr. Yang is an associate editor for ACM Transactions on Information Systems, has held leadership roles in top conferences like SIGIR, ACL, WSDM, and WWW, and was general co-chair of SIGIR 2024. She is a recipient of the NSF CAREER Award and authored the book Dynamic Information Retrieval Modeling (2016).