Learning Open Domain Multi-hop Search Using Reinforcement Learning
Citation
Enrique Noriega-Atala, Mihai Surdeanu, and Clayton Morrison. 2022. Learning Open Domain Multi-hop Search Using Reinforcement Learning. In Proceedings of the Workshop on Structured and Unstructured Knowledge Integration (SUKI), pages 26–35, Seattle, USA. Association for Computational Linguistics.Rights
Copyright © 2022 Association for Computational Linguistics. This is an open access article licensed on a Creative Commons Attribution 4.0 International License.Collection Information
This item from the UA Faculty Publications collection is made available by the University of Arizona with support from the University of Arizona Libraries. If you have questions, please contact us at repository@u.library.arizona.edu.Abstract
We propose a method to teach an automated agent to learn how to search for multi-hop paths of relations between entities in an open domain. The method learns a policy for directing existing information retrieval and machine reading resources to focus on relevant regions of a corpus. The approach formulates the learning problem as a Markov decision process with a state representation that encodes the dynamics of the search process and a reward structure that minimizes the number of documents that must be processed while still finding multi-hop paths. We implement the method in an actor-critic reinforcement learning algorithm and evaluate it on a dataset of search problems derived from a subset of English Wikipedia. The algorithm finds a family of policies that succeeds in extracting the desired information while processing fewer documents compared to several baseline heuristic algorithms. © 2022 Association for Computational Linguistics.Note
Open access journalISBN
9781955917865Version
Final published versionae974a485f413a2113503eed53cd6c53
10.18653/v1/2022.suki-1.4
Scopus Count
Collections
Except where otherwise noted, this item's license is described as Copyright © 2022 Association for Computational Linguistics. This is an open access article licensed on a Creative Commons Attribution 4.0 International License.

