Copy the page URI to the clipboard
Yu, Qian; Zhang, Peng; Hou, Yuexian; Song, Dawei and Wang, Jun
(2015).
DOI: https://doi.org/10.1007/978-3-319-16354-3_73
Abstract
Probabilistic language modelling has been widely used in information retrieval. It estimates document models under the multinomial distribution assumption, and uses query likelihood to rank documents. In this paper, we aim to generalize this distribution assumption by exploring the use of fully-observable Boltzmann Machines (BMs) for document modelling. BM is a stochastic recurrent network and is able to model the distribution of multi-dimensional variables. It yields a kind of Boltzmann distribution which is more general than multinomial distribution. We propose a Document Boltzmann Machine (DBM) that can naturally capture the intrinsic connections among terms and estimate query likelihood efficiently. We formally prove that under certain conditions (with 1-order parameters learnt only), DBM subsumes the traditional document language model. Its relations to other graphical models in IR, e.g., MRF model, are also discussed. Our experiments on the document re-ranking demonstrate the potential of the proposed DBM.
Viewing alternatives
Metrics
Public Attention
Altmetrics from AltmetricNumber of Citations
Citations from Dimensions- Published Version (PDF) This file is not available for public download
Item Actions
Export
About
- Item ORO ID
- 42212
- Item Type
- Conference or Workshop Item
- ISBN
- 3-319-16353-1, 978-3-319-16353-6
- Academic Unit or School
-
Faculty of Science, Technology, Engineering and Mathematics (STEM) > Computing and Communications
Faculty of Science, Technology, Engineering and Mathematics (STEM) - Copyright Holders
- © 2015 Springer International Publishing Switzerland
- Depositing User
- Dawei Song