APS Logo

Explainable Natural Language Processing with Matrix Product States

ORAL

Abstract

Deep Learning (DL) has empowered computers with superior performance in modern Natural Language Processing (NLP) tasks, such as sentiment analysis and machine translation. Even for texts with long-range correlations such as sequences of characters in Wikipedia, DL can effectively express the power-law decay in the mutual information between two distant characters [1]. Despite empirical successes, its intrinsic non-linearity complicates the analysis of algorithmic behaviours. Which network architectures and how many parameters are essential to reproduce long-range correlations are important yet theoretically challenging questions to tackle. Here, we attempt to provide systematic answers through the mapping between DL and its matrix product state (MPS) counterpart [2]. By recasting DL as MPS, we show that the number of parameters required to achieve high performance in sentiment analysis, and to reproduce power-law decay in the mutual information in Wikipedia texts, can be efficiently extracted from the entanglement entropy in the dual MPS. Our work utilises tools in many-body quantum physics to resolve explainability issues of NLP, and more generally of sequence modelling.

[1] H. W. Lin, M. Tegmark, Entropy, 19, 299 (2017)
[2] Y. Levine et al. Phys. Rev. Lett, 122, 065301 (2019)

Presenters

  • thiparat chotibut

    Physics, Chulalongkorn University

Authors

  • Pradeep Bhadola

    The Institute for Fundamental Study, Naresuan University

  • Jirawat Tangpanitanon

    Quantum Technology Foundation (Thailand), Centre for Quantum Technologies, NUS, Natl Univ of Singapore

  • Chanatip Mangkang

    Physics, Chulalongkorn University

  • Yuichiro Minato

    blueqat Inc.

  • Dimitris Angelakis

    Centre for Quantum Technologies, National University of Singapore, Centre for Quantum Technologies, NUS, Natl Univ of Singapore

  • thiparat chotibut

    Physics, Chulalongkorn University