Using bidirectional LSTM with BERT for Chinese punctuation prediction

Mingfeng Fang, Haifeng Zhao, Xiao Song, Xin Wang, Shilei Huang

Research output: Chapter in Book/Conference Proceeding/ReportConference Paper published in a bookpeer-review

8 Citations (Scopus)

Abstract

Punctuation prediction is an important step in the post processing of ASR systems. Lack of punctuation text is usually difficult to read and understand. In this paper, we propose a method based on Chinese punctuation prediction by combining the Bidirectional Long Short-Term Memory (BLSTM) and the Bidirectional Encoder Representations from Transformers (BERT), which makes the use of BERT as text encoding layers for learning contextualized word representations for improving the performance of BLSTM network. Compared with the previous punctuation prediction methods based on Recurrent Neural Network (RNN), our method improves the performance of punctuation prediction with the powerful ability of capturing semantics and long-distance dependencies in Chinese unsegmented text. Our experimental results on Chinese news datasets that our BERT-BLSTM based method outperforms the baseline by up to 31.07% absolute in overall micro-F1.

Original languageEnglish
Title of host publicationICSIDP 2019 - IEEE International Conference on Signal, Information and Data Processing 2019
PublisherInstitute of Electrical and Electronics Engineers Inc.
ISBN (Electronic)9781728123455
DOIs
Publication statusPublished - Dec 2019
Externally publishedYes
Event2019 IEEE International Conference on Signal, Information and Data Processing, ICSIDP 2019 - Chongqing, China
Duration: 11 Dec 201913 Dec 2019

Publication series

NameICSIDP 2019 - IEEE International Conference on Signal, Information and Data Processing 2019

Conference

Conference2019 IEEE International Conference on Signal, Information and Data Processing, ICSIDP 2019
Country/TerritoryChina
CityChongqing
Period11/12/1913/12/19

Bibliographical note

Publisher Copyright:
© 2019 IEEE.

Keywords

  • neural network
  • pre-trained language model
  • punctuation prediction

Fingerprint

Dive into the research topics of 'Using bidirectional LSTM with BERT for Chinese punctuation prediction'. Together they form a unique fingerprint.

Cite this