Shuailong Liang
2019
Subword Encoding in Lattice LSTM for Chinese Word SegmentationLSTM for Chinese Word Segmentation
Jie Yang
|
Yue Zhang
|
Shuailong Liang
Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 1 (Long and Short Papers)
We investigate subword information for Chinese word segmentation, by integrating sub word embeddings trained using byte-pair encoding into a Lattice LSTM (LaLSTM) network over a character sequence. Experiments on standard benchmark show that subword information brings significant gains over strong character-based segmentation models. To our knowledge, this is the first research on the effectiveness of subwords on neural word segmentation.
Does it Make Sense? And Why? A Pilot Study for Sense Making and Explanation
Cunxiang Wang
|
Shuailong Liang
|
Yue Zhang
|
Xiaonan Li
|
Tian Gao
Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics
Introducing common sense to natural language understanding systems has received increasing research attention. It remains a fundamental question on how to evaluate whether a system has the sense-making capability. Existing benchmarks measure common sense knowledge indirectly or without reasoning. In this paper, we release a benchmark to directly test whether a system can differentiate natural language statements that make sense from those that do not make sense. In addition, a system is asked to identify the most crucial reason why a statement does not make sense. We evaluate models trained over large-scale language modeling tasks as well as human performance, showing that there are different challenges for system sense-making.
2018
Design Challenges and Misconceptions in Neural Sequence Labeling
Jie Yang
|
Shuailong Liang
|
Yue Zhang
Proceedings of the 27th International Conference on Computational Linguistics
We investigate the design challenges of constructing effective and efficient neural sequence labeling systems, by reproducing twelve neural sequence labeling models, which include most of the state-of-the-art structures, and conduct a systematic model comparison on three benchmarks (i.e. NER, Chunking, and POS tagging). Misconceptions and inconsistent conclusions in existing literature are examined and clarified under statistical experiments. In the comparison and analysis process, we reach several practical conclusions which can be useful to practitioners.
Search