5G for Future Wireless Networks. First International Conference, 5GWN 2017, Beijing, China, April 21-23, 2017, Proceedings

Research Article

Tree-LSTM Guided Attention Pooling of DCNN for Semantic Sentence Modeling

Download
187 downloads
  • @INPROCEEDINGS{10.1007/978-3-319-72823-0_6,
        author={Liu Chen and Guangping Zeng and Qingchuan Zhang and Xingyu Chen},
        title={Tree-LSTM Guided Attention Pooling of DCNN for Semantic Sentence Modeling},
        proceedings={5G for Future Wireless Networks. First International Conference, 5GWN 2017, Beijing, China, April 21-23, 2017, Proceedings},
        proceedings_a={5GWN},
        year={2018},
        month={1},
        keywords={Dynamic Convolutional Neural Network (CNN) Tree-Structured Long-Short Term Memory (Tree-LSTM) Attention Pooling Semantic Sentence Modeling},
        doi={10.1007/978-3-319-72823-0_6}
    }
    
  • Liu Chen
    Guangping Zeng
    Qingchuan Zhang
    Xingyu Chen
    Year: 2018
    Tree-LSTM Guided Attention Pooling of DCNN for Semantic Sentence Modeling
    5GWN
    Springer
    DOI: 10.1007/978-3-319-72823-0_6
Liu Chen,*, Guangping Zeng,*, Qingchuan Zhang,*, Xingyu Chen,*
    *Contact email: chenliueve@163.com, zgp@ustb.edu.cn, zqc1982@126.com, cscserer@sina.com

    Abstract

    The ability to explicitly represent sentences is central to natural language processing. Convolutional neural network (CNN), recurrent neural network and recursive neural networks are mainstream architectures. We introduce a novel structure to combine the strength of them for semantic modelling of sentences. Sentence representations are generated by Dynamic CNN (DCNN, a variant of CNN). At pooling stage, attention pooling is adopted to capture most significant information with the guide of Tree-LSTM (a variant of Recurrent NN) sentence representations. Comprehensive information is extracted by the pooling scheme and the combination of the convolutional layer and the tree long-short term memory. We evaluate the model on sentiment classification task. Experiment results show that utilization of the given structures and combination of Tree-LSTM and DCNN outperforms both Tree-LSTM and DCNN and achieves outstanding performance.