Research Article
A Stacked BiLSTM Neural Network Based on Coattention Mechanism for Question Answering
Table 3
Experimental results of different baselines and our proposed model on Train-All data.
| | Idx | Model | MAP | MRR |
| | 1 | Probabilistic quasi-synchronous grammar [35] | 0.6029 | 0.6852 | | 2 | Tree edit models [2] | 0.6091 | 0.6917 | | 3 | Linear-chain CRF [17] | 0.6307 | 0.7477 | | 4 | LCLR [18] | 0.7092 | 0.7700 | | 5 | Bigram + count [38] | 0.7113 | 0.7846 | | 6 | Three-layer BiLSTM + BM25 [6] | 0.7134 | 0.7913 | | 7 | Convolutional deep neural networks [39] | 0.7459 | 0.8078 | | 8 | BiLSTM/CNN with attention [7] | 0.7111 | 0.8322 | | 9 | Attentive LSTM [1] | 0.7530 | 0.8300 | | 10 | BiLSTM encoder-decoder with step attention [8] | 0.7261 | 0.8018 | | 11 | BiLSTM | 0.6982 | 0.7764 | | 12 | Stacked BiLSTM | 0.7127 | 0.7893 | | 13 | BiLSTM with coattention | 0.7325 | 0.7962 | | 14 | Stacked BiLSTM with coattention | 0.7451 | 0.8114 | | 15 | Stacked BiLSTM with coattention (cosine + Euclidean) | 0.7613 | 0.8401 |
|
|