Dependency-Based Convolutional Neural Networks for Sentence Embedding

Dependency-Based Convolutional Neural Networks for Sentence Embedding

Dependency-based Convolutional Neural Networks for Sentence Embedding ROOT What is Hawaii ’s state flower ? Mingbo Ma Liang Huang Bing Xiang Bowen Zhou CUNY=>OSU IBM T. J. Watson ACL 2015 Beijing Convolutional Neural Network for Vision “deep learning” Layer 1 16,000 CPU cores (on 1,000 machines) => 12 GPUs (on 3 machines) Google/Stanford (2012) Stanford/NVIDIA (2013) 2 Convolutional Neural Network for Go “deep learning” Google (Silver et al., Nature 2016) and Facebook (Tian and Zhu, ICLR 2016) 48 CPUs and 8 GPUs 16 CPUs and 4 GPUs 3 Convolutional Neural Network for Language Kalchbrenner et al. (2014) and Kim (2014) apply CNNs to sentence modeling • alleviates data sparsity by word embedding • sequential order (sentence) instead of spatial order (image) Should use more linguistic and structural information! (implementations based on Python-based deep learning toolkit “Theano”) 4 Sequential Convolution Sequential convolution What is Hawaii ’s state flower 1 2 3 4 5 6 word rep. convolution direction 5 Sequential Convolution Sequential convolution What is Hawaii ’s state flower 1 2 3 4 5 6 word rep. convolution direction 6 Sequential Convolution Sequential convolution What is Hawaii ’s state flower 1 2 3 4 5 6 word rep. convolution direction 7 Sequential Convolution Sequential convolution What is Hawaii ’s state flower 1 2 3 4 5 6 word rep. convolution direction 8 Sequential Convolution Sequential convolution What is Hawaii ’s state flower 1 2 3 4 5 6 word rep. convolution direction 9 Try different convolution filters and repeat the same process 10 Sequential Convolution Sequential convolution What is Hawaii ’s state flower 1 2 3 4 5 6 word rep. convolution direction 11 Sequential Convolution Sequential convolution What is Hawaii ’s state flower 1 2 3 4 5 6 word rep. convolution direction Max pooling 12 Sequential Convolution Sequential convolution What is Hawaii ’s state flower 1 2 3 4 5 6 word rep. convolution direction Max pooling Classification Feed into NN 13 Example: Question Type Classification (TREC) Sequential Convolution: Location What is Hawaii 's state flower ? Gold standard: Entity 14 Sequential Convolution Sequential convolution What is Hawaii ’s state flower 1 2 3 4 5 6 word rep. 15 Sequential Convolution Sequential convolution What is Hawaii ’s state flower 1 2 3 4 5 6 word rep. Loc Loc 16 Sequential Convolution Sequential convolution What is Hawaii ’s state flower 1 2 3 4 5 6 word rep. Loc Loc Loc Loc 17 Sequential Convolution Sequential convolution What is Hawaii ’s state flower 1 2 3 4 5 6 word rep. Loc Loc Loc Loc Enty 18 Convolution on Tree Sequential convolution ROOT word What is Hawaii ’s state flower rep. 1 2 3 4 5 6 19 Sequential Convolution Sequential convolution: • Traditional convolution operates in surface order • Cons: No structural information is captured No long distance relationships 20 Dependency-based Convolution Sequential convolution: • Traditional convolution operates in surface order • Cons: No structural information is captured No long distance relationships Structural Convolution: • operates the convolution filters on dependency tree • more “important” words are convolved more often • long distance relationships is naturally obtained 21 Convolution on Tree dependency convolution ROOT child parent What is Hawaii ’s state flower 1 2 3 4 5 6 word rep. convolution direction 22 Convolution on Tree dependency convolution ROOT child parent What is Hawaii ’s state flower 1 2 3 4 5 6 word rep. convolution direction 23 Convolution on Tree dependency convolution ROOT child parent What is Hawaii ’s state flower 1 2 3 4 5 6 word rep. convolution direction 24 Convolution on Tree dependency convolution ROOT child parent What is Hawaii ’s state flower 1 2 3 4 5 6 word rep. convolution direction 25 Convolution on Tree dependency convolution ROOT child parent What is Hawaii ’s state flower 1 2 3 4 5 6 word rep. convolution direction 26 Convolution on Tree dependency convolution ROOT child parent What is Hawaii ’s state flower 1 2 3 4 5 6 word rep. convolution direction 27 Convolution on Tree dependency convolution ROOT child parent What is Hawaii ’s state flower 1 2 3 4 5 6 word rep. convolution direction 28 Try different Bigram convolution filters and repeat the same process 29 Convolution on Tree dependency convolution ROOT child parent What is Hawaii ’s state flower 1 2 3 4 5 6 word rep. convolution direction 30 Convolution on Tree dependency convolution ROOT child parent What is Hawaii ’s state flower 1 2 3 4 5 6 word rep. convolution direction Max pooling 31 Convolution on Tree dependency convolution ROOT child parent What is Hawaii ’s state flower 1 2 3 4 5 6 word rep. convolution direction Max pooling 32 Convolution on Tree dependency convolution ROOT child parent What is Hawaii ’s state flower 1 2 3 4 5 6 word rep. convolution direction Max pooling 33 Convolution on Tree dependency convolution ROOT child parent What is Hawaii ’s state flower 1 2 3 4 5 6 word rep. convolution direction Max pooling 34 Convolution on Tree ROOT* ROOT** Trigram convolution child parent grand parent What is Hawaii ’s state flower 1 2 3 4 5 6 word rep. convolution direction more important words are convolved more often! 35 Convolution on Tree ROOT* ROOT** Trigram convolution child parent grand parent What is Hawaii ’s state flower 1 2 3 4 5 6 word rep. convolution direction Max pooling 36 Convolution on Tree ROOT What is Hawaii ’s state flower 1 2 3 4 5 6 bigram Fully connected NN with softmax output trigram 37 Experiment 1: Sentiment Classification Sentimental analysis from Rotten Tomatoes (MR & SST-1) straightforward statements: simplistic, silly and tedious Negative subtle statements: the film tunes into a grief that could lead a Positive man across centuries sentences with adversative: not for everyone, but for those with whom it Positive will connect, it's a nice departure from standard moviegoing fare 38 Sentimental Analysis Experiments Results Category Model MR SST-1 ancestor 80.4 47.7 This work ancestor+sibling 81.7 48.3 ancestor+sibling+sequential 81.9 49.5 CNNs-non-static (Kim ’14) — baseline 81.5 48.0 CNNs CNNs-multichannel (Kim ’14) 81.1 47.4 Deep CNNs (Kalchbrenner+ ’14) - 48.5 Recursive Autoencoder (Socher+ ’11) 77.7 43.2 Recursive NNs Recursive Neural Tensor (Socher+ ’13) - 45.7 Deep Recursive NNs (Irsoy+ ’14) - 49.8 Recurrent NNs LSTM on tree (Zhu+ ’15) 81.9 48.0 Other Paragraph-Vec (Le+ ’14) - 48.7 39 Experiment 2: Question Classificaiton Top-level Fine-grained Sentence (TREC) (TREC-2) How did serfdom develop in and then leave Russia? DESC manner What is Hawaii 's state flower ? ENTY plant What sprawling U.S. state boasts the most airports ? LOC state When was Algeria colonized ? NUM date What person 's head is on a dime ? HUM ind What does the technical term ISDN mean ? ABBR exp 40 Question Classification Experiments Results Category Model TREC TREC2 ancestor 95.4 88.4 This work ancestor+sibling 95.6 89.0 ancestor+sibling+sequential 95.4 88.8 CNNs-non-static (Kim ’14) — baseline 93.6 86.4 CNNs CNNs-multichannel (Kim ’14) 92.2 86.0 Deep CNNs (Kalchbrenner+ ’14) 93.0 - Hand-coded SVMs (Silva+ ’11)* 95.0 90.8 we achieved the highest published accuracy on TREC. 41 Error Analysis :-) Cases which we do better than Baseline: Gold/Ours: Enty Baseline: Loc Gold/Ours: Enty Baseline: Desc Gold/Ours: Desc Baseline: Enty Gold/Ours: Mild Neg Baseline: Mild Pos http://cogcomp.cs.illinois.edu/Data/QA/QC/definition.html 42 Error Analysis :-( Cases which we make mistakes: Gold: Num Ours: Enty Baseline: Num Cases which we and baseline make mistakes: Gold: Num Ours: Enty Baseline: Desc http://cogcomp.cs.illinois.edu/Data/QA/QC/definition.html 43 Conclusions Pros: Dependency-based convolution captures long- distance information. It outperforms sequential CNN in all four datasets. highest published accuracy on TREC. Cons: Our model’s accuracy depends on parser quality. 44 Deep Learning can and should be combined with linguistic intuitions. ROOT What is Hawaii ’s state flower ? 6JCPM[QW Backup slides Difference with Mou et al. (2015) Similarity: we both use dependency tree for convolution Difference: • Mou et al.(2015) only capture a head and all of its modifiers and they do not consider ancestor paths. • Their convolution depth on tree is always 2. • Words are convolved at most twice. • We consider their model a special case of our sibling model. Mou et al.(2015) our model include all siblings on ancestor paths siblings (partial) s m _ _ s m h t s m _ _ t s m h s m h g Discriminative Neural Sentence Modeling by Tree-Based Convolution (arXiv 2015 v5).

View Full Text

Details

  • File Type
    pdf
  • Upload Time
    -
  • Content Languages
    English
  • Upload User
    Anonymous/Not logged-in
  • File Pages
    47 Page
  • File Size
    -

Download

Channel Download Status
Express Download Enable

Copyright

We respect the copyrights and intellectual property rights of all users. All uploaded documents are either original works of the uploader or authorized works of the rightful owners.

  • Not to be reproduced or distributed without explicit permission.
  • Not used for commercial purposes outside of approved use cases.
  • Not used to infringe on the rights of the original creators.
  • If you believe any content infringes your copyright, please contact us immediately.

Support

For help with questions, suggestions, or problems, please contact us