representing symbolic linguistic structures for neural
play

Representing symbolic linguistic structures for neural NLP: methods - PowerPoint PPT Presentation

Representing symbolic linguistic structures for neural NLP: methods and applications Alexander Panchenko Assistant Professor for NLP Structure and goals of this talk Publishing in ACL and similar conferences, e.g. NAACL, EMNLP , CoNLL:


  1. Representing symbolic linguistic structures for neural NLP: methods and applications Alexander Panchenko Assistant Professor for NLP

  2. Structure and goals of this talk • Publishing in ACL and similar conferences, e.g. NAACL, EMNLP , CoNLL: • this is the top conference in NLP —> your work is visible • Topics of this talk (all are based on forthcoming publications at ACL’19 and associated workshops): • Encoding and using linguistic structures in neural NLP models • Argument Mining

  3. About myself: a decade of fun R&D in NLP • 2019-now: Skoltech , Assistant Professor in NLP , MOSCOW • 2017-2019: University of Hamburg , Postdoc in NLP , GERMANY • 2015-2017: TU Darmstadt , Postdoc in NLP , GERMANY • 2013-2015: Startup in SNA , Research Engineer in NLP , MOSCOW • 2009-2013: Université catholique de Louvain , PhD in Computational Linguistics, BELGIUM • 2008: Xerox Research Centre Europe , Research Intern, FRANCE • 2002-2008: Bauman Moscow State Technical University , Engineer in Information Systems, MOSCOW

  4. About myself: a decade of fun R&D in NLP • Publications in int’l conferences & journals: • ACL • EMNLP • EACL • ECIR • NLE • Best papers at Representation learning workshop (ACL’2016) and SemEval’2019. • Editor and co-chair : • Cambridge Natural Language Engineering (NLE) • Springer LNCS/CCIS: AIST conf. • PC : • ACL, NAACL, EMNLP , CoNLL, LREC, RANLP , COLING, …

  5. How to inform neural architectures for NLP with symbolic linguistic knowledge? • Special issue of the Natural Language Engineering journal on informing neural architectures for NLP with linguistic and background knowledge: https://sites.google.com/view/nlesi

  6. How to inform neural architectures for NLP with symbolic linguistic knowledge? Some options : • Graph embeddings • Poincaré embeddings • Regularisers that access the resource • Structure of neural network is based on the structure of the resource • … other specialised embeddings? • … invented by you? • Special issue of the Natural Language Engineering journal on informing neural architectures for NLP with linguistic and background knowledge: https://sites.google.com/view/nlesi

  7. Text: a sparse symbolic representation Image source: https://www.tensorflow.org/tutorials/word2vec

  8. Graph: a sparse symbolic representation

  9. Embedding graph into a vector space From a survey on graph embeddings [Hamilton et al., 2017]:

  10. Learning with an autoencoder From a survey on graph embeddings [Hamilton et al., 2017]:

  11. A summary of well-known graph embedding algorithms From a survey on graph embeddings [Hamilton et al., 2017]:

  12. Graph Metric Embeddings • A short paper at ACL 2019 • Paper : https://arxiv.org/abs/ 1906.07040 • Code : http://github.com/uhh- lt/path2vec

  13. path2vec model

  14. Computational gains compare to graph-based algorithms Similarity computation: graph vs vectors

  15. path2vec: evaluation results on three different graphs Evaluation on di ff erent graphs on SimLex999 (left) and shortest path distance (middle, right).

  16. path2vec evaluation inside a graph- based WSD algorithm (WordNet graph)

  17. Graph embeddings for neural entity linking • A short paper at ACL 2019 Student Research Workshop (main conference) • Paper: https://www.inf.uni- hamburg.de/en/inst/ab/lt/ publications/2019-sevgilietal- aclsrw-graphemb.pdf • Code : https://github.com/ uhh-lt/kb2vec

  18. What is Entity Linking? Source of image: https://medium.com/asgard-ai/how-to-enhance-automatic-text-analysis- with-entity-linking-29128a12b

  19. Challenges of Entity Linking Michael Jordan (NBA) vs Michael Jordan (LDA), etc. Ambiguity ruin everything: Source of image: https://medium.com/asgard-ai/how-to-enhance-automatic-text-analysis- with-entity-linking-29128a12b

  20. Graph embeddings for neural entity linking

  21. Graph embeddings for neural entity linking Architecture of our feed-forward neural ED system: using Wikipedia hyperlink graph embeddings as an additional input representation of entity candidates

  22. Graph embeddings for neural entity linking

  23. end2end Entity Linking Model by Kolistas et al. (2018) • The final score is used for both the mention linking and entity disambiguation decisions. • SOTA entity linking results.

  24. Graph embeddings for neural entity linking

  25. Poincaré embeddings for various NLP tasks • ACL 2019 full paper • Paper : https://www.inf.uni- hamburg.de/en/inst/ab/lt/ publications/2019-janaetal- aclmain-poincompo.pdf • Code : https://github.com/ uhh-lt/poincare

  26. Poincaré embeddings for various NLP tasks Contributions: • We devise a straightforward and e ffi cient approach for combining distributional and hypernymy information for the task of noun phrase compositionality prediction . As far as we are aware, this is the first application of Poincaré embeddings to this task. • We demonstrate consistent and significant improvements on benchmark datasets in un- supervised and supervised settings.

  27. Poincaré embeddings for various NLP tasks Image source: • Poincaré ball: https://arxiv.org/pdf/1705.08039.pdf • Distance on a ball between two points: •

  28. Poincaré embeddings for various NLP tasks Training objective: Training data: • A set of relations (apple IsA fruit) • Can be taken from WordNet • … or extracted from text Source of the image: https://arxiv.org/pdf/1902.00913.pdf

  29. Poincaré embeddings for noun compositionally hot dog —> food BUT dog —> animal green apple —> fruit AND apple —> fruit Evaluation results: comparison to the distributional models

  30. Noun compositionality for the Russian language • The Balto-Slavic NLP workshop at ACL 2019 • Paper: http://panchenko.me/ papers/bsnlp19.pdf • Code: https://github.com/ slangtech/ru-comps • A dataset for evaluation of noun compositionally for Russian.

  31. Noun Compositionality for Russian: Results

  32. Poincaré embeddings for taxonomy induction • A short paper at ACL 2019 • Paper : https://www.inf.uni- hamburg.de/en/inst/ab/lt/ publications/2019-alyetal- aclshort-hypertaxi.pdf • Code : https://github.com/ uhh-lt/ Taxonomy_Refinement_Embe ddings

  33. Abandoned children in a taxonomy problem Attaching unconnected nodes in taxonomy provides large boosts in performance:

  34. Poincaré embeddings for taxonomy induction Outline of our taxonomy refinement method:

  35. Poincaré embeddings for taxonomy induction

  36. Comparative Argument Mining • 6th Workshop on Argument Mining at ACL 2019 . • Paper : https://www.inf.uni- hamburg.de/en/inst/ab/lt/ publications/2019-panchenkoetal- argminingws-compsent.pdf • Code : https://github.com/uhh-lt/ comparative

  37. Comparative Argument Mining • Sentiment analysis ++ • … not only opinions but also objective arguments. • … from text only. • Retrieve pros and cons to make some informed decisions.

  38. Comparative Argument Mining • Sentiment analysis ++ • … not only opinions but also objective arguments. • … from text only. • Retrieve pros and cons to make some informed decisions.

  39. Categorizing Comparative Sentences Contributions: • We release a new dataset consisting of 7,199 sentences containing item pairs (27% of the sentences are tagged as comparative and annotated with a preference); • We present an experimental study of supervised classifiers and a strong rule-based baseline from prior work.

  40. Categorizing Comparative Sentences

  41. Categorizing Comparative Sentences

  42. Categorizing Comparative Sentences

  43. Argument Mining Demo • Demo paper at ACL 2019 • Paper : https://www.inf.uni- hamburg.de/en/inst/ab/lt/ publications/2019-chernodubetal- acl19demo-targer.pdf • Code : • A neural sequence tagger: https://github.com/uhh-lt/targer • A web application for AM: http:// github.com/achernodub/targer/ • Demo : http:// ltdemos.informatik.uni- hamburg.de/targer/

  44. Argument Mining Demo Tagger: • BiLSTM-CNN-CRF • A custom PyTorch implementation • CoNLL as input: can be easily used for any sequence labelling task

  45. Argument Mining Demo Analyze Text: input field, drop-down model selection, colorised labels, and tagged result. http://ltdemos.informatik.uni-hamburg.de/targer/

  46. Argument Mining Demo Search Arguments: query box, field selectors, and result with link to the original document. http://ltdemos.informatik.uni-hamburg.de/targer/

  47. Argument Mining Demo

  48. How to publish NLP research in conferences like *ACL? Select a relevant topic: • “Hot topic” may boost interest to your work • … but many people may be working on it at the same time (you need to be fast). • … especially if the idea is fairly straightforward extension of existing stu ff . Collaborate: • Find strong collaborators which already published in the conferences you are aiming at. • Ideally your competences should complement one another. • Splitting work into “experiments”, “writing”, “related work”, etc.

Download Presentation
Download Policy: The content available on the website is offered to you 'AS IS' for your personal information and use only. It cannot be commercialized, licensed, or distributed on other websites without prior consent from the author. To download a presentation, simply click this link. If you encounter any difficulties during the download process, it's possible that the publisher has removed the file from their server.

Recommend


More recommend