Lecture 8. Suffix trees for NLP — различия между версиями

Материал из Wiki - Факультет компьютерных наук
Перейти к: навигация, поиск
(Reference graph)
(Reference graph)
Строка 247: Строка 247:
 
[[Файл:L8 15.jpg|слева]]
 
[[Файл:L8 15.jpg|слева]]
  
<br> <br> <br> <br> <br> <br> <br> <br> <br> <br> <br> <br> <br> <br> <br>
+
<br> <br> <br> <br> <br> <br> <br> <br> <br> <br> <br> <br> <br> <br> <br> <br> <br>
  
 
== Profound filtering ==
 
== Profound filtering ==
  
 
== Implementation ==
 
== Implementation ==

Версия 00:22, 1 сентября 2015

Text (collection) representation model

Vector Space Models

Vector Space Models, VSM [Salton, Wong, Yang, 1975] : every text is a vector in a space of terms. Usually is used along with cosine similarity measure.

Pro:

  • Makes linear algebra operation and thus machine learning applicable
  • Is simple and based on human intuition

Contra:

  • The more the words are there in the the text, the higher is the cosinesimilarity
  • Word order is lost
  • Need some additional hacks to use bigrams and ngrams as terms

Applications:

  • Information retrieval
  • Topic, genre and maybe sentiment classification
  • Extractive summarization
  • Spam filtering

Generalized Vector Space Models

Generalized Vector Space Models, GVSM [Wong, Ziarko, Wong, 1985]: introduces a term to term correlation in the VSM.

Pro:

  • Is still easy
  • Synonyms or near synonyms are introduced into the model
  • Term to term correlation may be computed in several ways: co-occurrences in the collection under consideration, co-occurrences in a large corpus, thesaurus-based relations

Contra:

  • A lot of computations: if jVj = n (the number of terms), there are 2ndimensions in the GVSM
  • Word order is lost
  • Need some additional hacks to use bigrams and ngrams as terms

Applications: mainly information retrieval

Vector Space Model of Semantics

Vector Space Model of Semantics [Turney, Pantel, 2010]: terms are vector in the space of contexts (usually, also terms).

Pro:

  • Is still easy
  • Allows to find semantically similar words
  • Very loved by scientific community

Contra:

  • Word order is lost
  • Need some additional hacks to use bigrams and ngrams as terms

Applications:

  • Word similarity detection, clustering, classification
  • Automatic thesaurus generation
  • Word sense disambiguation
  • Context-sensitive spelling correction
  • Textual advertising

Language Model [Ponte, Croft, 1998]

Language Model [Ponte, Croft, 1998]: a probability distribution over sequences of words.

Unigram model: p(t 1 t 2 t 3 ) = p(t 1 )p(t 2 )p(t 3 )

Bigram model: p(t 1 t 2 t 3 ) = p(t 1 |BOS)p(t 2 |t 1 )p(t 3 |t 2 )p(EOS|t 3 )

L8 1.jpg






Language Model [Ponte, Croft, 1998]: a probability distribution over sequences of words.

Pro:

  • The word order is not lost
  • Makes Markov chain theory applicable

Contra:

  • The problem of unseen words
  • How to choose n?
  • Long relations between words or sentences (for example, anaphora) are lost

Applications:

  • Information retrieval
  • Text generation
  • Abstractive summarization
  • Speech recognition
  • Speech generation

To sum up:

  • VSM-like models: no word order, algebraic structures, classification of any kind, stemming and lemmatization for dimension reduction
  • LM-like models: partially word order, Markov chains, generation of text and speech
  • What if new word appears? What if there is a typo? What if stemming is not always correct? Use edit distance on words or stems or use symbol ngrams instead as terms or ...

Suffix tree

Annotated suffix tree (AST)

The suffix tree is a data structure used for storing of and searching for strings of characters and their fragments.

Suffix trees in NLP:

  • Text clustering [Zamir, Etzioni,1998]
  • Language model for MT [Kennington et al., 2012] and IR [Huang, Powers, 2008]
  • Feature generation for text classification [Zhang, Lee, 2006]

Suffix tree: Example from [Gus�led, 1998]

Annotated suffix tree

Annotated suffix tree (AST) [Pampapathi, Mirkin, Levene, 2006]:

An annotated suffix tree is a data structure used for computing and storing all fragments of the text and their frequencies. It is a rooted tree in which:

  • Every node corresponds to one character
  • Every node is labeled by the frequency of the text fragment encoded by the path from the root to the node.

L8 3.jpg

Annotated suffix tree construction

"MINING" has 6 suffixes: "MINING", "INING", "NING", "ING", "NG", "G"

Naive algorithm for AST construction

  • Start with "MINING", add it as a chain of nodes with frequencies equal to unity. Add "INING" and "NING" the same way.
  • When adding "ING", note that there is already a path from root that encodes / reads "I N". Increase the frequencies in this path by 1. Add "G" to the end of the path. When adding "NG", note "G" and react in the same way.
  • Add "G" as a single node with frequency equal to unity.

L8 4.jpg

Using Naive algorithm we add the second string "DINING". Complexity: both O(n^2) time and space, where n is the number of input strings. Frequency of the root node is the sum of the first level node frequencies.

Annotated suffix tree scoring

L8 5.jpg
L8 6.jpg


















































Applications of AST

  • Spam filtering [Pampapathi, Mirkin, Levene, 2006]
  • Text categorization and clustering [Chernyak, Chugunova, Mirkin, 2010]
  • Text summarization [Yakovlev, Chernyak, 2014]
  • Taxonomy refinement [Chernyak, Mirkin, 2013]
  • German compound splitting [Provatorova, Chernyak, 201x]
  • AST as a string kernel [Dubov, 201x]
  • Reference graph construction [Dubov, Chernyak, Mirkin, 201x]
  • Profound filtering
  • Duplicate detection

Spam filtering

  1. Construct two huge ASTs: spam AST and ham AST
  2. Introduce match permutation (swaping symbols in the match) to SCORE
  3. Test versus Naive Bayes classifier on a standart datasets (The Ling-Spam corpus, Spam Assassin public corpus, The BBKSpam04 corpus)
  4. Beat it (in terms of spam precision and spam recall)!

Unsupervised text categorization

Input 1: 5k+ abstracts from ACM journal and ACM CCS (English). Every abstract is annotated with some of the ACM CCS topics.

Input 2: 100k+ web page description from Yandex Categories (YaCa) and Yandex Categories (Russian). Every description belongs to a YaCa categories.

Note: We know right answers! We try to reproduce them computationally. Pure scientific joy. No business applications.

Task: different models of text collection representation, relevance measures and AST scoringrelevance() Score relevance(taxonomy topic) and relevance(YaCa category) to abstracts and descriptions. Score the taxonomy topics and YaCa categories in descending order. Calculate how many right answers are there. Use nDCG and MAP to evaluate the quality.

Text summarization

TextRank is an efficient algorithm for extractive text summarization. Let us construct graph, where nodes stand for single sentences and edges connect sequential sentences. Than we can use the VSM and cosine similarity to compute the similarity between sentences and run PageRank to get the most important sentences. Let us replace the cosine similarity with scoring common AST.

A common AST of two ASTs is a tree, that consists of all chains of nodes that coincide in the ASTs. The frequencies are computed as average. Scoring the common AST (i.e.scoring all the paths in the tree according to SCORE) suits as sentence similarity measure.

Input 1: DUC 2004 (a text collection for text summarization)

Input 2: home made collection of gazeta.ru papers.

Quality measures: precision

L8 7.jpg










Taxonomy refinement

No taxonomy of mathematics in Russian! Let us construct one. Still we have something in educational standards. We can use Wikipedia to refine it.

L8 8.jpg
L8 9.jpg
The refining scheme. Initial taxonomy topics are in rectangles, the Wikipedia categories and subcategories are in rounded rectangles, the Wikipedia articles are in the ovals, and the leaf descriptors are in the clouds.





























































  1. Specify the domain of taxonomy to be refined and set the frame of taxonomy manually.
  2. Download, from the Wikipedia, the category tree and articles from the domain under consideration.
  3. Clean the category subtree of irrelevant articles.
  4. Clean the category subtree of irrelevant subcategories.
  5. Assign the remaining Wikipedia categories to the taxonomy topics.
  6. Form the intermediate layers of the taxonomy
  7. Use Wikipedia articles in each of the added category nodes as its leaves.
  8. Extract relevant keywords from Wikipedia articles and use them as leaf descriptors.
L8 11.jpg
The fragment of refined PTMS taxonomy. Lower layers are shown.
The fragment of refined PTMS taxonomy. Lower layers are shown.





































































German compound splitting

Liebeskummer — heart-sickness — Lieb-es-kumm-er Bilderrahmen — picture frame — Bild-er-rahm-en Schmerzensgeld — compensation — Schmerz-ens-geld Gedankenfreiheit — freedom of thought — Gedank-en-frei-heit Schweineeisch — pork — Schwein-e-eisch Trinkgeld — tip — Trink-geld

To split a compound:

  • write complex linguistics rules
  • use machine learning algorithms
  • (hypothesis) construct an AST from a list of simple words, than score compounds to this tree

It is a problem to find a list of German words that does not contain compounds.

Reference graph

L8 14.jpg













Reference graph is a graph of association rules. Not interesting itself, but:

  • Dynamics
  • Link analysis
  • Visualization with different options
  • Data: Russian newspapers
  • Nodes: Key words and phrases
  • Edges: A ⇒ B: the key word or phrase B occurs with a higher probability if the key word or phrase A occurs in the same text Features: Easy extension to temporal case and possibility of well developed graph analysis methods

Reference graphs can be used as a tool of information discovery and search and as a tool for temporal analysis.

L8 15.jpg


















Profound filtering

Implementation