uni/prosem/prosem-ki.bib

192 lines
20 KiB
BibTeX
Raw Normal View History

2014-01-07 16:35:04 +01:00
% This file was created with JabRef 2.10b2.
% Encoding: UTF8
2014-01-07 16:35:04 +01:00
@InProceedings{Brin1998,
Title = {The Anatomy of a Large-Scale Hypertextual Web Search Engine},
Author = {Brin, Sergey and Page, Lawrence},
Booktitle = {Seventh World Wide Web Conference},
Year = {1998},
Keywords = {World Wide Web, Search Engines, Information Retrieval, PageRank, Google},
Owner = {jim},
Quality = {1},
Timestamp = {2013.10.29}
}
@Conference{Clark2004,
Title = {Parsing the {WSJ} using {CCG} and Log-Linear Models},
Author = {Clark, Stephen and Curran, James R.},
Booktitle = {Proceedings of the 42nd Annual Meeting of the Association for Computational Linguistics},
Year = {2004},
Pages = {104-111},
Owner = {jim},
Quality = {1},
Timestamp = {2013.10.29}
}
@Conference{Gnjatovic2012,
Title = {A {Cognitively-Inspired} {Method} for {Meaning Representation} in {Dialogue Systems}},
Author = {Gnjatović, Milan and Delić, Vlado},
Booktitle = {3rd IEEE International Conference on Cognitive Infocommunications},
Year = {2012},
Month = {December},
Pages = {383-388},
Owner = {jim},
Timestamp = {2014.01.18}
}
2014-01-07 16:35:04 +01:00
@InBook{Jurafsky2009,
Title = {Speech and Language Processing},
Author = {Jurafsky, Daniel and Martin, James H.},
Chapter = {18},
Pages = {617--644},
Publisher = {Pearson},
Year = {2009},
Edition = {Second},
Series = {Prentice-Hall series in artificial intelligence},
Abstract = {Sentences get their meanings from the words they contain and the syntactic order of the words. Therefore the meaning of a sentence is partially based on the words and its syntactic structure. The composition of meaning representation is guided by the syntactic components and relations provided by grammars such as CFGs. A meaning representation is generated by first sending the input through a parser which results in the syntactic analysis and second passing this analysis as input to a semantic analyzer. In the syntax-driven semantic analysis it is assumed that syntactic, lexical and anaphoric ambiguities are not a problem. The semantic meanings are attached to the grammar rules and lexical entries from which trees are generated in the first place. This is called rule-to-rule hypothesis. The semantic attachments are written in braces after the syntactic rules themselves. After the syntactic analysis has been created, every word receives a FOL predicate and/or term. The semantic analyzer goes the tree up until the complete FOL term has been created. On the way lambda reduction is used to replace predicates and terms with their proper meanings, received from other parts of the tree.},
2014-01-07 16:35:04 +01:00
Booktitle = {Speech and Language Processing},
Owner = {jim},
Quality = {1},
Timestamp = {2013.11.16}
}
2014-01-07 16:35:04 +01:00
@InBook{Jurafsky2009a,
Title = {Speech and Language Processing},
Author = {Jurafsky, Daniel and Martin, James H.},
Chapter = {17},
Pages = {579--616},
Publisher = {Pearson},
Year = {2009},
Edition = {Second},
Series = {Prentice-Hall series in artificial intelligence},
Abstract = {Lambda notation is used to bind variables dynamically to later appearing contents. lambda x P(x)(y) results in P(y) after a lambda reduction as x has been bound to y. lambda P P(x)(lambda x Restaurant(x)) results in lambda x Restaurant(x)(x) which results in Restaurant(x)},
2014-01-07 16:35:04 +01:00
Booktitle = {Speech and Language Processing},
Owner = {jim},
Quality = {1},
Timestamp = {2013.11.16}
2013-11-17 15:49:47 +01:00
}
2014-01-07 16:35:04 +01:00
@InBook{Jurafsky2009b,
Title = {Speech and Language Processing},
Author = {Jurafsky, Daniel and Martin, James H.},
Chapter = {13},
Pages = {461--492},
Publisher = {Pearson},
Year = {2009},
Edition = {Second},
Series = {Prentice-Hall series in artificial intelligence},
Owner = {jim},
Quality = {1},
Timestamp = {2013.11.17}
2013-11-17 15:49:47 +01:00
}
2014-01-07 16:35:04 +01:00
@Conference{Kessler1997,
Title = {Automatic Detection of Text Genre},
Author = {Kessler, Brett and Nunberg, Geoffrey and Schuetze, Hinrich},
Booktitle = {Proceedings of the 35th Annual Meeting of the Association for Computational Linguistics},
Year = {1997},
Pages = {32-38},
Owner = {jim},
Quality = {1},
Timestamp = {2013.10.29}
2013-11-17 15:49:47 +01:00
}
2014-01-07 16:35:04 +01:00
@Conference{Klein2003,
Title = {Named Entity Recognition with Character-Level Models},
Author = {Klein, Dan and Smarr, Joseph and Nguyen, Huy and Manning, Christopher D.},
Booktitle = {Conference on Natural Learning (CoNLL)},
Year = {2003},
Pages = {180-183},
Owner = {jim},
Quality = {1},
Timestamp = {2013.10.29}
}
2014-01-07 16:35:04 +01:00
@Article{Li,
Title = {A Comparison of {CYK} and {Earley} {Parsing} {Algorithms}},
Author = {Li, Te and Alagappan, Devi},
Institution = {Arizona State University},
Owner = {jim},
Timestamp = {2014.01.07}
}
2014-01-07 16:35:04 +01:00
@TechReport{Paskin2001,
Title = {Cubic-time Parsing and Learning Algorithms for Grammatical Bigram Models},
Author = {Paskin, Mark A.},
Institution = {University of California},
Year = {2001},
Month = {June},
Number = {UCB/CSD-01-1148},
Abstract = {In Dependency Grammar there are head words and dependents. Each phrase has only one head word. The head word determines how all of its dependents may be syntactically combined with other words to form a sentence. A head word and all of its dependents form a constituent. In every sentence there may be one or more dependency relationships with one head word each. Dependents that precede their head are called predependents and dependents that follow their head are called postdependents. A dependency parse consists of a set of dependency relationships that satisfies three constraints: 1. Every word except one (the root) is dependent to exactly one head. 2. The dependency relationships are acyclic; no word is, through a sequence of dependency relationships, dependent to itself. 3. When drawn as a graph above the sentence, no two dependency relations cross - a property known as projectivity or planarity. The Grammatical Bigram Probability Model assumes that all the dependents of a head word are independent of one another and their relative order. This is a strong approximation as in full English there are argument structure constraints that rely on the order of dependents. This simplification allows for a reduced computational complexity for parsing and learning. The grammar model falls into the class of "Bilexical grammars". A dependency parse consists of multiple spans. A span has at least two words up to n words. Spans have one property: No word in the span has a parent outside the span. Spans can be joined and closed. To join the span one of them has to be connected (both end words are connected with an edge) and both spans have to share one endword. The new span will be connected if both subspans were connected. If that is not the case, it can be closed by adding an edge between the endwords of the new span. Every dependency parse has a unique span decomposition. For joining the left subspan has be simple. That means it has to have an edge between its endwords or consist of two words only. Relying on this ensures that each span is derived only once. Every span has a signature. This signature states the indexes of its endwords, if it is simple and whether the left or right endword have parents within the span. Spans where both the left and right endword have the parent within the string are called toplevel signatures as such signatures characterize valid parses. Parser operations take signatures as input rather than spans. They produce signatures as well. SEED creates an unconnected and simple span with two adjacent words. CLOSE-LEFT adds an edge between the endwords and makes the left endword the parent of the right one. CLOSE-RIGHT does the opposite and makes the right endword the parent of the left one. These operators require that neither the left nor the right endword have a parent within the span. JOIN takes two input spans and joins them. It requires that the spans share an endword (1.), the shared endword has one parent (2.) and the left input is simple (3.). The JOIN rule applies only if the left span doesn't start the sentence. These operators constitute an algebra over span signatures called span signature algebra. A derivation D is an expression in this algebra. Like operations it evaluates to span signatures. These expressions can be represented as trees where the nodes are operations. There is an isomorphism between dependency parses and their corresponding derivations. Optimal derivation must consist of an operation over the results of optimal sub-derivations. Therefore it is enough to record the parse operation with the most likely derivation of a given signature in order to reconstruct the most likely derivation of the entire sentence. The chart-parse algorithm returns the optimal parse. It uses a subprocedure called EXTRACT-OPT-PARSE that constructs the optimal parse by finding the top-level signature (sigma) with maximum optimal probability (pi*). It backtracks then recursively through the optimal derivation defined by (omega*). If CLOSE operations are encountered edges are
2014-01-07 16:35:04 +01:00
Owner = {jim},
Quality = {1},
Timestamp = {2013.10.29}
}
@Misc{Portalarium2013,
Title = {Shroud of the {Avatar} {Six Month Progress Video}},
Author = {Portalarium},
HowPublished = {\url{https://www.youtube.com/watch?v=yGa6hR4a87U}},
Month = {November},
Note = {Accessed on 18.01.2014 12:07},
Year = {2013},
Owner = {jim},
Timestamp = {2014.01.12}
}
2014-01-07 16:35:04 +01:00
@InBook{Russel2010,
Title = {Artificial intelligence: A Modern Approach},
Author = {Russel, Stuart J. and Norvig, Peter},
Chapter = {23},
Pages = {888--927},
Publisher = {Pearson},
Year = {2009},
Edition = {Third},
Series = {Prentice-Hall series in artificial intelligence},
Abstract = {The first method to understanding natural language is syntactic analysis or parsing. The goal is to find the phrase structure of a sequence of words according to the rules of the applied grammar. A strict top-to-bottom or bottom-to-top parsing can be inefficient. Given two sentences with the same first 10 words and a difference only from the 11th word on, parsing from left-to-right would force the parser to make a guess about the nature of the sentence. But it doesn't know if it's right until the 11th word. From there it had to backtrack and reanalyze the sentence. To prevent that dynamic programming is used. Every analyzed substring gets stored for later. Once it is discovered that for example "the students in section 2 of Computer Science 101" is a noun phrase, this information can be stored in a structure known as chart. Algorithms that do such storing are called chart parsers. One of this chart parsers is a bottom-up version called CYK algorithm after its inventors John Cocke, Daniel Younger and Tadeo Kasami. This algorithm requires a grammar in the Chomsky Normal Form. The algorithm takes O(n²m) space for the P table with n being the number of words in the sentence and m the number of nonterminal symbols in the grammar. It takes O(n³m) time whereas m is constant for a particular grammar. That's why it is commonly described as O(n³). There is no faster algorithm for general context-free grammars. The CYK algorithm only co mputes the probability of the most probable tree. The subtrees are all represented in P table. PCFGs (Probabilistic context free grammars) have many rules with a probability for each one of them. Learning the grammar from data is better than a knowledge engineering approach. Learning is easiest if we are given a corpus of correctly parsed sentences; commonly known as a treebank. The best known treebank is the Penn Treebank as it consists of 3 million words which have been annotated with part of speech and parse-tree structure. Given an amount of trees, a PCFG can be created just by counting and smoothing. If no treebank is given it is still possible to learn the grammar but it is more difficult. In such a case there are actually two problems: First learning the structure of the grammar rules and second learning the probabilities associated with them. PCFGs have the problem that they are context-free. Combining a PCFG and Markov model will get the best of both. This leads ultimately to lexicalized PCFGs. But another problem of PCFGs is there preference for short sentences. Lexicalized PCFGs introduce so called head words. Such words are the most important words in a phrase and the probabilities are calculated between the head words. Example: "eat a banana" "eat" is the head of the verb phrase "eat a banana", whereas "banana" is the head of the noun phrase "a banana". Probability P1 now depends on "eat" and "banana" and the result would be very high. If the head of the noun phrase were "bandanna", the result would be significantly lower. The next step are definite clause grammars. They can be used to parse in a way of logical inference and makes it possible to reason about languages and strings in many different ways. Furthermore augmentations allow for distinctions in a single subphrase. For example the noun phrase (NP) depends on the subject case and the person and number of persons. A real world example would be "to smell". It is "I smell", "you smell", "we smell", "you smell" and "they smell" but "he/she/it smells". It depends on the person what version is taken. Semantic interpretation is used to give sentences a meaning. This is achieved through logical sentences. The semantics can be added to an already augmented grammar (created during the previous step), resulting in multiple augmentations at the same time. Chill is an inductive logic programming program that can learn to achieve 70% to 85% accuracy on various database query tasks. But there are several complications as English is endlessly complex. First there is the time at which things happened (present, p
2014-01-07 16:35:04 +01:00
Bookauthor = {Russel, Stuart J. and Norvig, Peter},
Booktitle = {Artificial intelligence: A Modern Approach},
Date = {December 11},
Owner = {jim},
Timestamp = {2013.10.24}
}
2014-01-07 16:35:04 +01:00
@InProceedings{Sleator1993,
Title = {Parsing English with a Link Grammar},
Author = {Sleator, Daniel D. K. and Temperley, Davy},
Booktitle = {Third Annual Workshop on Parsing technologies},
Year = {1993},
Owner = {jim},
Quality = {1},
Timestamp = {2013.10.29}
}
2014-01-07 16:35:04 +01:00
@Conference{Smith2008,
Title = {Dependency Parsing by Belief Propagation},
Author = {Smith, David A. and Eisner, Jason},
Booktitle = {Conference on Empirical Methods in Natural Language Processing},
Year = {2008},
Pages = {145-156},
Date = {October 25 - October 27},
Owner = {jim},
Quality = {1},
Timestamp = {2013.10.29}
}