Browsing by Subject "Machine translation"
Now showing 1 - 2 of 2
- Results Per Page
- Sort Options
Item Experiments in non-factoid question answering(2013-08) Kulkarni, Sameer RajendraQuestion Answering (QA) is the task of generating or extracting an answer for a user query from a corpus of documents. Factoid QA is the most popular and studied form of QA and has received maximum focus from the scientific community. As is apparent from the name, the information requested from these factoid questions is a bare fact and in most cases is a named entity. In a majority of cases, such information is found in a single document and does not require sentence extraction and sentence reordering. However, most interesting questions are not factoid questions. Users might request a summary of a recent event from a news article, or they might want to know about a recent remedial cure for some observed symptoms that require text extraction from five different medical documents. All such queries require sentence extraction from a single or (often) multiple documents and require sentence reordering to generate a readable answer. This task is non-trivial, and hence there is more to non-factoid QA than meets the eye. Non-factoid QA has recently drawn attention from both the Information Retrieval (IR) and Natural Language Processing (NLP) communities, but most of the research has focused on developing learning models for re-ranking the answers from a set of question-answer pairs. This thesis explores the use of different natural language (NL) structures to complement the traditional bag-of-words model to generate answers for non-factoid questions. We find that complex linguistic features like semantic role labels outperform the traditional bag-of-words model. In fact, we find that the combination of different NL structures with the bag-of-words model performs best in our experiments. We also use Feature Engineering for extracting different sets of features from a given corpus. We find that using similarity features, translation features and occurrence features produces a higher ranked result as compared to the bag-of-words model and may help bridge the semantic gap between non-factoid questions and answers.Item An incremental syntactic language model for statistical phrase-based translation.(2012-02) Schwartz, Lane Oscar BingamanModern machine translation techniques typically incorporate both a translation model, which guides how individual words and phrases can be translated, and a language model (LM), which promotes fluency as translated words and phrases are combined into a translated sentence. Most attempts to inform the translation process with linguistic knowledge have focused on infusing syntax into translation models. We present a novel technique for incorporating syntactic knowledge as a language model in the context of statistical phrase-based machine translation (Koehn et al., 2003), one of the most widely used modern translation paradigms. The major contributions of this work are as follows: #15; We present a formal definition of an incremental syntactic language model as a Hierarchical Hidden Markov Model (HHMM), and detail how this model is estimated from a treebank corpus of labelled data. #15; The HHMM syntactic language model has been used in prior work involving parsing, speech recognition, and semantic role labelling. We present the first complete algorithmic definition of the HHMM as a language model. #15; We develop a novel and general method for incorporating any generative incremental language model into phrase-based machine translation. We integrate our HHMM incremental syntactic language model into Moses, the prevailing phrase-based decoder. #15; We present empirical results that demonstrate substantial improvements in perplexity for our syntactic language model over traditional n-gram language models; we also present empirical results on a constrained Urdu-English translation task that demonstrate the use of our syntactic LM.A standard measure of language model quality is average per-word perplexity. We present empirical results evaluating perplexity of various n-gram language models and our syntactic language model on both in-domain and out-of-domain test sets. On an in-domain test set, a traditional 5-gram language model trained on the same data as our syntactic language model outperforms the syntactic language model in terms of perplexity. We find that interpolating the 5-gram LM with the syntactic LM results in improved perplexity results, a 10% absolute reduction in perplexity compared to the 5-gram LM alone. On an out-of-domain test set, we find that our syntactic LM substantially outperforms all other LMs trained on the same training data. The syntactic LM demonstrates a 58% absolute reduction in perplexity over a 5-gram language model trained on the same training data. On this same out-of-domain test set, we further show that interpolating our syntactic language model with a large Gigaword-scale 5-gram language model results in the best overall perplexity results — a 61% absolute reduction in perplexity compared to the Gigaword-scale 5-gram language model alone, a 76% absolute reduction in perplexity compared to the syntactic LM alone, and a 90% absolute reduction in perplexity compared to the original smaller 5-gram language model. A language model with low perplexity is a theoretically good model of the language; it is expected that using an LM with low perplexity as a component of a machine translation system should result in more fluent translations. We present empirical results on a constrained Urdu-English translation task and perform an informal manual evaluation of translation results which suggests that the use of our incremental syntactic language model is indeed serving to guide the translation algorithm towards more fluent target language translations.