Titles
stringlengths 6
220
| Abstracts
stringlengths 37
3.26k
| Years
int64 1.99k
2.02k
| Categories
stringclasses 1
value |
---|---|---|---|
Beyond Word-based Language Model in Statistical Machine Translation | Language model is one of the most important modules in statistical machine
translation and currently the word-based language model dominants this
community. However, many translation models (e.g. phrase-based models) generate
the target language sentences by rendering and compositing the phrases rather
than the words. Thus, it is much more reasonable to model dependency between
phrases, but few research work succeed in solving this problem. In this paper,
we tackle this problem by designing a novel phrase-based language model which
attempts to solve three key sub-problems: 1, how to define a phrase in language
model; 2, how to determine the phrase boundary in the large-scale monolingual
data in order to enlarge the training set; 3, how to alleviate the data
sparsity problem due to the huge vocabulary size of phrases. By carefully
handling these issues, the extensive experiments on Chinese-to-English
translation show that our phrase-based language model can significantly improve
the translation quality by up to +1.47 absolute BLEU score.
| 2,015 | Computation and Language |
Use of Modality and Negation in Semantically-Informed Syntactic MT | This paper describes the resource- and system-building efforts of an
eight-week Johns Hopkins University Human Language Technology Center of
Excellence Summer Camp for Applied Language Exploration (SCALE-2009) on
Semantically-Informed Machine Translation (SIMT). We describe a new
modality/negation (MN) annotation scheme, the creation of a (publicly
available) MN lexicon, and two automated MN taggers that we built using the
annotation scheme and lexicon. Our annotation scheme isolates three components
of modality and negation: a trigger (a word that conveys modality or negation),
a target (an action associated with modality or negation) and a holder (an
experiencer of modality). We describe how our MN lexicon was semi-automatically
produced and we demonstrate that a structure-based MN tagger results in
precision around 86% (depending on genre) for tagging of a standard LDC data
set.
We apply our MN annotation scheme to statistical machine translation using a
syntactic framework that supports the inclusion of semantic annotations.
Syntactic tags enriched with semantic annotations are assigned to parse trees
in the target-language training texts through a process of tree grafting. While
the focus of our work is modality and negation, the tree grafting procedure is
general and supports other types of semantic information. We exploit this
capability by including named entities, produced by a pre-existing tagger, in
addition to the MN elements produced by the taggers described in this paper.
The resulting system significantly outperformed a linguistically naive baseline
model (Hiero), and reached the highest scores yet reported on the NIST 2009
Urdu-English test set. This finding supports the hypothesis that both syntactic
and semantic information can improve translation quality.
| 2,012 | Computation and Language |
Monitoring Term Drift Based on Semantic Consistency in an Evolving
Vector Field | Based on the Aristotelian concept of potentiality vs. actuality allowing for
the study of energy and dynamics in language, we propose a field approach to
lexical analysis. Falling back on the distributional hypothesis to
statistically model word meaning, we used evolving fields as a metaphor to
express time-dependent changes in a vector space model by a combination of
random indexing and evolving self-organizing maps (ESOM). To monitor semantic
drifts within the observation period, an experiment was carried out on the term
space of a collection of 12.8 million Amazon book reviews. For evaluation, the
semantic consistency of ESOM term clusters was compared with their respective
neighbourhoods in WordNet, and contrasted with distances among term vectors by
random indexing. We found that at 0.05 level of significance, the terms in the
clusters showed a high level of semantic consistency. Tracking the drift of
distributional patterns in the term space across time periods, we found that
consistency decreased, but not at a statistically significant level. Our method
is highly scalable, with interpretations in philosophy.
| 2,015 | Computation and Language |
An investigation into language complexity of World-of-Warcraft
game-external texts | We present a language complexity analysis of World of Warcraft (WoW)
community texts, which we compare to texts from a general corpus of web
English. Results from several complexity types are presented, including lexical
diversity, density, readability and syntactic complexity. The language of WoW
texts is found to be comparable to the general corpus on some complexity
measures, yet more specialized on other measures. Our findings can be used by
educators willing to include game-related activities into school curricula.
| 2,015 | Computation and Language |
Boost Phrase-level Polarity Labelling with Review-level Sentiment
Classification | Sentiment analysis on user reviews helps to keep track of user reactions
towards products, and make advices to users about what to buy. State-of-the-art
review-level sentiment classification techniques could give pretty good
precisions of above 90%. However, current phrase-level sentiment analysis
approaches might only give sentiment polarity labelling precisions of around
70%~80%, which is far from satisfaction and restricts its application in many
practical tasks. In this paper, we focus on the problem of phrase-level
sentiment polarity labelling and attempt to bridge the gap between phrase-level
and review-level sentiment analysis. We investigate the inconsistency between
the numerical star ratings and the sentiment orientation of textual user
reviews. Although they have long been treated as identical, which serves as a
basic assumption in previous work, we find that this assumption is not
necessarily true. We further propose to leverage the results of review-level
sentiment classification to boost the performance of phrase-level polarity
labelling using a novel constrained convex optimization framework. Besides, the
framework is capable of integrating various kinds of information sources and
heuristics, while giving the global optimal solution due to its convexity.
Experimental results on both English and Chinese reviews show that our
framework achieves high labelling precisions of up to 89%, which is a
significant improvement from current approaches.
| 2,015 | Computation and Language |
Phrase-based Image Captioning | Generating a novel textual description of an image is an interesting problem
that connects computer vision and natural language processing. In this paper,
we present a simple model that is able to generate descriptive sentences given
a sample image. This model has a strong focus on the syntax of the
descriptions. We train a purely bilinear model that learns a metric between an
image representation (generated from a previously trained Convolutional Neural
Network) and phrases that are used to described them. The system is then able
to infer phrases from a given image sample. Based on caption syntax statistics,
we propose a simple language model that can produce relevant descriptions for a
given test image using the phrases inferred. Our approach, which is
considerably simpler than state-of-the-art models, achieves comparable results
in two popular datasets for the task: Flickr30k and the recently proposed
Microsoft COCO.
| 2,015 | Computation and Language |
Applying deep learning techniques on medical corpora from the World Wide
Web: a prototypical system and evaluation | BACKGROUND: The amount of biomedical literature is rapidly growing and it is
becoming increasingly difficult to keep manually curated knowledge bases and
ontologies up-to-date. In this study we applied the word2vec deep learning
toolkit to medical corpora to test its potential for identifying relationships
from unstructured text. We evaluated the efficiency of word2vec in identifying
properties of pharmaceuticals based on mid-sized, unstructured medical text
corpora available on the web. Properties included relationships to diseases
('may treat') or physiological processes ('has physiological effect'). We
compared the relationships identified by word2vec with manually curated
information from the National Drug File - Reference Terminology (NDF-RT)
ontology as a gold standard. RESULTS: Our results revealed a maximum accuracy
of 49.28% which suggests a limited ability of word2vec to capture linguistic
regularities on the collected medical corpora compared with other published
results. We were able to document the influence of different parameter settings
on result accuracy and found and unexpected trade-off between ranking quality
and accuracy. Pre-processing corpora to reduce syntactic variability proved to
be a good strategy for increasing the utility of the trained vector models.
CONCLUSIONS: Word2vec is a very efficient implementation for computing vector
representations and for its ability to identify relationships in textual data
without any prior domain knowledge. We found that the ranking and retrieved
results generated by word2vec were not of sufficient quality for automatic
population of knowledge bases and ontologies, but could serve as a starting
point for further manual curation.
| 2,015 | Computation and Language |
A new hybrid metric for verifying parallel corpora of Arabic-English | This paper discusses a new metric that has been applied to verify the quality
in translation between sentence pairs in parallel corpora of Arabic-English.
This metric combines two techniques, one based on sentence length and the other
based on compression code length. Experiments on sample test parallel
Arabic-English corpora indicate the combination of these two techniques
improves accuracy of the identification of satisfactory and unsatisfactory
sentence pairs compared to sentence length and compression code length alone.
The new method proposed in this research is effective at filtering noise and
reducing mis-translations resulting in greatly improved quality.
| 2,015 | Computation and Language |
Probabilistic Models for High-Order Projective Dependency Parsing | This paper presents generalized probabilistic models for high-order
projective dependency parsing and an algorithmic framework for learning these
statistical models involving dependency trees. Partition functions and
marginals for high-order dependency trees can be computed efficiently, by
adapting our algorithms which extend the inside-outside algorithm to
higher-order cases. To show the effectiveness of our algorithms, we perform
experiments on three languages---English, Chinese and Czech, using maximum
conditional likelihood estimation for model training and L-BFGS for parameter
estimation. Our methods achieve competitive performance for English, and
outperform all previously reported dependency parsers for Chinese and Czech.
| 2,015 | Computation and Language |
A Survey of Word Reordering in Statistical Machine Translation:
Computational Models and Language Phenomena | Word reordering is one of the most difficult aspects of statistical machine
translation (SMT), and an important factor of its quality and efficiency.
Despite the vast amount of research published to date, the interest of the
community in this problem has not decreased, and no single method appears to be
strongly dominant across language pairs. Instead, the choice of the optimal
approach for a new translation task still seems to be mostly driven by
empirical trials. To orientate the reader in this vast and complex research
area, we present a comprehensive survey of word reordering viewed as a
statistical modeling challenge and as a natural language phenomenon. The survey
describes in detail how word reordering is modeled within different
string-based and tree-based SMT frameworks and as a stand-alone task, including
systematic overviews of the literature in advanced reordering modeling. We then
question why some approaches are more successful than others in different
language pairs. We argue that, besides measuring the amount of reordering, it
is important to understand which kinds of reordering occur in a given language
pair. To this end, we conduct a qualitative analysis of word reordering
phenomena in a diverse sample of language pairs, based on a large collection of
linguistic knowledge. Empirical results in the SMT literature are shown to
support the hypothesis that a few linguistic facts can be very useful to
anticipate the reordering characteristics of a language pair and to select the
SMT framework that best suits them.
| 2,016 | Computation and Language |
Using NLP to measure democracy | This paper uses natural language processing to create the first machine-coded
democracy index, which I call Automated Democracy Scores (ADS). The ADS are
based on 42 million news articles from 6,043 different sources and cover all
independent countries in the 1993-2012 period. Unlike the democracy indices we
have today the ADS are replicable and have standard errors small enough to
actually distinguish between cases.
The ADS are produced with supervised learning. Three approaches are tried: a)
a combination of Latent Semantic Analysis and tree-based regression methods; b)
a combination of Latent Dirichlet Allocation and tree-based regression methods;
and c) the Wordscores algorithm. The Wordscores algorithm outperforms the
alternatives, so it is the one on which the ADS are based.
There is a web application where anyone can change the training set and see
how the results change: democracy-scores.org
| 2,015 | Computation and Language |
Deep Sentence Embedding Using Long Short-Term Memory Networks: Analysis
and Application to Information Retrieval | This paper develops a model that addresses sentence embedding, a hot topic in
current natural language processing research, using recurrent neural networks
with Long Short-Term Memory (LSTM) cells. Due to its ability to capture long
term memory, the LSTM-RNN accumulates increasingly richer information as it
goes through the sentence, and when it reaches the last word, the hidden layer
of the network provides a semantic representation of the whole sentence. In
this paper, the LSTM-RNN is trained in a weakly supervised manner on user
click-through data logged by a commercial web search engine. Visualization and
analysis are performed to understand how the embedding process works. The model
is found to automatically attenuate the unimportant words and detects the
salient keywords in the sentence. Furthermore, these detected keywords are
found to automatically activate different cells of the LSTM-RNN, where words
belonging to a similar topic activate the same cell. As a semantic
representation of the sentence, the embedding vector can be used in many
different applications. These automatic keyword detection and topic allocation
abilities enabled by the LSTM-RNN allow the network to perform document
retrieval, a difficult language processing task, where the similarity between
the query and documents can be measured by the distance between their
corresponding sentence embedding vectors computed by the LSTM-RNN. On a web
search task, the LSTM-RNN embedding is shown to significantly outperform
several existing state of the art methods. We emphasize that the proposed model
generates sentence embedding vectors that are specially useful for web document
retrieval tasks. A comparison with a well known general sentence embedding
method, the Paragraph Vector, is performed. The results show that the proposed
method in this paper significantly outperforms it for web document retrieval
task.
| 2,016 | Computation and Language |
Web-scale Surface and Syntactic n-gram Features for Dependency Parsing | We develop novel first- and second-order features for dependency parsing
based on the Google Syntactic Ngrams corpus, a collection of subtree counts of
parsed sentences from scanned books. We also extend previous work on surface
$n$-gram features from Web1T to the Google Books corpus and from first-order to
second-order, comparing and analysing performance over newswire and web
treebanks.
Surface and syntactic $n$-grams both produce substantial and complementary
gains in parsing accuracy across domains. Our best system combines the two
feature sets, achieving up to 0.8% absolute UAS improvements on newswire and
1.4% on web text.
| 2,015 | Computation and Language |
Breaking Sticks and Ambiguities with Adaptive Skip-gram | Recently proposed Skip-gram model is a powerful method for learning
high-dimensional word representations that capture rich semantic relationships
between words. However, Skip-gram as well as most prior work on learning word
representations does not take into account word ambiguity and maintain only
single representation per word. Although a number of Skip-gram modifications
were proposed to overcome this limitation and learn multi-prototype word
representations, they either require a known number of word meanings or learn
them using greedy heuristic approaches. In this paper we propose the Adaptive
Skip-gram model which is a nonparametric Bayesian extension of Skip-gram
capable to automatically learn the required number of representations for all
words at desired semantic resolution. We derive efficient online variational
learning algorithm for the model and empirically demonstrate its efficiency on
word-sense induction task.
| 2,015 | Computation and Language |
Rational Kernels for Arabic Stemming and Text Classification | In this paper, we address the problems of Arabic Text Classification and
stemming using Transducers and Rational Kernels. We introduce a new stemming
technique based on the use of Arabic patterns (Pattern Based Stemmer). Patterns
are modelled using transducers and stemming is done without depending on any
dictionary. Using transducers for stemming, documents are transformed into
finite state transducers. This document representation allows us to use and
explore rational kernels as a framework for Arabic Text Classification.
Stemming experiments are conducted on three word collections and classification
experiments are done on the Saudi Press Agency dataset. Results show that our
approach, when compared with other approaches, is promising specially in terms
of Accuracy, Recall and F1.
| 2,015 | Computation and Language |
Local Translation Prediction with Global Sentence Representation | Statistical machine translation models have made great progress in improving
the translation quality. However, the existing models predict the target
translation with only the source- and target-side local context information. In
practice, distinguishing good translations from bad ones does not only depend
on the local features, but also rely on the global sentence-level information.
In this paper, we explore the source-side global sentence-level features for
target-side local translation prediction. We propose a novel
bilingually-constrained chunk-based convolutional neural network to learn
sentence semantic representations. With the sentence-level feature
representation, we further design a feed-forward neural network to better
predict translations using both local and global information. The large-scale
experiments show that our method can obtain substantial improvements in
translation quality over the strong baseline: the hierarchical phrase-based
translation model augmented with the neural network joint model.
| 2,015 | Computation and Language |
Parsing as Reduction | We reduce phrase-representation parsing to dependency parsing. Our reduction
is grounded on a new intermediate representation, "head-ordered dependency
trees", shown to be isomorphic to constituent trees. By encoding order
information in the dependency labels, we show that any off-the-shelf, trainable
dependency parser can be used to produce constituents. When this parser is
non-projective, we can perform discontinuous parsing in a very natural manner.
Despite the simplicity of our approach, experiments show that the resulting
parsers are on par with strong baselines, such as the Berkeley parser for
English and the best single system in the SPMRL-2014 shared task. Results are
particularly striking for discontinuous parsing of German, where we surpass the
current state of the art by a wide margin.
| 2,015 | Computation and Language |
Improved Semantic Representations From Tree-Structured Long Short-Term
Memory Networks | Because of their superior ability to preserve sequence information over time,
Long Short-Term Memory (LSTM) networks, a type of recurrent neural network with
a more complex computational unit, have obtained strong results on a variety of
sequence modeling tasks. The only underlying LSTM structure that has been
explored so far is a linear chain. However, natural language exhibits syntactic
properties that would naturally combine words to phrases. We introduce the
Tree-LSTM, a generalization of LSTMs to tree-structured network topologies.
Tree-LSTMs outperform all existing systems and strong LSTM baselines on two
tasks: predicting the semantic relatedness of two sentences (SemEval 2014, Task
1) and sentiment classification (Stanford Sentiment Treebank).
| 2,015 | Computation and Language |
Task-Oriented Learning of Word Embeddings for Semantic Relation
Classification | We present a novel learning method for word embeddings designed for relation
classification. Our word embeddings are trained by predicting words between
noun pairs using lexical relation-specific features on a large unlabeled
corpus. This allows us to explicitly incorporate relation-specific information
into the word embeddings. The learned word embeddings are then used to
construct feature vectors for a relation classification model. On a
well-established semantic relation classification task, our method
significantly outperforms a baseline based on a previously introduced word
embedding method, and compares favorably to previous state-of-the-art models
that use syntactic information or manually constructed external resources.
| 2,015 | Computation and Language |
Non-linear Learning for Statistical Machine Translation | Modern statistical machine translation (SMT) systems usually use a linear
combination of features to model the quality of each translation hypothesis.
The linear combination assumes that all the features are in a linear
relationship and constrains that each feature interacts with the rest features
in an linear manner, which might limit the expressive power of the model and
lead to a under-fit model on the current data. In this paper, we propose a
non-linear modeling for the quality of translation hypotheses based on neural
networks, which allows more complex interaction between features. A learning
framework is presented for training the non-linear models. We also discuss
possible heuristics in designing the network structure which may improve the
non-linear learning performance. Experimental results show that with the basic
features of a hierarchical phrase-based machine translation system, our method
produce translations that are better than a linear model.
| 2,015 | Computation and Language |
The NLP Engine: A Universal Turing Machine for NLP | It is commonly accepted that machine translation is a more complex task than
part of speech tagging. But how much more complex? In this paper we make an
attempt to develop a general framework and methodology for computing the
informational and/or processing complexity of NLP applications and tasks. We
define a universal framework akin to a Turning Machine that attempts to fit
(most) NLP tasks into one paradigm. We calculate the complexities of various
NLP tasks using measures of Shannon Entropy, and compare `simple' ones such as
part of speech tagging to `complex' ones such as machine translation. This
paper provides a first, though far from perfect, attempt to quantify NLP tasks
under a uniform paradigm. We point out current deficiencies and suggest some
avenues for fruitful research.
| 2,015 | Computation and Language |
Variation of word frequencies in Russian literary texts | We study the variation of word frequencies in Russian literary texts. Our
findings indicate that the standard deviation of a word's frequency across
texts depends on its average frequency according to a power law with exponent
$0.62,$ showing that the rarer words have a relatively larger degree of
frequency volatility (i.e., "burstiness").
Several latent factors models have been estimated to investigate the
structure of the word frequency distribution. The dependence of a word's
frequency volatility on its average frequency can be explained by the asymmetry
in the distribution of latent factors.
| 2,016 | Computation and Language |
Bayesian Optimization of Text Representations | When applying machine learning to problems in NLP, there are many choices to
make about how to represent input texts. These choices can have a big effect on
performance, but they are often uninteresting to researchers or practitioners
who simply need a module that performs well. We propose an approach to
optimizing over this space of choices, formulating the problem as global
optimization. We apply a sequential model-based optimization technique and show
that our method makes standard linear models competitive with more
sophisticated, expensive state-of-the-art methods based on latent variable
models or neural networks on various topic classification and sentiment
analysis problems. Our approach is a first step towards black-box NLP systems
that work with raw text and do not require manual tuning.
| 2,015 | Computation and Language |
Robustly Leveraging Prior Knowledge in Text Classification | Prior knowledge has been shown very useful to address many natural language
processing tasks. Many approaches have been proposed to formalise a variety of
knowledge, however, whether the proposed approach is robust or sensitive to the
knowledge supplied to the model has rarely been discussed. In this paper, we
propose three regularization terms on top of generalized expectation criteria,
and conduct extensive experiments to justify the robustness of the proposed
methods. Experimental results demonstrate that our proposed methods obtain
remarkable improvements and are much more robust than baselines.
| 2,015 | Computation and Language |
Complexity and universality in the long-range order of words | As is the case of many signals produced by complex systems, language presents
a statistical structure that is balanced between order and disorder. Here we
review and extend recent results from quantitative characterisations of the
degree of order in linguistic sequences that give insights into two relevant
aspects of language: the presence of statistical universals in word ordering,
and the link between semantic information and the statistical linguistic
structure. We first analyse a measure of relative entropy that assesses how
much the ordering of words contributes to the overall statistical structure of
language. This measure presents an almost constant value close to 3.5 bits/word
across several linguistic families. Then, we show that a direct application of
information theory leads to an entropy measure that can quantify and extract
semantic structures from linguistic samples, even without prior knowledge of
the underlying language.
| 2,015 | Computation and Language |
Statistical modality tagging from rule-based annotations and
crowdsourcing | We explore training an automatic modality tagger. Modality is the attitude
that a speaker might have toward an event or state. One of the main hurdles for
training a linguistic tagger is gathering training data. This is particularly
problematic for training a tagger for modality because modality triggers are
sparse for the overwhelming majority of sentences. We investigate an approach
to automatically training a modality tagger where we first gathered sentences
based on a high-recall simple rule-based modality tagger and then provided
these sentences to Mechanical Turk annotators for further annotation. We used
the resulting set of training data to train a precise modality tagger using a
multi-class SVM that delivers good performance.
| 2,012 | Computation and Language |
What's Cookin'? Interpreting Cooking Videos using Text, Speech and
Vision | We present a novel method for aligning a sequence of instructions to a video
of someone carrying out a task. In particular, we focus on the cooking domain,
where the instructions correspond to the recipe. Our technique relies on an HMM
to align the recipe steps to the (automatically generated) speech transcript.
We then refine this alignment using a state-of-the-art visual food detector,
based on a deep convolutional neural network. We show that our technique
outperforms simpler techniques based on keyword spotting. It also enables
interesting applications, such as automatically illustrating recipes with
keyframes, and searching within a video for events of interest.
| 2,015 | Computation and Language |
Studying the Wikipedia Hyperlink Graph for Relatedness and
Disambiguation | Hyperlinks and other relations in Wikipedia are a extraordinary resource
which is still not fully understood. In this paper we study the different types
of links in Wikipedia, and contrast the use of the full graph with respect to
just direct links. We apply a well-known random walk algorithm on two tasks,
word relatedness and named-entity disambiguation. We show that using the full
graph is more effective than just direct links by a large margin, that
non-reciprocal links harm performance, and that there is no benefit from
categories and infoboxes, with coherent results on both tasks. We set new
state-of-the-art figures for systems based on Wikipedia links, comparable to
systems exploiting several information sources and/or supervised machine
learning. Our approach is open source, with instruction to reproduce results,
and amenable to be integrated with complementary text-based methods.
| 2,015 | Computation and Language |
Encoding Source Language with Convolutional Neural Network for Machine
Translation | The recently proposed neural network joint model (NNJM) (Devlin et al., 2014)
augments the n-gram target language model with a heuristically chosen source
context window, achieving state-of-the-art performance in SMT. In this paper,
we give a more systematic treatment by summarizing the relevant source
information through a convolutional architecture guided by the target
information. With different guiding signals during decoding, our specifically
designed convolution+gating architectures can pinpoint the parts of a source
sentence that are relevant to predicting a target word, and fuse them with the
context of entire source sentence to form a unified representation. This
representation, together with target language words, are fed to a deep neural
network (DNN) to form a stronger NNJM. Experiments on two NIST Chinese-English
translation tasks show that the proposed model can achieve significant
improvements over the previous NNJM by up to +1.08 BLEU points on average
| 2,015 | Computation and Language |
Identifying missing dictionary entries with frequency-conserving context
models | In an effort to better understand meaning from natural language texts, we
explore methods aimed at organizing lexical objects into contexts. A number of
these methods for organization fall into a family defined by word ordering.
Unlike demographic or spatial partitions of data, these collocation models are
of special importance for their universal applicability. While we are
interested here in text and have framed our treatment appropriately, our work
is potentially applicable to other areas of research (e.g., speech, genomics,
and mobility patterns) where one has ordered categorical data, (e.g., sounds,
genes, and locations). Our approach focuses on the phrase (whether word or
larger) as the primary meaning-bearing lexical unit and object of study. To do
so, we employ our previously developed framework for generating word-conserving
phrase-frequency data. Upon training our model with the Wiktionary---an
extensive, online, collaborative, and open-source dictionary that contains over
100,000 phrasal-definitions---we develop highly effective filters for the
identification of meaningful, missing phrase-entries. With our predictions we
then engage the editorial community of the Wiktionary and propose short lists
of potential missing entries for definition, developing a breakthrough, lexical
extraction technique, and expanding our knowledge of the defined English
lexicon of phrases.
| 2,015 | Computation and Language |
An Unsupervised Method for Uncovering Morphological Chains | Most state-of-the-art systems today produce morphological analysis based only
on orthographic patterns. In contrast, we propose a model for unsupervised
morphological analysis that integrates orthographic and semantic views of
words. We model word formation in terms of morphological chains, from base
words to the observed words, breaking the chains into parent-child relations.
We use log-linear models with morpheme and word-level features to predict
possible parents, including their modifications, for each word. The limited set
of candidate parents for each word render contrastive estimation feasible. Our
model consistently matches or outperforms five state-of-the-art systems on
Arabic, English and Turkish.
| 2,015 | Computation and Language |
Context-Dependent Translation Selection Using Convolutional Neural
Network | We propose a novel method for translation selection in statistical machine
translation, in which a convolutional neural network is employed to judge the
similarity between a phrase pair in two languages. The specifically designed
convolutional architecture encodes not only the semantic similarity of the
translation pair, but also the context containing the phrase in the source
language. Therefore, our approach is able to capture context-dependent semantic
similarities of translation pairs. We adopt a curriculum learning strategy to
train the model: we classify the training examples into easy, medium, and
difficult categories, and gradually build the ability of representing phrase
and sentence level context by using training examples from easy to difficult.
Experimental results show that our approach significantly outperforms the
baseline system by up to 1.4 BLEU points.
| 2,015 | Computation and Language |
Neural Responding Machine for Short-Text Conversation | We propose Neural Responding Machine (NRM), a neural network-based response
generator for Short-Text Conversation. NRM takes the general encoder-decoder
framework: it formalizes the generation of response as a decoding process based
on the latent representation of the input text, while both encoding and
decoding are realized with recurrent neural networks (RNN). The NRM is trained
with a large amount of one-round conversation data collected from a
microblogging service. Empirical study shows that NRM can generate
grammatically correct and content-wise appropriate responses to over 75% of the
input text, outperforming state-of-the-arts in the same setting, including
retrieval-based and SMT-based models.
| 2,015 | Computation and Language |
Syntax-based Deep Matching of Short Texts | Many tasks in natural language processing, ranging from machine translation
to question answering, can be reduced to the problem of matching two sentences
or more generally two short texts. We propose a new approach to the problem,
called Deep Match Tree (DeepMatch$_{tree}$), under a general setting. The
approach consists of two components, 1) a mining algorithm to discover patterns
for matching two short-texts, defined in the product space of dependency trees,
and 2) a deep neural network for matching short texts using the mined patterns,
as well as a learning algorithm to build the network having a sparse structure.
We test our algorithm on the problem of matching a tweet and a response in
social media, a hard matching problem proposed in [Wang et al., 2013], and show
that DeepMatch$_{tree}$ can outperform a number of competitor models including
one without using dependency trees and one based on word-embedding, all with
large margins
| 2,015 | Computation and Language |
Compositional Distributional Semantics with Long Short Term Memory | We are proposing an extension of the recursive neural network that makes use
of a variant of the long short-term memory architecture. The extension allows
information low in parse trees to be stored in a memory register (the `memory
cell') and used much later higher up in the parse tree. This provides a
solution to the vanishing gradient problem and allows the network to capture
long range dependencies. Experimental results show that our composition
outperformed the traditional neural-network composition on the Stanford
Sentiment Treebank.
| 2,015 | Computation and Language |
Convolutional Neural Network Architectures for Matching Natural Language
Sentences | Semantic matching is of central importance to many natural language tasks
\cite{bordes2014semantic,RetrievalQA}. A successful matching algorithm needs to
adequately model the internal structures of language objects and the
interaction between them. As a step toward this goal, we propose convolutional
neural network models for matching two sentences, by adapting the convolutional
strategy in vision and speech. The proposed models not only nicely represent
the hierarchical structures of sentences with their layer-by-layer composition
and pooling, but also capture the rich matching patterns at different levels.
Our models are rather generic, requiring no prior knowledge on language, and
can hence be applied to matching tasks of different nature and in different
languages. The empirical study on a variety of matching tasks demonstrates the
efficacy of the proposed model on a variety of matching tasks and its
superiority to competitor models.
| 2,015 | Computation and Language |
Is language evolution grinding to a halt? The scaling of lexical
turbulence in English fiction suggests it is not | Of basic interest is the quantification of the long term growth of a
language's lexicon as it develops to more completely cover both a culture's
communication requirements and knowledge space. Here, we explore the usage
dynamics of words in the English language as reflected by the Google Books 2012
English Fiction corpus. We critique an earlier method that found decreasing
birth and increasing death rates of words over the second half of the 20th
Century, showing death rates to be strongly affected by the imposed time cutoff
of the arbitrary present and not increasing dramatically. We provide a robust,
principled approach to examining lexical evolution by tracking the volume of
word flux across various relative frequency thresholds. We show that while the
overall statistical structure of the English language remains stable over time
in terms of its raw Zipf distribution, we find evidence of an enduring `lexical
turbulence': The flux of words across frequency thresholds from decade to
decade scales superlinearly with word rank and exhibits a scaling break we
connect to that of Zipf's law. To better understand the changing lexicon, we
examine the contributions to the Jensen-Shannon divergence of individual words
crossing frequency thresholds. We also find indications that scholarly works
about fiction are strongly represented in the 2012 English Fiction corpus, and
suggest that a future revision of the corpus should attempt to separate
critical works from fiction itself.
| 2,017 | Computation and Language |
On Using Monolingual Corpora in Neural Machine Translation | Recent work on end-to-end neural network-based architectures for machine
translation has shown promising results for En-Fr and En-De translation.
Arguably, one of the major factors behind this success has been the
availability of high quality parallel corpora. In this work, we investigate how
to leverage abundant monolingual corpora for neural machine translation.
Compared to a phrase-based and hierarchical baseline, we obtain up to $1.96$
BLEU improvement on the low-resource language pair Turkish-English, and $1.59$
BLEU on the focused domain task of Chinese-English chat messages. While our
method was initially targeted toward such tasks with less parallel data, we
show that it also extends to high resource languages such as Cs-En and De-En
where we obtain an improvement of $0.39$ and $0.47$ BLEU scores over the neural
machine translation baselines, respectively.
| 2,015 | Computation and Language |
An implementation of Apertium based Assamese morphological analyzer | Morphological Analysis is an important branch of linguistics for any Natural
Language Processing Technology. Morphology studies the word structure and
formation of word of a language. In current scenario of NLP research,
morphological analysis techniques have become more popular day by day. For
processing any language, morphology of the word should be first analyzed.
Assamese language contains very complex morphological structure. In our work we
have used Apertium based Finite-State-Transducers for developing morphological
analyzer for Assamese Language with some limited domain and we get 72.7%
accuracy
| 2,015 | Computation and Language |
Long Short-Term Memory Over Tree Structures | The chain-structured long short-term memory (LSTM) has showed to be effective
in a wide range of problems such as speech recognition and machine translation.
In this paper, we propose to extend it to tree structures, in which a memory
cell can reflect the history memories of multiple child cells or multiple
descendant cells in a recursive process. We call the model S-LSTM, which
provides a principled way of considering long-distance interaction over
hierarchies, e.g., language or image parse structures. We leverage the models
for semantic composition to understand the meaning of text, a fundamental
problem in natural language understanding, and show that it outperforms a
state-of-the-art recursive model by replacing its composition layers with the
S-LSTM memory blocks. We also show that utilizing the given structures is
helpful in achieving a performance better than that without considering the
structures.
| 2,015 | Computation and Language |
$gen$CNN: A Convolutional Architecture for Word Sequence Prediction | We propose a novel convolutional architecture, named $gen$CNN, for word
sequence prediction. Different from previous work on neural network-based
language modeling and generation (e.g., RNN or LSTM), we choose not to greedily
summarize the history of words as a fixed length vector. Instead, we use a
convolutional neural network to predict the next word with the history of words
of variable length. Also different from the existing feedforward networks for
language modeling, our model can effectively fuse the local correlation and
global correlation in the word sequence, with a convolution-gating strategy
specifically designed for the task. We argue that our model can give adequate
representation of the history, and therefore can naturally exploit both the
short and long range dependencies. Our model is fast, easy to train, and
readily parallelized. Our extensive experiments on text generation and $n$-best
re-ranking in machine translation show that $gen$CNN outperforms the
state-of-the-arts with big margins.
| 2,015 | Computation and Language |
Prediction Using Note Text: Synthetic Feature Creation with word2vec | word2vec affords a simple yet powerful approach of extracting quantitative
variables from unstructured textual data. Over half of healthcare data is
unstructured and therefore hard to model without involved expertise in data
engineering and natural language processing. word2vec can serve as a bridge to
quickly gather intelligence from such data sources.
In this study, we ran 650 megabytes of unstructured, medical chart notes from
the Providence Health & Services electronic medical record through word2vec. We
used two different approaches in creating predictive variables and tested them
on the risk of readmission for patients with COPD (Chronic Obstructive Lung
Disease). As a comparative benchmark, we ran the same test using the LACE risk
model (a single score based on length of stay, acuity, comorbid conditions, and
emergency department visits).
Using only free text and mathematical might, we found word2vec comparable to
LACE in predicting the risk of readmission of COPD patients.
| 2,015 | Computation and Language |
Text Segmentation based on Semantic Word Embeddings | We explore the use of semantic word embeddings in text segmentation
algorithms, including the C99 segmentation algorithm and new algorithms
inspired by the distributed word vector representation. By developing a general
framework for discussing a class of segmentation objectives, we study the
effectiveness of greedy versus exact optimization approaches and suggest a new
iterative refinement technique for improving the performance of greedy
strategies. We compare our results to known benchmarks, using known metrics. We
demonstrate state-of-the-art performance for an untrained method with our
Content Vector Segmentation (CVS) on the Choi test set. Finally, we apply the
segmentation procedure to an in-the-wild dataset consisting of text extracted
from scholarly articles in the arXiv.org database.
| 2,015 | Computation and Language |
Learning to Search for Dependencies | We demonstrate that a dependency parser can be built using a credit
assignment compiler which removes the burden of worrying about low-level
machine learning details from the parser implementation. The result is a simple
parser which robustly applies to many languages that provides similar
statistical and computational performance with best-to-date transition-based
parsing approaches, while avoiding various downsides including randomization,
extra feature requirements, and custom learning algorithms.
| 2,015 | Computation and Language |
Phrase database Approach to structural and semantic disambiguation in
English-Korean Machine Translation | In machine translation it is common phenomenon that machine-readable
dictionaries and standard parsing rules are not enough to ensure accuracy in
parsing and translating English phrases into Korean language, which is revealed
in misleading translation results due to consequent structural and semantic
ambiguities. This paper aims to suggest a solution to structural and semantic
ambiguities due to the idiomaticity and non-grammaticalness of phrases commonly
used in English language by applying bilingual phrase database in
English-Korean Machine Translation (EKMT). This paper firstly clarifies what
the phrase unit in EKMT is based on the definition of the English phrase,
secondly clarifies what kind of language unit can be the target of the phrase
database for EKMT, thirdly suggests a way to build the phrase database by
presenting the format of the phrase database with examples, and finally
discusses briefly the method to apply this bilingual phrase database to the
EKMT for structural and semantic disambiguation.
| 2,015 | Computation and Language |
Syntagma Lexical Database | This paper discusses the structure of Syntagma's Lexical Database (focused on
Italian). The basic database consists in four tables. Table Forms contains word
inflections, used by the POS-tagger for the identification of input-words.
Forms is related to Lemma. Table Lemma stores all kinds of grammatical features
of words, word-level semantic data and restrictions. In the table Meanings
meaning-related data are stored: definition, examples, domain, and semantic
information. Table Valency contains the argument structure of each meaning,
with syntactic and semantic features for each argument. The extended version of
SLD contains the links to Syntagma's Semantic Net and to the WordNet synsets of
other languages.
| 2,015 | Computation and Language |
On measuring linguistic intelligence | This work addresses the problem of measuring how many languages a person
"effectively" speaks given that some of the languages are close to each other.
In other words, to assign a meaningful number to her language portfolio.
Intuition says that someone who speaks fluently Spanish and Portuguese is
linguistically less proficient compared to someone who speaks fluently Spanish
and Chinese since it takes more effort for a native Spanish speaker to learn
Chinese than Portuguese. As the number of languages grows and their proficiency
levels vary, it gets even more complicated to assign a score to a language
portfolio. In this article we propose such a measure ("linguistic quotient" -
LQ) that can account for these effects.
We define the properties that such a measure should have. They are based on
the idea of coherent risk measures from the mathematical finance. Having laid
down the foundation, we propose one such a measure together with the algorithm
that works on languages classification tree as input.
The algorithm together with the input is available online at lingvometer.com
| 2,015 | Computation and Language |
Multilingual Open Relation Extraction Using Cross-lingual Projection | Open domain relation extraction systems identify relation and argument
phrases in a sentence without relying on any underlying schema. However,
current state-of-the-art relation extraction systems are available only for
English because of their heavy reliance on linguistic tools such as
part-of-speech taggers and dependency parsers. We present a cross-lingual
annotation projection method for language independent relation extraction. We
evaluate our method on a manually annotated test set and present results on
three typologically different languages. We release these manual annotations
and extracted relations in 61 languages from Wikipedia.
| 2,021 | Computation and Language |
Yara Parser: A Fast and Accurate Dependency Parser | Dependency parsers are among the most crucial tools in natural language
processing as they have many important applications in downstream tasks such as
information retrieval, machine translation and knowledge acquisition. We
introduce the Yara Parser, a fast and accurate open-source dependency parser
based on the arc-eager algorithm and beam search. It achieves an unlabeled
accuracy of 93.32 on the standard WSJ test set which ranks it among the top
dependency parsers. At its fastest, Yara can parse about 4000 sentences per
second when in greedy mode (1 beam). When optimizing for accuracy (using 64
beams and Brown cluster features), Yara can parse 45 sentences per second. The
parser can be trained on any syntactic dependency treebank and different
options are provided in order to make it more flexible and tunable for specific
tasks. It is released with the Apache version 2.0 license and can be used for
both commercial and academic purposes. The parser can be found at
https://github.com/yahoo/YaraParser.
| 2,015 | Computation and Language |
Unsupervised POS Induction with Word Embeddings | Unsupervised word embeddings have been shown to be valuable as features in
supervised learning problems; however, their role in unsupervised problems has
been less thoroughly explored. In this paper, we show that embeddings can
likewise add value to the problem of unsupervised POS induction. In two
representative models of POS induction, we replace multinomial distributions
over the vocabulary with multivariate Gaussian distributions over word
embeddings and observe consistent improvements in eight languages. We also
analyze the effect of various choices while inducing word embeddings on
"downstream" POS induction results.
| 2,015 | Computation and Language |
Morphological Analyzer and Generator for Russian and Ukrainian Languages | pymorphy2 is a morphological analyzer and generator for Russian and Ukrainian
languages. It uses large efficiently encoded lexi- cons built from OpenCorpora
and LanguageTool data. A set of linguistically motivated rules is developed to
enable morphological analysis and generation of out-of-vocabulary words
observed in real-world documents. For Russian pymorphy2 provides
state-of-the-arts morphological analysis quality. The analyzer is implemented
in Python programming language with optional C++ extensions. Emphasis is put on
ease of use, documentation and extensibility. The package is distributed under
a permissive open-source license, encouraging its use in both academic and
commercial setting.
| 2,015 | Computation and Language |
Using Latent Semantic Analysis to Identify Quality in Use (QU)
Indicators from User Reviews | The paper describes a novel approach to categorize users' reviews according
to the three Quality in Use (QU) indicators defined in ISO: effectiveness,
efficiency and freedom from risk. With the tremendous amount of reviews
published each day, there is a need to automatically summarize user reviews to
inform us if any of the software able to meet requirement of a company
according to the quality requirements. We implemented the method of Latent
Semantic Analysis (LSA) and its subspace to predict QU indicators. We build a
reduced dimensionality universal semantic space from Information System
journals and Amazon reviews. Next, we projected set of indicators' measurement
scales into the universal semantic space and represent them as subspace. In the
subspace, we can map similar measurement scales to the unseen reviews and
predict the QU indicators. Our preliminary study able to obtain the average of
F-measure, 0.3627.
| 2,015 | Computation and Language |
Unsupervised authorship attribution | We describe a technique for attributing parts of a written text to a set of
unknown authors. Nothing is assumed to be known a priori about the writing
styles of potential authors. We use multiple independent clusterings of an
input text to identify parts that are similar and dissimilar to one another. We
describe algorithms necessary to combine the multiple clusterings into a
meaningful output. We show results of the application of the technique on texts
having multiple writing styles.
| 2,015 | Computation and Language |
Normalization of Non-Standard Words in Croatian Texts | This paper presents text normalization which is an integral part of any
text-to-speech synthesis system. Text normalization is a set of methods with a
task to write non-standard words, like numbers, dates, times, abbreviations,
acronyms and the most common symbols, in their full expanded form are
presented. The whole taxonomy for classification of non-standard words in
Croatian language together with rule-based normalization methods combined with
a lookup dictionary are proposed. Achieved token rate for normalization of
Croatian texts is 95%, where 80% of expanded words are in correct morphological
form.
| 2,015 | Computation and Language |
Towards Using Machine Translation Techniques to Induce Multilingual
Lexica of Discourse Markers | Discourse markers are universal linguistic events subject to language
variation. Although an extensive literature has already reported language
specific traits of these events, little has been said on their cross-language
behavior and on building an inventory of multilingual lexica of discourse
markers. This work describes new methods and approaches for the description,
classification, and annotation of discourse markers in the specific domain of
the Europarl corpus. The study of discourse markers in the context of
translation is crucial due to the idiomatic nature of these structures.
Multilingual lexica together with the functional analysis of such structures
are useful tools for the hard task of translating discourse markers into
possible equivalents from one language to another. Using Daniel Marcu's
validated discourse markers for English, extracted from the Brown Corpus, our
purpose is to build multilingual lexica of discourse markers for other
languages, based on machine translation techniques. The major assumption in
this study is that the usage of a discourse marker is independent of the
language, i.e., the rhetorical function of a discourse marker in a sentence in
one language is equivalent to the rhetorical function of the same discourse
marker in another language.
| 2,015 | Computation and Language |
Learning to Understand Phrases by Embedding the Dictionary | Distributional models that learn rich semantic word representations are a
success story of recent NLP research. However, developing models that learn
useful representations of phrases and sentences has proved far harder. We
propose using the definitions found in everyday dictionaries as a means of
bridging this gap between lexical and phrasal semantics. Neural language
embedding models can be effectively trained to map dictionary definitions
(phrases) to (lexical) representations of the words defined by those
definitions. We present two applications of these architectures: "reverse
dictionaries" that return the name of a concept given a definition or
description and general-knowledge crossword question answerers. On both tasks,
neural language embedding models trained on definitions from a handful of
freely-available lexical resources perform as well or better than existing
commercial systems that rely on significant task-specific engineering. The
results highlight the effectiveness of both neural embedding architectures and
definition-based training for developing models that understand phrases and
sentences.
| 2,016 | Computation and Language |
A Unified Deep Neural Network for Speaker and Language Recognition | Learned feature representations and sub-phoneme posteriors from Deep Neural
Networks (DNNs) have been used separately to produce significant performance
gains for speaker and language recognition tasks. In this work we show how
these gains are possible using a single DNN for both speaker and language
recognition. The unified DNN approach is shown to yield substantial performance
improvements on the the 2013 Domain Adaptation Challenge speaker recognition
task (55% reduction in EER for the out-of-domain condition) and on the NIST
2011 Language Recognition Evaluation (48% reduction in EER for the 30s test
condition).
| 2,015 | Computation and Language |
Discriminative Neural Sentence Modeling by Tree-Based Convolution | This paper proposes a tree-based convolutional neural network (TBCNN) for
discriminative sentence modeling. Our models leverage either constituency trees
or dependency trees of sentences. The tree-based convolution process extracts
sentences' structural features, and these features are aggregated by max
pooling. Such architecture allows short propagation paths between the output
layer and underlying feature detectors, which enables effective structural
feature learning and extraction. We evaluate our models on two tasks: sentiment
analysis and question classification. In both experiments, TBCNN outperforms
previous state-of-the-art results, including existing neural networks and
dedicated feature/rule engineering. We also make efforts to visualize the
tree-based convolution process, shedding light on how our models work.
| 2,015 | Computation and Language |
Bengali to Assamese Statistical Machine Translation using Moses (Corpus
Based) | Machine dialect interpretation assumes a real part in encouraging man-machine
correspondence and in addition men-men correspondence in Natural Language
Processing (NLP). Machine Translation (MT) alludes to utilizing machine to
change one dialect to an alternate. Statistical Machine Translation is a type
of MT consisting of Language Model (LM), Translation Model (TM) and decoder. In
this paper, Bengali to Assamese Statistical Machine Translation Model has been
created by utilizing Moses. Other translation tools like IRSTLM for Language
Model and GIZA-PP-V1.0.7 for Translation model are utilized within this
framework which is accessible in Linux situations. The purpose of the LM is to
encourage fluent output and the purpose of TM is to encourage similarity
between input and output, the decoder increases the probability of translated
text in target language. A parallel corpus of 17100 sentences in Bengali and
Assamese has been utilized for preparing within this framework. Measurable MT
procedures have not so far been generally investigated for Indian dialects. It
might be intriguing to discover to what degree these models can help the
immense continuous MT deliberations in the nation.
| 2,015 | Computation and Language |
QUOTUS: The Structure of Political Media Coverage as Revealed by Quoting
Patterns | Given the extremely large pool of events and stories available, media outlets
need to focus on a subset of issues and aspects to convey to their audience.
Outlets are often accused of exhibiting a systematic bias in this selection
process, with different outlets portraying different versions of reality.
However, in the absence of objective measures and empirical evidence, the
direction and extent of systematicity remains widely disputed.
In this paper we propose a framework based on quoting patterns for
quantifying and characterizing the degree to which media outlets exhibit
systematic bias. We apply this framework to a massive dataset of news articles
spanning the six years of Obama's presidency and all of his speeches, and
reveal that a systematic pattern does indeed emerge from the outlet's quoting
behavior. Moreover, we show that this pattern can be successfully exploited in
an unsupervised prediction setting, to determine which new quotes an outlet
will select to broadcast. By encoding bias patterns in a low-rank space we
provide an analysis of the structure of political media coverage. This reveals
a latent media bias space that aligns surprisingly well with political ideology
and outlet type. A linguistic analysis exposes striking differences across
these latent dimensions, showing how the different types of media outlets
portray different realities even when reporting on the same events. For
example, outlets mapped to the mainstream conservative side of the latent space
focus on quotes that portray a presidential persona disproportionately
characterized by negativity.
| 2,015 | Computation and Language |
A Metric to Classify Style of Spoken Speech | The ability to classify spoken speech based on the style of speaking is an
important problem. With the advent of BPO's in recent times, specifically those
that cater to a population other than the local population, it has become
necessary for BPO's to identify people with certain style of speaking
(American, British etc). Today BPO's employ accent analysts to identify people
having the required style of speaking. This process while involving human bias,
it is becoming increasingly infeasible because of the high attrition rate in
the BPO industry. In this paper, we propose a new metric, which robustly and
accurately helps classify spoken speech based on the style of speaking. The
role of the proposed metric is substantiated by using it to classify real
speech data collected from over seventy different people working in a BPO. We
compare the performance of the metric against human experts who independently
carried out the classification process. Experimental results show that the
performance of the system using the novel metric performs better than two
different human expert.
| 2,015 | Computation and Language |
Voice based self help System: User Experience Vs Accuracy | In general, self help systems are being increasingly deployed by service
based industries because they are capable of delivering better customer service
and increasingly the switch is to voice based self help systems because they
provide a natural interface for a human to interact with a machine. A speech
based self help system ideally needs a speech recognition engine to convert
spoken speech to text and in addition a language processing engine to take care
of any misrecognitions by the speech recognition engine. Any off-the-shelf
speech recognition engine is generally a combination of acoustic processing and
speech grammar. While this is the norm, we believe that ideally a speech
recognition application should have in addition to a speech recognition engine
a separate language processing engine to give the system better performance. In
this paper, we discuss ways in which the speech recognition engine and the
language processing engine can be combined to give a better user experience.
| 2,015 | Computation and Language |
Jointly Embedding Relations and Mentions for Knowledge Population | This paper contributes a joint embedding model for predicting relations
between a pair of entities in the scenario of relation inference. It differs
from most stand-alone approaches which separately operate on either knowledge
bases or free texts. The proposed model simultaneously learns low-dimensional
vector representations for both triplets in knowledge repositories and the
mentions of relations in free texts, so that we can leverage the evidence both
resources to make more accurate predictions. We use NELL to evaluate the
performance of our approach, compared with cutting-edge methods. Results of
extensive experiments show that our model achieves significant improvement on
relation extraction.
| 2,015 | Computation and Language |
Mining and discovering biographical information in Difangzhi with a
language-model-based approach | We present results of expanding the contents of the China Biographical
Database by text mining historical local gazetteers, difangzhi. The goal of the
database is to see how people are connected together, through kinship, social
connections, and the places and offices in which they served. The gazetteers
are the single most important collection of names and offices covering the Song
through Qing periods. Although we begin with local officials we shall
eventually include lists of local examination candidates, people from the
locality who served in government, and notable local figures with biographies.
The more data we collect the more connections emerge. The value of doing
systematic text mining work is that we can identify relevant connections that
are either directly informative or can become useful without deep historical
research. Academia Sinica is developing a name database for officials in the
central governments of the Ming and Qing dynasties.
| 2,015 | Computation and Language |
Exploring Lexical, Syntactic, and Semantic Features for Chinese Textual
Entailment in NTCIR RITE Evaluation Tasks | We computed linguistic information at the lexical, syntactic, and semantic
levels for Recognizing Inference in Text (RITE) tasks for both traditional and
simplified Chinese in NTCIR-9 and NTCIR-10. Techniques for syntactic parsing,
named-entity recognition, and near synonym recognition were employed, and
features like counts of common words, statement lengths, negation words, and
antonyms were considered to judge the entailment relationships of two
statements, while we explored both heuristics-based functions and
machine-learning approaches. The reported systems showed robustness by
simultaneously achieving second positions in the binary-classification subtasks
for both simplified and traditional Chinese in NTCIR-10 RITE-2. We conducted
more experiments with the test data of NTCIR-9 RITE, with good results. We also
extended our work to search for better configurations of our classifiers and
investigated contributions of individual features. This extended work showed
interesting results and should encourage further discussion.
| 2,015 | Computation and Language |
Concentric network symmetry grasps authors' styles in word adjacency
networks | Several characteristics of written texts have been inferred from statistical
analysis derived from networked models. Even though many network measurements
have been adapted to study textual properties at several levels of complexity,
some textual aspects have been disregarded. In this paper, we study the
symmetry of word adjacency networks, a well-known representation of text as a
graph. A statistical analysis of the symmetry distribution performed in several
novels showed that most of the words do not display symmetric patterns of
connectivity. More specifically, the merged symmetry displayed a distribution
similar to the ubiquitous power-law distribution. Our experiments also revealed
that the studied metrics do not correlate with other traditional network
measurements, such as the degree or betweenness centrality. The effectiveness
of the symmetry measurements was verified in the authorship attribution task.
Interestingly, we found that specific authors prefer particular types of
symmetric motifs. As a consequence, the authorship of books could be accurately
identified in 82.5% of the cases, in a dataset comprising books written by 8
authors. Because the proposed measurements for text analysis are complementary
to the traditional approach, they can be used to improve the characterization
of text networks, which might be useful for related applications, such as those
relying on the identification of topical words and information retrieval.
| 2,015 | Computation and Language |
Leveraging Twitter for Low-Resource Conversational Speech Language
Modeling | In applications involving conversational speech, data sparsity is a limiting
factor in building a better language model. We propose a simple,
language-independent method to quickly harvest large amounts of data from
Twitter to supplement a smaller training set that is more closely matched to
the domain. The techniques lead to a significant reduction in perplexity on
four low-resource languages even though the presence on Twitter of these
languages is relatively small. We also find that the Twitter text is more
useful for learning word classes than the in-domain text and that use of these
word classes leads to further reductions in perplexity. Additionally, we
introduce a method of using social and textual information to prioritize the
download queue during the Twitter crawling. This maximizes the amount of useful
data that can be collected, impacting both perplexity and vocabulary coverage.
| 2,015 | Computation and Language |
Temporal ordering of clinical events | This report describes a minimalistic set of methods engineered to anchor
clinical events onto a temporal space. Specifically, we describe methods to
extract clinical events (e.g., Problems, Treatments and Tests), temporal
expressions (i.e., time, date, duration, and frequency), and temporal links
(e.g., Before, After, Overlap) between events and temporal entities. These
methods are developed and validated using high quality datasets.
| 2,015 | Computation and Language |
Unsupervised Dependency Parsing: Let's Use Supervised Parsers | We present a self-training approach to unsupervised dependency parsing that
reuses existing supervised and unsupervised parsing algorithms. Our approach,
called `iterated reranking' (IR), starts with dependency trees generated by an
unsupervised parser, and iteratively improves these trees using the richer
probability models used in supervised parsing that are in turn trained on these
trees. Our system achieves 1.8% accuracy higher than the state-of-the-part
parser of Spitkovsky et al. (2013) on the WSJ corpus.
| 2,015 | Computation and Language |
Gap Analysis of Natural Language Processing Systems with respect to
Linguistic Modality | Modality is one of the important components of grammar in linguistics. It
lets speaker to express attitude towards, or give assessment or potentiality of
state of affairs. It implies different senses and thus has different
perceptions as per the context. This paper presents an account showing the gap
in the functionality of the current state of art Natural Language Processing
(NLP) systems. The contextual nature of linguistic modality is studied. In this
paper, the works and logical approaches employed by Natural Language Processing
systems dealing with modality are reviewed. It sees human cognition and
intelligence as multi-layered approach that can be implemented by intelligent
systems for learning. Lastly, current flow of research going on within this
field is talked providing futurology.
| 2,015 | Computation and Language |
A Knowledge-poor Pronoun Resolution System for Turkish | A pronoun resolution system which requires limited syntactic knowledge to
identify the antecedents of personal and reflexive pronouns in Turkish is
presented. As in its counterparts for languages like English, Spanish and
French, the core of the system is the constraints and preferences determined
empirically. In the evaluation phase, it performed considerably better than the
baseline algorithm used for comparison. The system is significant for its being
the first fully specified knowledge-poor computational framework for pronoun
resolution in Turkish where Turkish possesses different structural properties
from the languages for which knowledge-poor systems had been developed.
| 2,015 | Computation and Language |
Online Inference for Relation Extraction with a Reduced Feature Set | Access to web-scale corpora is gradually bringing robust automatic knowledge
base creation and extension within reach. To exploit these large
unannotated---and extremely difficult to annotate---corpora, unsupervised
machine learning methods are required. Probabilistic models of text have
recently found some success as such a tool, but scalability remains an obstacle
in their application, with standard approaches relying on sampling schemes that
are known to be difficult to scale. In this report, we therefore present an
empirical assessment of the sublinear time sparse stochastic variational
inference (SSVI) scheme applied to RelLDA. We demonstrate that online inference
leads to relatively strong qualitative results but also identify some of its
pathologies---and those of the model---which will need to be overcome if SSVI
is to be used for large-scale relation extraction.
| 2,015 | Computation and Language |
Self-Adaptive Hierarchical Sentence Model | The ability to accurately model a sentence at varying stages (e.g.,
word-phrase-sentence) plays a central role in natural language processing. As
an effort towards this goal we propose a self-adaptive hierarchical sentence
model (AdaSent). AdaSent effectively forms a hierarchy of representations from
words to phrases and then to sentences through recursive gated local
composition of adjacent segments. We design a competitive mechanism (through
gating networks) to allow the representations of the same sentence to be
engaged in a particular learning task (e.g., classification), therefore
effectively mitigating the gradient vanishing problem persistent in other
recursive models. Both qualitative and quantitative analysis shows that AdaSent
can automatically form and select the representations suitable for the task at
hand during training, yielding superior classification performance over
competitor models on 5 benchmark data sets.
| 2,015 | Computation and Language |
Big Data Small Data, In Domain Out-of Domain, Known Word Unknown Word:
The Impact of Word Representation on Sequence Labelling Tasks | Word embeddings -- distributed word representations that can be learned from
unlabelled data -- have been shown to have high utility in many natural
language processing applications. In this paper, we perform an extrinsic
evaluation of five popular word embedding methods in the context of four
sequence labelling tasks: POS-tagging, syntactic chunking, NER and MWE
identification. A particular focus of the paper is analysing the effects of
task-based updating of word representations. We show that when using word
embeddings as features, as few as several hundred training instances are
sufficient to achieve competitive results, and that word embeddings lead to
improvements over OOV words and out of domain. Perhaps more surprisingly, our
results indicate there is little difference between the different word
embedding methods, and that simple Brown clusters are often competitive with
word embeddings across all tasks we consider.
| 2,015 | Computation and Language |
A Hierarchical Distance-dependent Bayesian Model for Event Coreference
Resolution | We present a novel hierarchical distance-dependent Bayesian model for event
coreference resolution. While existing generative models for event coreference
resolution are completely unsupervised, our model allows for the incorporation
of pairwise distances between event mentions -- information that is widely used
in supervised coreference models to guide the generative clustering processing
for better event clustering both within and across documents. We model the
distances between event mentions using a feature-rich learnable distance
function and encode them as Bayesian priors for nonparametric clustering.
Experiments on the ECB+ corpus show that our model outperforms state-of-the-art
methods for both within- and cross-document event coreference resolution.
| 2,015 | Computation and Language |
x.ent: R Package for Entities and Relations Extraction based on
Unsupervised Learning and Document Structure | Relation extraction with accurate precision is still a challenge when
processing full text databases. We propose an approach based on cooccurrence
analysis in each document for which we used document organization to improve
accuracy of relation extraction. This approach is implemented in a R package
called \emph{x.ent}. Another facet of extraction relies on use of extracted
relation into a querying system for expert end-users. Two datasets had been
used. One of them gets interest from specialists of epidemiology in plant
health. For this dataset usage is dedicated to plant-disease exploration
through agricultural information news. An open-data platform exploits exports
from \emph{x.ent} and is publicly available.
| 2,015 | Computation and Language |
On the Stability of Online Language Features: How Much Text do you Need
to know a Person? | In recent years, numerous studies have inferred personality and other traits
from people's online writing. While these studies are encouraging, more
information is needed in order to use these techniques with confidence. How do
linguistic features vary across different online media, and how much text is
required to have a representative sample for a person? In this paper, we
examine several large sets of online, user-generated text, drawn from Twitter,
email, blogs, and online discussion forums. We examine and compare
population-wide results for the linguistic measure LIWC, and the inferred
traits of Big5 Personality and Basic Human Values. We also empirically measure
the stability of these traits across different sized samples for each
individual. Our results highlight the importance of tuning models to each
online medium, and include guidelines for the minimum amount of text required
for a representative result.
| 2,015 | Computation and Language |
Classifying Relations by Ranking with Convolutional Neural Networks | Relation classification is an important semantic processing task for which
state-ofthe-art systems still rely on costly handcrafted features. In this work
we tackle the relation classification task using a convolutional neural network
that performs classification by ranking (CR-CNN). We propose a new pairwise
ranking loss function that makes it easy to reduce the impact of artificial
classes. We perform experiments using the the SemEval-2010 Task 8 dataset,
which is designed for the task of classifying the relationship between two
nominals marked in a sentence. Using CRCNN, we outperform the state-of-the-art
for this dataset and achieve a F1 of 84.1 without using any costly handcrafted
features. Additionally, our experimental results show that: (1) our approach is
more effective than CNN followed by a softmax classifier; (2) omitting the
representation of the artificial class Other improves both precision and
recall; and (3) using only word embeddings as input features is enough to
achieve state-of-the-art results if we consider only the text between the two
target nominals.
| 2,015 | Computation and Language |
Learning Dictionaries for Named Entity Recognition using Minimal
Supervision | This paper describes an approach for automatic construction of dictionaries
for Named Entity Recognition (NER) using large amounts of unlabeled data and a
few seed examples. We use Canonical Correlation Analysis (CCA) to obtain lower
dimensional embeddings (representations) for candidate phrases and classify
these phrases using a small number of labeled examples. Our method achieves
16.5% and 11.3% F-1 score improvement over co-training on disease and virus NER
respectively. We also show that by adding candidate phrase embeddings as
features in a sequence tagger gives better performance compared to using word
embeddings.
| 2,015 | Computation and Language |
Efficient Non-parametric Estimation of Multiple Embeddings per Word in
Vector Space | There is rising interest in vector-space word embeddings and their use in
NLP, especially given recent methods for their fast estimation at very large
scale. Nearly all this work, however, assumes a single vector per word type
ignoring polysemy and thus jeopardizing their usefulness for downstream tasks.
We present an extension to the Skip-gram model that efficiently learns multiple
embeddings per word type. It differs from recent related work by jointly
performing word sense discrimination and embedding learning, by
non-parametrically estimating the number of senses per word type, and by its
efficiency and scalability. We present new state-of-the-art results in the word
similarity in context task and demonstrate its scalability by training with one
machine on a corpus of nearly 1 billion tokens in less than 6 hours.
| 2,015 | Computation and Language |
Inferring Missing Entity Type Instances for Knowledge Base Completion:
New Dataset and Methods | Most of previous work in knowledge base (KB) completion has focused on the
problem of relation extraction. In this work, we focus on the task of inferring
missing entity type instances in a KB, a fundamental task for KB competition
yet receives little attention. Due to the novelty of this task, we construct a
large-scale dataset and design an automatic evaluation methodology. Our
knowledge base completion method uses information within the existing KB and
external information from Wikipedia. We show that individual methods trained
with a global objective that considers unobserved cells from both the entity
and the type side gives consistently higher quality predictions compared to
baseline methods. We also perform manual evaluation on a small subset of the
data to verify the effectiveness of our knowledge base completion methods and
the correctness of our proposed automatic evaluation method.
| 2,015 | Computation and Language |
Compositional Vector Space Models for Knowledge Base Completion | Knowledge base (KB) completion adds new facts to a KB by making inferences
from existing facts, for example by inferring with high likelihood
nationality(X,Y) from bornIn(X,Y). Most previous methods infer simple one-hop
relational synonyms like this, or use as evidence a multi-hop relational path
treated as an atomic feature, like bornIn(X,Z) -> containedIn(Z,Y). This paper
presents an approach that reasons about conjunctions of multi-hop relations
non-atomically, composing the implications of a path using a recursive neural
network (RNN) that takes as inputs vector embeddings of the binary relation in
the path. Not only does this allow us to generalize to paths unseen at training
time, but also, with a single high-capacity RNN, to predict new relation types
not seen when the compositional model was trained (zero-shot learning). We
assemble a new dataset of over 52M relational triples, and show that our method
improves over a traditional classifier by 11%, and a method leveraging
pre-trained embeddings by 7%.
| 2,015 | Computation and Language |
Using Syntax-Based Machine Translation to Parse English into Abstract
Meaning Representation | We present a parser for Abstract Meaning Representation (AMR). We treat
English-to-AMR conversion within the framework of string-to-tree, syntax-based
machine translation (SBMT). To make this work, we transform the AMR structure
into a form suitable for the mechanics of SBMT and useful for modeling. We
introduce an AMR-specific language model and add data and features drawn from
semantic resources. Our resulting AMR parser improves upon state-of-the-art
results by 7 Smatch points.
| 2,015 | Computation and Language |
Exploring semantically-related concepts from Wikipedia: the case of SeRE | In this paper we present our web application SeRE designed to explore
semantically related concepts. Wikipedia and DBpedia are rich data sources to
extract related entities for a given topic, like in- and out-links, broader and
narrower terms, categorisation information etc. We use the Wikipedia full text
body to compute the semantic relatedness for extracted terms, which results in
a list of entities that are most relevant for a topic. For any given query, the
user interface of SeRE visualizes these related concepts, ordered by semantic
relatedness; with snippets from Wikipedia articles that explain the connection
between those two entities. In a user study we examine how SeRE can be used to
find important entities and their relationships for a given topic and to answer
the question of how the classification system can be used for filtering.
| 2,015 | Computation and Language |
Correlational Neural Networks | Common Representation Learning (CRL), wherein different descriptions (or
views) of the data are embedded in a common subspace, is receiving a lot of
attention recently. Two popular paradigms here are Canonical Correlation
Analysis (CCA) based approaches and Autoencoder (AE) based approaches. CCA
based approaches learn a joint representation by maximizing correlation of the
views when projected to the common subspace. AE based methods learn a common
representation by minimizing the error of reconstructing the two views. Each of
these approaches has its own advantages and disadvantages. For example, while
CCA based approaches outperform AE based approaches for the task of transfer
learning, they are not as scalable as the latter. In this work we propose an AE
based approach called Correlational Neural Network (CorrNet), that explicitly
maximizes correlation among the views when projected to the common subspace.
Through a series of experiments, we demonstrate that the proposed CorrNet is
better than the above mentioned approaches with respect to its ability to learn
correlated common representations. Further, we employ CorrNet for several cross
language tasks and show that the representations learned using CorrNet perform
better than the ones learned using other state of the art approaches.
| 2,015 | Computation and Language |
Document Classification by Inversion of Distributed Language
Representations | There have been many recent advances in the structure and measurement of
distributed language models: those that map from words to a vector-space that
is rich in information about word choice and composition. This vector-space is
the distributed language representation. The goal of this note is to point out
that any distributed representation can be turned into a classifier through
inversion via Bayes rule. The approach is simple and modular, in that it will
work with any language representation whose training can be formulated as
optimizing a probability model. In our application to 2 million sentences from
Yelp reviews, we also find that it performs as well as or better than complex
purpose-built algorithms.
| 2,015 | Computation and Language |
Reader-Aware Multi-Document Summarization via Sparse Coding | We propose a new MDS paradigm called reader-aware multi-document
summarization (RA-MDS). Specifically, a set of reader comments associated with
the news reports are also collected. The generated summaries from the reports
for the event should be salient according to not only the reports but also the
reader comments. To tackle this RA-MDS problem, we propose a
sparse-coding-based method that is able to calculate the salience of the text
units by jointly considering news reports and reader comments. Another
reader-aware characteristic of our framework is to improve linguistic quality
via entity rewriting. The rewriting consideration is jointly assessed together
with other summarization requirements under a unified optimization model. To
support the generation of compressive summaries via optimization, we explore a
finer syntactic unit, namely, noun/verb phrase. In this work, we also generate
a data set for conducting RA-MDS. Extensive experiments on this data set and
some classical data sets demonstrate the effectiveness of our proposed
approach.
| 2,015 | Computation and Language |
Lexical Translation Model Using a Deep Neural Network Architecture | In this paper we combine the advantages of a model using global source
sentence contexts, the Discriminative Word Lexicon, and neural networks. By
using deep neural networks instead of the linear maximum entropy model in the
Discriminative Word Lexicon models, we are able to leverage dependencies
between different source words due to the non-linearity. Furthermore, the
models for different target words can share parameters and therefore data
sparsity problems are effectively reduced.
By using this approach in a state-of-the-art translation system, we can
improve the performance by up to 0.5 BLEU points for three different language
pairs on the TED translation task.
| 2,014 | Computation and Language |
CommentWatcher: An Open Source Web-based platform for analyzing
discussions on web forums | We present CommentWatcher, an open source tool aimed at analyzing discussions
on web forums. Constructed as a web platform, CommentWatcher features automatic
mass fetching of user posts from forum on multiple sites, extracting topics,
visualizing the topics as an expression cloud and exploring their temporal
evolution. The underlying social network of users is simultaneously constructed
using the citation relations between users and visualized as a graph structure.
Our platform addresses the issues of the diversity and dynamics of structures
of webpages hosting the forums by implementing a parser architecture that is
independent of the HTML structure of webpages. This allows easy on-the-fly
adding of new websites. Two types of users are targeted: end users who seek to
study the discussed topics and their temporal evolution, and researchers in
need of establishing a forum benchmark dataset and comparing the performances
of analysis tools.
| 2,015 | Computation and Language |
Leveraging Deep Neural Networks and Knowledge Graphs for Entity
Disambiguation | Entity Disambiguation aims to link mentions of ambiguous entities to a
knowledge base (e.g., Wikipedia). Modeling topical coherence is crucial for
this task based on the assumption that information from the same semantic
context tends to belong to the same topic. This paper presents a novel deep
semantic relatedness model (DSRM) based on deep neural networks (DNN) and
semantic knowledge graphs (KGs) to measure entity semantic relatedness for
topical coherence modeling. The DSRM is directly trained on large-scale KGs and
it maps heterogeneous types of knowledge of an entity from KGs to numerical
feature vectors in a latent space such that the distance between two
semantically-related entities is minimized. Compared with the state-of-the-art
relatedness approach proposed by (Milne and Witten, 2008a), the DSRM obtains
19.4% and 24.5% reductions in entity disambiguation errors on two publicly
available datasets respectively.
| 2,015 | Computation and Language |
Detecting Concept-level Emotion Cause in Microblogging | In this paper, we propose a Concept-level Emotion Cause Model (CECM), instead
of the mere word-level models, to discover causes of microblogging users'
diversified emotions on specific hot event. A modified topic-supervised biterm
topic model is utilized in CECM to detect emotion topics' in event-related
tweets, and then context-sensitive topical PageRank is utilized to detect
meaningful multiword expressions as emotion causes. Experimental results on a
dataset from Sina Weibo, one of the largest microblogging websites in China,
show CECM can better detect emotion causes than baseline methods.
| 2,015 | Computation and Language |
Detecting and ordering adjectival scalemates | This paper presents a pattern-based method that can be used to infer
adjectival scales, such as <lukewarm, warm, hot>, from a corpus. Specifically,
the proposed method uses lexical patterns to automatically identify and order
pairs of scalemates, followed by a filtering phase in which unrelated pairs are
discarded. For the filtering phase, several different similarity measures are
implemented and compared. The model presented in this paper is evaluated using
the current standard, along with a novel evaluation set, and shown to be at
least as good as the current state-of-the-art.
| 2,015 | Computation and Language |
Texts in, meaning out: neural language models in semantic similarity
task for Russian | Distributed vector representations for natural language vocabulary get a lot
of attention in contemporary computational linguistics. This paper summarizes
the experience of applying neural network language models to the task of
calculating semantic similarity for Russian. The experiments were performed in
the course of Russian Semantic Similarity Evaluation track, where our models
took from the 2nd to the 5th position, depending on the task.
We introduce the tools and corpora used, comment on the nature of the shared
task and describe the achieved results. It was found out that Continuous
Skip-gram and Continuous Bag-of-words models, previously successfully applied
to English material, can be used for semantic modeling of Russian as well.
Moreover, we show that texts in Russian National Corpus (RNC) provide an
excellent training material for such models, outperforming other, much larger
corpora. It is especially true for semantic relatedness tasks (although
stacking models trained on larger corpora on top of RNC models improves
performance even more).
High-quality semantic vectors learned in such a way can be used in a variety
of linguistic tasks and promise an exciting field for further study.
| 2,015 | Computation and Language |
Parsing Linear Context-Free Rewriting Systems with Fast Matrix
Multiplication | We describe a matrix multiplication recognition algorithm for a subset of
binary linear context-free rewriting systems (LCFRS) with running time
$O(n^{\omega d})$ where $M(m) = O(m^{\omega})$ is the running time for $m
\times m$ matrix multiplication and $d$ is the "contact rank" of the LCFRS --
the maximal number of combination and non-combination points that appear in the
grammar rules. We also show that this algorithm can be used as a subroutine to
get a recognition algorithm for general binary LCFRS with running time
$O(n^{\omega d + 1})$. The currently best known $\omega$ is smaller than
$2.38$. Our result provides another proof for the best known result for parsing
mildly context sensitive formalisms such as combinatory categorial grammars,
head grammars, linear indexed grammars, and tree adjoining grammars, which can
be parsed in time $O(n^{4.76})$. It also shows that inversion transduction
grammars can be parsed in time $O(n^{5.76})$. In addition, binary LCFRS
subsumes many other formalisms and types of grammars, for some of which we also
improve the asymptotic complexity of parsing.
| 2,016 | Computation and Language |
Compositional Distributional Semantics with Compact Closed Categories
and Frobenius Algebras | This thesis contributes to ongoing research related to the categorical
compositional model for natural language of Coecke, Sadrzadeh and Clark in
three ways: Firstly, I propose a concrete instantiation of the abstract
framework based on Frobenius algebras (joint work with Sadrzadeh). The theory
improves shortcomings of previous proposals, extends the coverage of the
language, and is supported by experimental work that improves existing results.
The proposed framework describes a new class of compositional models that find
intuitive interpretations for a number of linguistic phenomena. Secondly, I
propose and evaluate in practice a new compositional methodology which
explicitly deals with the different levels of lexical ambiguity (joint work
with Pulman). A concrete algorithm is presented, based on the separation of
vector disambiguation from composition in an explicit prior step. Extensive
experimental work shows that the proposed methodology indeed results in more
accurate composite representations for the framework of Coecke et al. in
particular and every other class of compositional models in general. As a last
contribution, I formalize the explicit treatment of lexical ambiguity in the
context of the categorical framework by resorting to categorical quantum
mechanics (joint work with Coecke). In the proposed extension, the concept of a
distributional vector is replaced with that of a density matrix, which
compactly represents a probability distribution over the potential different
meanings of the specific word. Composition takes the form of quantum
measurements, leading to interesting analogies between quantum physics and
linguistics.
| 2,015 | Computation and Language |
Embedding Semantic Relations into Word Representations | Learning representations for semantic relations is important for various
tasks such as analogy detection, relational search, and relation
classification. Although there have been several proposals for learning
representations for individual words, learning word representations that
explicitly capture the semantic relations between words remains under
developed. We propose an unsupervised method for learning vector
representations for words such that the learnt representations are sensitive to
the semantic relations that exist between two words. First, we extract lexical
patterns from the co-occurrence contexts of two words in a corpus to represent
the semantic relations that exist between those two words. Second, we represent
a lexical pattern as the weighted sum of the representations of the words that
co-occur with that lexical pattern. Third, we train a binary classifier to
detect relationally similar vs. non-similar lexical pattern pairs. The proposed
method is unsupervised in the sense that the lexical pattern pairs we use as
train data are automatically sampled from a corpus, without requiring any
manual intervention. Our proposed method statistically significantly
outperforms the current state-of-the-art word representations on three
benchmark datasets for proportional analogy detection, demonstrating its
ability to accurately capture the semantic relations among words.
| 2,015 | Computation and Language |
Grounded Discovery of Coordinate Term Relationships between Software
Entities | We present an approach for the detection of coordinate-term relationships
between entities from the software domain, that refer to Java classes. Usually,
relations are found by examining corpus statistics associated with text
entities. In some technical domains, however, we have access to additional
information about the real-world objects named by the entities, suggesting that
coupling information about the "grounded" entities with corpus statistics might
lead to improved methods for relation discovery. To this end, we develop a
similarity measure for Java classes using distributional information about how
they are used in software, which we combine with corpus statistics on the
distribution of contexts in which the classes appear in text. Using our
approach, cross-validation accuracy on this dataset can be improved
dramatically, from around 60% to 88%. Human labeling results show that our
classifier has an F1 score of 86% over the top 1000 predicted pairs.
| 2,015 | Computation and Language |
VQA: Visual Question Answering | We propose the task of free-form and open-ended Visual Question Answering
(VQA). Given an image and a natural language question about the image, the task
is to provide an accurate natural language answer. Mirroring real-world
scenarios, such as helping the visually impaired, both the questions and
answers are open-ended. Visual questions selectively target different areas of
an image, including background details and underlying context. As a result, a
system that succeeds at VQA typically needs a more detailed understanding of
the image and complex reasoning than a system producing generic image captions.
Moreover, VQA is amenable to automatic evaluation, since many open-ended
answers contain only a few words or a closed set of answers that can be
provided in a multiple-choice format. We provide a dataset containing ~0.25M
images, ~0.76M questions, and ~10M answers (www.visualqa.org), and discuss the
information it provides. Numerous baselines and methods for VQA are provided
and compared with human performance. Our VQA demo is available on CloudCV
(http://cloudcv.org/vqa).
| 2,016 | Computation and Language |
Mining Measured Information from Text | We present an approach to extract measured information from text (e.g., a
1370 degrees C melting point, a BMI greater than 29.9 kg/m^2 ). Such
extractions are critically important across a wide range of domains -
especially those involving search and exploration of scientific and technical
documents. We first propose a rule-based entity extractor to mine measured
quantities (i.e., a numeric value paired with a measurement unit), which
supports a vast and comprehensive set of both common and obscure measurement
units. Our method is highly robust and can correctly recover valid measured
quantities even when significant errors are introduced through the process of
converting document formats like PDF to plain text. Next, we describe an
approach to extracting the properties being measured (e.g., the property "pixel
pitch" in the phrase "a pixel pitch as high as 352 {\mu}m"). Finally, we
present MQSearch: the realization of a search engine with full support for
measured information.
| 2,015 | Computation and Language |
Contextual Analysis for Middle Eastern Languages with Hidden Markov
Models | Displaying a document in Middle Eastern languages requires contextual
analysis due to different presentational forms for each character of the
alphabet. The words of the document will be formed by the joining of the
correct positional glyphs representing corresponding presentational forms of
the characters. A set of rules defines the joining of the glyphs. As usual,
these rules vary from language to language and are subject to interpretation by
the software developers.
In this paper, we propose a machine learning approach for contextual analysis
based on the first order Hidden Markov Model. We will design and build a model
for the Farsi language to exhibit this technology. The Farsi model achieves 94
\% accuracy with the training based on a short list of 89 Farsi vocabularies
consisting of 2780 Farsi characters.
The experiment can be easily extended to many languages including Arabic,
Urdu, and Sindhi. Furthermore, the advantage of this approach is that the same
software can be used to perform contextual analysis without coding complex
rules for each specific language. Of particular interest is that the languages
with fewer speakers can have greater representation on the web, since they are
typically ignored by software developers due to lack of financial incentives.
| 2,015 | Computation and Language |