Titles
stringlengths 6
220
| Abstracts
stringlengths 37
3.26k
| Years
int64 1.99k
2.02k
| Categories
stringclasses 1
value |
---|---|---|---|
Preparing Korean Data for the Shared Task on Parsing Morphologically
Rich Languages | This document gives a brief description of Korean data prepared for the SPMRL
2013 shared task. A total of 27,363 sentences with 350,090 tokens are used for
the shared task. All constituent trees are collected from the KAIST Treebank
and transformed to the Penn Treebank style. All dependency trees are converted
from the transformed constituent trees using heuristics and labeling rules de-
signed specifically for the KAIST Treebank. In addition to the gold-standard
morphological analysis provided by the KAIST Treebank, two sets of automatic
morphological analysis are provided for the shared task, one is generated by
the HanNanum morphological analyzer, and the other is generated by the Sejong
morphological analyzer.
| 2,013 | Computation and Language |
The placement of the head that minimizes online memory: a complex
systems approach | It is well known that the length of a syntactic dependency determines its
online memory cost. Thus, the problem of the placement of a head and its
dependents (complements or modifiers) that minimizes online memory is
equivalent to the problem of the minimum linear arrangement of a star tree.
However, how that length is translated into cognitive cost is not known. This
study shows that the online memory cost is minimized when the head is placed at
the center, regardless of the function that transforms length into cost,
provided only that this function is strictly monotonically increasing. Online
memory defines a quasi-convex adaptive landscape with a single central minimum
if the number of elements is odd and two central minima if that number is even.
We discuss various aspects of the dynamics of word order of subject (S), verb
(V) and object (O) from a complex systems perspective and suggest that word
orders tend to evolve by swapping adjacent constituents from an initial or
early SOV configuration that is attracted towards a central word order by
online memory minimization. We also suggest that the stability of SVO is due to
at least two factors, the quasi-convex shape of the adaptive landscape in the
online memory dimension and online memory adaptations that avoid regression to
SOV. Although OVS is also optimal for placing the verb at the center, its low
frequency is explained by its long distance to the seminal SOV in the
permutation space.
| 2,015 | Computation and Language |
Implementation of nlization framework for verbs, pronouns and
determiners with eugene | UNL system is designed and implemented by a nonprofit organization, UNDL
Foundation at Geneva in 1999. UNL applications are application softwares that
allow end users to accomplish natural language tasks, such as translating,
summarizing, retrieving or extracting information, etc. Two major web based
application softwares are Interactive ANalyzer (IAN), which is a natural
language analysis system. It represents natural language sentences as semantic
networks in the UNL format. Other application software is dEep-to-sUrface
GENErator (EUGENE), which is an open-source interactive NLizer. It generates
natural language sentences out of semantic networks represented in the UNL
format. In this paper, NLization framework with EUGENE is focused, while using
UNL system for accomplishing the task of machine translation. In whole
NLization process, EUGENE takes a UNL input and delivers an output in natural
language without any human intervention. It is language-independent and has to
be parametrized to the natural language input through a dictionary and a
grammar, provided as separate interpretable files. In this paper, it is
explained that how UNL input is syntactically and semantically analyzed with
the UNL-NL T-Grammar for NLization of UNL sentences involving verbs, pronouns
and determiners for Punjabi natural language.
| 2,013 | Computation and Language |
General Purpose Textual Sentiment Analysis and Emotion Detection Tools | Textual sentiment analysis and emotion detection consists in retrieving the
sentiment or emotion carried by a text or document. This task can be useful in
many domains: opinion mining, prediction, feedbacks, etc. However, building a
general purpose tool for doing sentiment analysis and emotion detection raises
a number of issues, theoretical issues like the dependence to the domain or to
the language but also pratical issues like the emotion representation for
interoperability. In this paper we present our sentiment/emotion analysis
tools, the way we propose to circumvent the di culties and the applications
they are used for.
| 2,013 | Computation and Language |
Mapping Mutable Genres in Structurally Complex Volumes | To mine large digital libraries in humanistically meaningful ways, scholars
need to divide them by genre. This is a task that classification algorithms are
well suited to assist, but they need adjustment to address the specific
challenges of this domain. Digital libraries pose two problems of scale not
usually found in the article datasets used to test these algorithms. 1) Because
libraries span several centuries, the genres being identified may change
gradually across the time axis. 2) Because volumes are much longer than
articles, they tend to be internally heterogeneous, and the classification task
needs to begin with segmentation. We describe a multi-layered solution that
trains hidden Markov models to segment volumes, and uses ensembles of
overlapping classifiers to address historical change. We test this approach on
a collection of 469,200 volumes drawn from HathiTrust Digital Library. To
demonstrate the humanistic value of these methods, we extract 32,209 volumes of
fiction from the digital library, and trace the changing proportions of first-
and third-person narration in the corpus. We note that narrative points of view
seem to have strong associations with particular themes and genres.
| 2,016 | Computation and Language |
Domain and Function: A Dual-Space Model of Semantic Relations and
Compositions | Given appropriate representations of the semantic relations between carpenter
and wood and between mason and stone (for example, vectors in a vector space
model), a suitable algorithm should be able to recognize that these relations
are highly similar (carpenter is to wood as mason is to stone; the relations
are analogous). Likewise, with representations of dog, house, and kennel, an
algorithm should be able to recognize that the semantic composition of dog and
house, dog house, is highly similar to kennel (dog house and kennel are
synonymous). It seems that these two tasks, recognizing relations and
compositions, are closely connected. However, up to now, the best models for
relations are significantly different from the best models for compositions. In
this paper, we introduce a dual-space model that unifies these two tasks. This
model matches the performance of the best previous models for relations and
compositions. The dual-space model consists of a space for measuring domain
similarity and a space for measuring function similarity. Carpenter and wood
share the same domain, the domain of carpentry. Mason and stone share the same
domain, the domain of masonry. Carpenter and mason share the same function, the
function of artisans. Wood and stone share the same function, the function of
materials. In the composition dog house, kennel has some domain overlap with
both dog and house (the domains of pets and buildings). The function of kennel
is similar to the function of house (the function of shelters). By combining
domain and function similarities in various ways, we can model relations,
compositions, and other aspects of semantics.
| 2,012 | Computation and Language |
Why SOV might be initially preferred and then lost or recovered? A
theoretical framework | Little is known about why SOV order is initially preferred and then discarded
or recovered. Here we present a framework for understanding these and many
related word order phenomena: the diversity of dominant orders, the existence
of free words orders, the need of alternative word orders and word order
reversions and cycles in evolution. Under that framework, word order is
regarded as a multiconstraint satisfaction problem in which at least two
constraints are in conflict: online memory minimization and maximum
predictability.
| 2,014 | Computation and Language |
Exploiting Similarities among Languages for Machine Translation | Dictionaries and phrase tables are the basis of modern statistical machine
translation systems. This paper develops a method that can automate the process
of generating and extending dictionaries and phrase tables. Our method can
translate missing word and phrase entries by learning language structures based
on large monolingual data and mapping between languages from small bilingual
data. It uses distributed representation of words and learns a linear mapping
between vector spaces of languages. Despite its simplicity, our method is
surprisingly effective: we can achieve almost 90% precision@5 for translation
of words between English and Spanish. This method makes little assumption about
the languages, so it can be used to extend and refine dictionaries and
translation tables for any language pairs.
| 2,013 | Computation and Language |
Text segmentation with character-level text embeddings | Learning word representations has recently seen much success in computational
linguistics. However, assuming sequences of word tokens as input to linguistic
analysis is often unjustified. For many languages word segmentation is a
non-trivial task and naturally occurring text is sometimes a mixture of natural
language strings and other character data. We propose to learn text
representations directly from raw character sequences by training a Simple
recurrent Network to predict the next character in text. The network uses its
hidden layer to evolve abstract representations of the character sequences it
sees. To demonstrate the usefulness of the learned text embeddings, we use them
as features in a supervised character level text segmentation and labeling
task: recognizing spans of text containing programming language code. By using
the embeddings as features we are able to substantially improve over a baseline
which uses only surface character n-grams.
| 2,013 | Computation and Language |
JRC EuroVoc Indexer JEX - A freely available multi-label categorisation
tool | EuroVoc (2012) is a highly multilingual thesaurus consisting of over 6,700
hierarchically organised subject domains used by European Institutions and many
authorities in Member States of the European Union (EU) for the classification
and retrieval of official documents. JEX is JRC-developed multi-label
classification software that learns from manually labelled data to
automatically assign EuroVoc descriptors to new documents in a profile-based
category-ranking task. The JEX release consists of trained classifiers for 22
official EU languages, of parallel training data in the same languages, of an
interface that allows viewing and amending the assignment results, and of a
module that allows users to re-train the tool on their own document
collections. JEX allows advanced users to change the document representation so
as to possibly improve the categorisation result through linguistic
pre-processing. JEX can be used as a tool for interactive EuroVoc descriptor
assignment to increase speed and consistency of the human categorisation
process, or it can be used fully automatically. The output of JEX is a
language-independent EuroVoc feature vector lending itself also as input to
various other Language Technology tasks, including cross-lingual clustering and
classification, cross-lingual plagiarism detection, sentence selection and
ranking, and more.
| 2,012 | Computation and Language |
DGT-TM: A freely Available Translation Memory in 22 Languages | The European Commission's (EC) Directorate General for Translation, together
with the EC's Joint Research Centre, is making available a large translation
memory (TM; i.e. sentences and their professionally produced translations)
covering twenty-two official European Union (EU) languages and their 231
language pairs. Such a resource is typically used by translation professionals
in combination with TM software to improve speed and consistency of their
translations. However, this resource has also many uses for translation studies
and for language technology applications, including Statistical Machine
Translation (SMT), terminology extraction, Named Entity Recognition (NER),
multilingual classification and clustering, and many more. In this reference
paper for DGT-TM, we introduce this new resource, provide statistics regarding
its size, and explain how it was produced and how to use it.
| 2,012 | Computation and Language |
An introduction to the Europe Media Monitor family of applications | Most large organizations have dedicated departments that monitor the media to
keep up-to-date with relevant developments and to keep an eye on how they are
represented in the news. Part of this media monitoring work can be automated.
In the European Union with its 23 official languages, it is particularly
important to cover media reports in many languages in order to capture the
complementary news content published in the different countries. It is also
important to be able to access the news content across languages and to merge
the extracted information. We present here the four publicly accessible systems
of the Europe Media Monitor (EMM) family of applications, which cover between
19 and 50 languages (see http://press.jrc.it/overview.html). We give an
overview of their functionality and discuss some of the implications of the
fact that they cover quite so many languages. We discuss design issues
necessary to be able to achieve this high multilinguality, as well as the
benefits of this multilinguality.
| 2,009 | Computation and Language |
Recognizing Speech in a Novel Accent: The Motor Theory of Speech
Perception Reframed | The motor theory of speech perception holds that we perceive the speech of
another in terms of a motor representation of that speech. However, when we
have learned to recognize a foreign accent, it seems plausible that recognition
of a word rarely involves reconstruction of the speech gestures of the speaker
rather than the listener. To better assess the motor theory and this
observation, we proceed in three stages. Part 1 places the motor theory of
speech perception in a larger framework based on our earlier models of the
adaptive formation of mirror neurons for grasping, and for viewing extensions
of that mirror system as part of a larger system for neuro-linguistic
processing, augmented by the present consideration of recognizing speech in a
novel accent. Part 2 then offers a novel computational model of how a listener
comes to understand the speech of someone speaking the listener's native
language with a foreign accent. The core tenet of the model is that the
listener uses hypotheses about the word the speaker is currently uttering to
update probabilities linking the sound produced by the speaker to phonemes in
the native language repertoire of the listener. This, on average, improves the
recognition of later words. This model is neutral regarding the nature of the
representations it uses (motor vs. auditory). It serve as a reference point for
the discussion in Part 3, which proposes a dual-stream neuro-linguistic
architecture to revisits claims for and against the motor theory of speech
perception and the relevance of mirror neurons, and extracts some implications
for the reframing of the motor theory.
| 2,013 | Computation and Language |
Even the Abstract have Colour: Consensus in Word-Colour Associations | Colour is a key component in the successful dissemination of information.
Since many real-world concepts are associated with colour, for example danger
with red, linguistic information is often complemented with the use of
appropriate colours in information visualization and product marketing. Yet,
there is no comprehensive resource that captures concept-colour associations.
We present a method to create a large word-colour association lexicon by
crowdsourcing. A word-choice question was used to obtain sense-level
annotations and to ensure data quality. We focus especially on abstract
concepts and emotions to show that even they tend to have strong colour
associations. Thus, using the right colours can not only improve semantic
coherence, but also inspire the desired emotional response.
| 2,013 | Computation and Language |
LDC Arabic Treebanks and Associated Corpora: Data Divisions Manual | The Linguistic Data Consortium (LDC) has developed hundreds of data corpora
for natural language processing (NLP) research. Among these are a number of
annotated treebank corpora for Arabic. Typically, these corpora consist of a
single collection of annotated documents. NLP research, however, usually
requires multiple data sets for the purposes of training models, developing
techniques, and final evaluation. Therefore it becomes necessary to divide the
corpora used into the required data sets (divisions). This document details a
set of rules that have been defined to enable consistent divisions for old and
new Arabic treebanks (ATB) and related corpora.
| 2,013 | Computation and Language |
A Hybrid Algorithm for Matching Arabic Names | In this paper, a new hybrid algorithm which combines both of token-based and
character-based approaches is presented. The basic Levenshtein approach has
been extended to token-based distance metric. The distance metric is enhanced
to set the proper granularity level behavior of the algorithm. It smoothly maps
a threshold of misspellings differences at the character level, and the
importance of token level errors in terms of token's position and frequency.
Using a large Arabic dataset, the experimental results show that the proposed
algorithm overcomes successfully many types of errors such as: typographical
errors, omission or insertion of middle name components, omission of
non-significant popular name components, and different writing styles character
variations. When compared the results with other classical algorithms, using
the same dataset, the proposed algorithm was found to increase the minimum
success level of best tested algorithms, while achieving higher upper limits .
| 2,013 | Computation and Language |
Sentiment Analysis: How to Derive Prior Polarities from SentiWordNet | Assigning a positive or negative score to a word out of context (i.e. a
word's prior polarity) is a challenging task for sentiment analysis. In the
literature, various approaches based on SentiWordNet have been proposed. In
this paper, we compare the most often used techniques together with newly
proposed ones and incorporate all of them in a learning framework to see
whether blending them can further improve the estimation of prior polarity
scores. Using two different versions of SentiWordNet and testing regression and
classification models across tasks and datasets, our learning approach
consistently outperforms the single metrics, providing a new state-of-the-art
approach in computing words' prior polarity for sentiment analysis. We conclude
our investigation showing interesting biases in calculated prior polarity
scores when word Part of Speech and annotator gender are considered.
| 2,013 | Computation and Language |
From Once Upon a Time to Happily Ever After: Tracking Emotions in Novels
and Fairy Tales | Today we have access to unprecedented amounts of literary texts. However,
search still relies heavily on key words. In this paper, we show how sentiment
analysis can be used in tandem with effective visualizations to quantify and
track emotions in both individual books and across very large collections. We
introduce the concept of emotion word density, and using the Brothers Grimm
fairy tales as example, we show how collections of text can be organized for
better search. Using the Google Books Corpus we show how to determine an
entity's emotion associations from co-occurring words. Finally, we compare
emotion words in fairy tales and novels, to show that fairy tales have a much
wider range of emotion word densities than novels.
| 2,011 | Computation and Language |
Colourful Language: Measuring Word-Colour Associations | Since many real-world concepts are associated with colour, for example danger
with red, linguistic information is often complimented with the use of
appropriate colours in information visualization and product marketing. Yet,
there is no comprehensive resource that captures concept-colour associations.
We present a method to create a large word-colour association lexicon by
crowdsourcing. We focus especially on abstract concepts and emotions to show
that even though they cannot be physically visualized, they too tend to have
strong colour associations. Finally, we show how word-colour associations
manifest themselves in language, and quantify usefulness of co-occurrence and
polarity cues in automatically detecting colour associations.
| 2,011 | Computation and Language |
JRC-Names: A freely available, highly multilingual named entity resource | This paper describes a new, freely available, highly multilingual named
entity resource for person and organisation names that has been compiled over
seven years of large-scale multilingual news analysis combined with Wikipedia
mining, resulting in 205,000 per-son and organisation names plus about the same
number of spelling variants written in over 20 different scripts and in many
more languages. This resource, produced as part of the Europe Media Monitor
activity (EMM, http://emm.newsbrief.eu/overview.html), can be used for a number
of purposes. These include improving name search in databases or on the
internet, seeding machine learning systems to learn named entity recognition
rules, improve machine translation results, and more. We describe here how this
resource was created; we give statistics on its current size; we address the
issue of morphological inflection; and we give details regarding its
functionality. Updates to this resource will be made available daily.
| 2,011 | Computation and Language |
Feature Learning with Gaussian Restricted Boltzmann Machine for Robust
Speech Recognition | In this paper, we first present a new variant of Gaussian restricted
Boltzmann machine (GRBM) called multivariate Gaussian restricted Boltzmann
machine (MGRBM), with its definition and learning algorithm. Then we propose
using a learned GRBM or MGRBM to extract better features for robust speech
recognition. Our experiments on Aurora2 show that both GRBM-extracted and
MGRBM-extracted feature performs much better than Mel-frequency cepstral
coefficient (MFCC) with either HMM-GMM or hybrid HMM-deep neural network (DNN)
acoustic model, and MGRBM-extracted feature is slightly better.
| 2,013 | Computation and Language |
Acronym recognition and processing in 22 languages | We are presenting work on recognising acronyms of the form Long-Form
(Short-Form) such as "International Monetary Fund (IMF)" in millions of news
articles in twenty-two languages, as part of our more general effort to
recognise entities and their variants in news text and to use them for the
automatic analysis of the news, including the linking of related news across
languages. We show how the acronym recognition patterns, initially developed
for medical terms, needed to be adapted to the more general news domain and we
present evaluation results. We describe our effort to automatically merge the
numerous long-form variants referring to the same short-form, while keeping
non-related long-forms separate. Finally, we provide extensive statistics on
the frequency and the distribution of short-form/long-form pairs across
languages.
| 2,013 | Computation and Language |
Sentiment Analysis in the News | Recent years have brought a significant growth in the volume of research in
sentiment analysis, mostly on highly subjective text types (movie or product
reviews). The main difference these texts have with news articles is that their
target is clearly defined and unique across the text. Following different
annotation efforts and the analysis of the issues encountered, we realised that
news opinion mining is different from that of other text types. We identified
three subtasks that need to be addressed: definition of the target; separation
of the good and bad news content from the good and bad sentiment expressed on
the target; and analysis of clearly marked opinion that is expressed
explicitly, not needing interpretation or the use of world knowledge.
Furthermore, we distinguish three different possible views on newspaper
articles - author, reader and text, which have to be addressed differently at
the time of analysing sentiment. Given these definitions, we present work on
mining opinions about entities in English language news, in which (a) we test
the relative suitability of various sentiment dictionaries and (b) we attempt
to separate positive or negative opinion from good or bad news. In the
experiments described here, we tested whether or not subject domain-defining
vocabulary should be ignored. Results showed that this idea is more appropriate
in the context of news opinion mining and that the approaches taking this into
consideration produce a better performance.
| 2,010 | Computation and Language |
Tracking Sentiment in Mail: How Genders Differ on Emotional Axes | With the widespread use of email, we now have access to unprecedented amounts
of text that we ourselves have written. In this paper, we show how sentiment
analysis can be used in tandem with effective visualizations to quantify and
track emotions in many types of mail. We create a large word--emotion
association lexicon by crowdsourcing, and use it to compare emotions in love
letters, hate mail, and suicide notes. We show that there are marked
differences across genders in how they use emotion words in work-place email.
For example, women use many words from the joy--sadness axis, whereas men
prefer terms from the fear--trust axis. Finally, we show visualizations that
can help people track emotions in their emails.
| 2,013 | Computation and Language |
Using Nuances of Emotion to Identify Personality | Past work on personality detection has shown that frequency of lexical
categories such as first person pronouns, past tense verbs, and sentiment words
have significant correlations with personality traits. In this paper, for the
first time, we show that fine affect (emotion) categories such as that of
excitement, guilt, yearning, and admiration are significant indicators of
personality. Additionally, we perform experiments to show that the gains
provided by the fine affect categories are not obtained by using coarse affect
categories alone or with specificity features alone. We employ these features
in five SVM classifiers for detecting five personality traits through essays.
We find that the use of fine emotion features leads to statistically
significant improvement over a competitive baseline, whereas the use of coarse
affect and specificity features does not.
| 2,013 | Computation and Language |
An Inter-lingual Reference Approach For Multi-Lingual Ontology Matching | Ontologies are considered as the backbone of the Semantic Web. With the
rising success of the Semantic Web, the number of participating communities
from different countries is constantly increasing. The growing number of
ontologies available in different natural languages leads to an
interoperability problem. In this paper, we discuss several approaches for
ontology matching; examine similarities and differences, identify weaknesses,
and compare the existing automated approaches with the manual approaches for
integrating multilingual ontologies. In addition to that, we propose a new
architecture for a multilingual ontology matching service. As a case study we
used an example of two multilingual enterprise ontologies - the university
ontology of Freie Universitaet Berlin and the ontology for Fayoum University in
Egypt.
| 2,013 | Computation and Language |
Domain-Specific Sentiment Word Extraction by Seed Expansion and Pattern
Generation | This paper focuses on the automatic extraction of domain-specific sentiment
word (DSSW), which is a fundamental subtask of sentiment analysis. Most
previous work utilizes manual patterns for this task. However, the performance
of those methods highly relies on the labelled patterns or selected seeds. In
order to overcome the above problem, this paper presents an automatic framework
to detect large-scale domain-specific patterns for DSSW extraction. To this
end, sentiment seeds are extracted from massive dataset of user comments.
Subsequently, these sentiment seeds are expanded by synonyms using a
bootstrapping mechanism. Simultaneously, a synonymy graph is built and the
graph propagation algorithm is applied on the built synonymy graph. Afterwards,
syntactic and sequential relations between target words and high-ranked
sentiment words are extracted automatically to construct large-scale patterns,
which are further used to extracte DSSWs. The experimental results in three
domains reveal the effectiveness of our method.
| 2,013 | Computation and Language |
Development and Transcription of Assamese Speech Corpus | A balanced speech corpus is the basic need for any speech processing task. In
this report we describe our effort on development of Assamese speech corpus. We
mainly focused on some issues and challenges faced during development of the
corpus. Being a less computationally aware language, this is the first effort
to develop speech corpus for Assamese. As corpus development is an ongoing
process, in this paper we report only the initial task.
| 2,013 | Computation and Language |
Cross-Recurrence Quantification Analysis of Categorical and Continuous
Time Series: an R package | This paper describes the R package crqa to perform cross-recurrence
quantification analysis of two time series of either a categorical or
continuous nature. Streams of behavioral information, from eye movements to
linguistic elements, unfold over time. When two people interact, such as in
conversation, they often adapt to each other, leading these behavioral levels
to exhibit recurrent states. In dialogue, for example, interlocutors adapt to
each other by exchanging interactive cues: smiles, nods, gestures, choice of
words, and so on. In order for us to capture closely the goings-on of dynamic
interaction, and uncover the extent of coupling between two individuals, we
need to quantify how much recurrence is taking place at these levels. Methods
available in crqa would allow researchers in cognitive science to pose such
questions as how much are two people recurrent at some level of analysis, what
is the characteristic lag time for one person to maximally match another, or
whether one person is leading another. First, we set the theoretical ground to
understand the difference between 'correlation' and 'co-visitation' when
comparing two time series, using an aggregative or cross-recurrence approach.
Then, we describe more formally the principles of cross-recurrence, and show
with the current package how to carry out analyses applying them. We end the
paper by comparing computational efficiency, and results' consistency, of crqa
R package, with the benchmark MATLAB toolbox crptoolbox. We show perfect
comparability between the two libraries on both levels.
| 2,013 | Computation and Language |
Improving the Quality of MT Output using Novel Name Entity Translation
Scheme | This paper presents a novel approach to machine translation by combining the
state of art name entity translation scheme. Improper translation of name
entities lapse the quality of machine translated output. In this work, name
entities are transliterated by using statistical rule based approach. This
paper describes the translation and transliteration of name entities from
English to Punjabi. We have experimented on four types of name entities which
are: Proper names, Location names, Organization names and miscellaneous.
Various rules for the purpose of syllabification have been constructed.
Transliteration of name entities is accomplished with the help of Probability
calculation. N-Gram probabilities for the extracted syllables have been
calculated using statistical machine translation toolkit MOSES.
| 2,013 | Computation and Language |
Development of Marathi Part of Speech Tagger Using Statistical Approach | Part-of-speech (POS) tagging is a process of assigning the words in a text
corresponding to a particular part of speech. A fundamental version of POS
tagging is the identification of words as nouns, verbs, adjectives etc. For
processing natural languages, Part of Speech tagging is a prominent tool. It is
one of the simplest as well as most constant and statistical model for many NLP
applications. POS Tagging is an initial stage of linguistics, text analysis
like information retrieval, machine translator, text to speech synthesis,
information extraction etc. In POS Tagging we assign a Part of Speech tag to
each word in a sentence and literature. Various approaches have been proposed
to implement POS taggers. In this paper we present a Marathi part of speech
tagger. It is morphologically rich language. Marathi is spoken by the native
people of Maharashtra. The general approach used for development of tagger is
statistical using Unigram, Bigram, Trigram and HMM Methods. It presents a clear
idea about all the algorithms with suitable examples. It also introduces a tag
set for Marathi which can be used for tagging Marathi text. In this paper we
have shown the development of the tagger as well as compared to check the
accuracy of taggers output. The three Marathi POS taggers viz. Unigram, Bigram,
Trigram and HMM gives the accuracy of 77.38%, 90.30%, 91.46% and 93.82%
respectively.
| 2,013 | Computation and Language |
Subjective and Objective Evaluation of English to Urdu Machine
Translation | Machine translation is research based area where evaluation is very important
phenomenon for checking the quality of MT output. The work is based on the
evaluation of English to Urdu Machine translation. In this research work we
have evaluated the translation quality of Urdu language which has been
translated by using different Machine Translation systems like Google, Babylon
and Ijunoon. The evaluation process is done by using two approaches - Human
evaluation and Automatic evaluation. We have worked for both the approaches
where in human evaluation emphasis is given to scales and parameters while in
automatic evaluation emphasis is given to some automatic metric such as BLEU,
GTM, METEOR and ATEC.
| 2,013 | Computation and Language |
Rule Based Stemmer in Urdu | Urdu is a combination of several languages like Arabic, Hindi, English,
Turkish, Sanskrit etc. It has a complex and rich morphology. This is the reason
why not much work has been done in Urdu language processing. Stemming is used
to convert a word into its respective root form. In stemming, we separate the
suffix and prefix from the word. It is useful in search engines, natural
language processing and word processing, spell checkers, word parsing, word
frequency and count studies. This paper presents a rule based stemmer for Urdu.
The stemmer that we have discussed here is used in information retrieval. We
have also evaluated our results by verifying it with a human expert.
| 2,013 | Computation and Language |
Stemmers for Tamil Language: Performance Analysis | Stemming is the process of extracting root word from the given inflection
word and also plays significant role in numerous application of Natural
Language Processing (NLP). Tamil Language raises several challenges to NLP,
since it has rich morphological patterns than other languages. The rule based
approach light-stemmer is proposed in this paper, to find stem word for given
inflection Tamil word. The performance of proposed approach is compared to a
rule based suffix removal stemmer based on correctly and incorrectly predicted.
The experimental result clearly show that the proposed approach light stemmer
for Tamil language perform better than suffix removal stemmer and also more
effective in Information Retrieval System (IRS).
| 2,013 | Computation and Language |
Semantic Measures for the Comparison of Units of Language, Concepts or
Instances from Text and Knowledge Base Analysis | Semantic measures are widely used today to estimate the strength of the
semantic relationship between elements of various types: units of language
(e.g., words, sentences, documents), concepts or even instances semantically
characterized (e.g., diseases, genes, geographical locations). Semantic
measures play an important role to compare such elements according to semantic
proxies: texts and knowledge representations, which support their meaning or
describe their nature. Semantic measures are therefore essential for designing
intelligent agents which will for example take advantage of semantic analysis
to mimic human ability to compare abstract or concrete objects. This paper
proposes a comprehensive survey of the broad notion of semantic measure for the
comparison of units of language, concepts or instances based on semantic proxy
analyses. Semantic measures generalize the well-known notions of semantic
similarity, semantic relatedness and semantic distance, which have been
extensively studied by various communities over the last decades (e.g.,
Cognitive Sciences, Linguistics, and Artificial Intelligence to mention a few).
| 2,016 | Computation and Language |
A State of the Art of Word Sense Induction: A Way Towards Word Sense
Disambiguation for Under-Resourced Languages | Word Sense Disambiguation (WSD), the process of automatically identifying the
meaning of a polysemous word in a sentence, is a fundamental task in Natural
Language Processing (NLP). Progress in this approach to WSD opens up many
promising developments in the field of NLP and its applications. Indeed,
improvement over current performance levels could allow us to take a first step
towards natural language understanding. Due to the lack of lexical resources it
is sometimes difficult to perform WSD for under-resourced languages. This paper
is an investigation on how to initiate research in WSD for under-resourced
languages by applying Word Sense Induction (WSI) and suggests some interesting
topics to focus on.
| 2,013 | Computation and Language |
Local Feature or Mel Frequency Cepstral Coefficients - Which One is
Better for MLN-Based Bangla Speech Recognition? | This paper discusses the dominancy of local features (LFs), as input to the
multilayer neural network (MLN), extracted from a Bangla input speech over mel
frequency cepstral coefficients (MFCCs). Here, LF-based method comprises three
stages: (i) LF extraction from input speech, (ii) phoneme probabilities
extraction using MLN from LF and (iii) the hidden Markov model (HMM) based
classifier to obtain more accurate phoneme strings. In the experiments on
Bangla speech corpus prepared by us, it is observed that the LFbased automatic
speech recognition (ASR) system provides higher phoneme correct rate than the
MFCC-based system. Moreover, the proposed system requires fewer mixture
components in the HMMs.
| 2,013 | Computation and Language |
Evolution of the Modern Phase of Written Bangla: A Statistical Study | Active languages such as Bangla (or Bengali) evolve over time due to a
variety of social, cultural, economic, and political issues. In this paper, we
analyze the change in the written form of the modern phase of Bangla
quantitatively in terms of character-level, syllable-level, morpheme-level and
word-level features. We collect three different types of corpora---classical,
newspapers and blogs---and test whether the differences in their features are
statistically significant. Results suggest that there are significant changes
in the length of a word when measured in terms of characters, but there is not
much difference in usage of different characters, syllables and morphemes in a
word or of different words in a sentence. To the best of our knowledge, this is
the first work on Bangla of this kind.
| 2,013 | Computation and Language |
Cross-lingual Pseudo-Projected Expectation Regularization for Weakly
Supervised Learning | We consider a multilingual weakly supervised learning scenario where
knowledge from annotated corpora in a resource-rich language is transferred via
bitext to guide the learning in other languages. Past approaches project labels
across bitext and use them as features or gold labels for training. We propose
a new method that projects model expectations rather than labels, which
facilities transfer of model uncertainty across language boundaries. We encode
expectations as constraints and train a discriminative CRF model using
Generalized Expectation Criteria (Mann and McCallum, 2010). Evaluated on
standard Chinese-English and German-English NER datasets, our method
demonstrates F1 scores of 64% and 60% when no labeled data is used. Attaining
the same accuracy with supervised CRFs requires 12k and 1.5k labeled sentences.
Furthermore, when combined with labeled examples, our method yields significant
improvements over state-of-the-art supervised methods, achieving best reported
numbers to date on Chinese OntoNotes and German CoNLL-03 datasets.
| 2,013 | Computation and Language |
Named entity recognition using conditional random fields with non-local
relational constraints | We begin by introducing the Computer Science branch of Natural Language
Processing, then narrowing the attention on its subbranch of Information
Extraction and particularly on Named Entity Recognition, discussing briefly its
main methodological approaches. It follows an introduction to state-of-the-art
Conditional Random Fields under the form of linear chains. Subsequently, the
idea of constrained inference as a way to model long-distance relationships in
a text is presented, based on an Integer Linear Programming representation of
the problem. Adding such relationships to the problem as automatically inferred
logical formulas, translatable into linear conditions, we propose to solve the
resulting more complex problem with the aid of Lagrangian relaxation, of which
some technical details are explained. Lastly, we give some experimental
results.
| 2,013 | Computation and Language |
ARKref: a rule-based coreference resolution system | ARKref is a tool for noun phrase coreference. It is a deterministic,
rule-based system that uses syntactic information from a constituent parser,
and semantic information from an entity recognition component. Its architecture
is based on the work of Haghighi and Klein (2009). ARKref was originally
written in 2009. At the time of writing, the last released version was in March
2011. This document describes that version, which is open-source and publicly
available at: http://www.ark.cs.cmu.edu/ARKref
| 2,013 | Computation and Language |
Treating clitics with minimalist grammars | We propose an extension of Stabler's version of clitics treatment for a wider
coverage of the French language. For this, we present the lexical entries
needed in the lexicon. Then, we show the recognition of complex syntactic
phenomena as (left and right) dislo- cation, clitic climbing over modal and
extraction from determiner phrase. The aim of this presentation is the
syntax-semantic interface for clitics analyses in which we will stress on
clitic climbing over verb and raising verb.
| 2,013 | Computation and Language |
Distributed Representations of Words and Phrases and their
Compositionality | The recently introduced continuous Skip-gram model is an efficient method for
learning high-quality distributed vector representations that capture a large
number of precise syntactic and semantic word relationships. In this paper we
present several extensions that improve both the quality of the vectors and the
training speed. By subsampling of the frequent words we obtain significant
speedup and also learn more regular word representations. We also describe a
simple alternative to the hierarchical softmax called negative sampling. An
inherent limitation of word representations is their indifference to word order
and their inability to represent idiomatic phrases. For example, the meanings
of "Canada" and "Air" cannot be easily combined to obtain "Air Canada".
Motivated by this example, we present a simple method for finding phrases in
text, and show that learning good vector representations for millions of
phrases is possible.
| 2,013 | Computation and Language |
A Logic-based Approach for Recognizing Textual Entailment Supported by
Ontological Background Knowledge | We present the architecture and the evaluation of a new system for
recognizing textual entailment (RTE). In RTE we want to identify automatically
the type of a logical relation between two input texts. In particular, we are
interested in proving the existence of an entailment between them. We conceive
our system as a modular environment allowing for a high-coverage syntactic and
semantic text analysis combined with logical inference. For the syntactic and
semantic analysis we combine a deep semantic analysis with a shallow one
supported by statistical models in order to increase the quality and the
accuracy of results. For RTE we use logical inference of first-order employing
model-theoretic techniques and automated reasoning tools. The inference is
supported with problem-relevant background knowledge extracted automatically
and on demand from external sources like, e.g., WordNet, YAGO, and OpenCyc, or
other, more experimental sources with, e.g., manually defined presupposition
resolutions, or with axiomatized general and common sense knowledge. The
results show that fine-grained and consistent knowledge coming from diverse
sources is a necessary condition determining the correctness and traceability
of results.
| 2,013 | Computation and Language |
The optimality of attaching unlinked labels to unlinked meanings | Vocabulary learning by children can be characterized by many biases. When
encountering a new word, children as well as adults, are biased towards
assuming that it means something totally different from the words that they
already know. To the best of our knowledge, the 1st mathematical proof of the
optimality of this bias is presented here. First, it is shown that this bias is
a particular case of the maximization of mutual information between words and
meanings. Second, the optimality is proven within a more general information
theoretic framework where mutual information maximization competes with other
information theoretic principles. The bias is a prediction from modern
information theory. The relationship between information theoretic principles
and the principles of contrast and mutual exclusivity is also shown.
| 2,017 | Computation and Language |
Sockpuppet Detection in Wikipedia: A Corpus of Real-World Deceptive
Writing for Linking Identities | This paper describes the corpus of sockpuppet cases we gathered from
Wikipedia. A sockpuppet is an online user account created with a fake identity
for the purpose of covering abusive behavior and/or subverting the editing
regulation process. We used a semi-automated method for crawling and curating a
dataset of real sockpuppet investigation cases. To the best of our knowledge,
this is the first corpus available on real-world deceptive writing. We describe
the process for crawling the data and some preliminary results that can be used
as baseline for benchmarking research. The dataset will be released under a
Creative Commons license from our project website: http://docsig.cis.uab.edu.
| 2,013 | Computation and Language |
Description and Evaluation of Semantic Similarity Measures Approaches | In recent years, semantic similarity measure has a great interest in Semantic
Web and Natural Language Processing (NLP). Several similarity measures have
been developed, being given the existence of a structured knowledge
representation offered by ontologies and corpus which enable semantic
interpretation of terms. Semantic similarity measures compute the similarity
between concepts/terms included in knowledge sources in order to perform
estimations. This paper discusses the existing semantic similarity methods
based on structure, information content and feature approaches. Additionally,
we present a critical evaluation of several categories of semantic similarity
approaches based on two standard benchmarks. The aim of this paper is to give
an efficient evaluation of all these measures which help researcher and
practitioners to select the measure that best fit for their requirements.
| 2,013 | Computation and Language |
A Comparative Study on Linguistic Feature Selection in Sentiment
Polarity Classification | Sentiment polarity classification is perhaps the most widely studied topic.
It classifies an opinionated document as expressing a positive or negative
opinion. In this paper, using movie review dataset, we perform a comparative
study with different single kind linguistic features and the combinations of
these features. We find that the classic topic-based classifier(Naive Bayes and
Support Vector Machine) do not perform as well on sentiment polarity
classification. And we find that with some combination of different linguistic
features, the classification accuracy can be boosted a lot. We give some
reasonable explanations about these boosting outcomes.
| 2,013 | Computation and Language |
Using Robust PCA to estimate regional characteristics of language use
from geo-tagged Twitter messages | Principal component analysis (PCA) and related techniques have been
successfully employed in natural language processing. Text mining applications
in the age of the online social media (OSM) face new challenges due to
properties specific to these use cases (e.g. spelling issues specific to texts
posted by users, the presence of spammers and bots, service announcements,
etc.). In this paper, we employ a Robust PCA technique to separate typical
outliers and highly localized topics from the low-dimensional structure present
in language use in online social networks. Our focus is on identifying
geospatial features among the messages posted by the users of the Twitter
microblogging service. Using a dataset which consists of over 200 million
geolocated tweets collected over the course of a year, we investigate whether
the information present in word usage frequencies can be used to identify
regional features of language use and topics of interest. Using the PCA pursuit
method, we are able to identify important low-dimensional features, which
constitute smoothly varying functions of the geographic location.
| 2,013 | Computation and Language |
Identifying Purpose Behind Electoral Tweets | Tweets pertaining to a single event, such as a national election, can number
in the hundreds of millions. Automatically analyzing them is beneficial in many
downstream natural language applications such as question answering and
summarization. In this paper, we propose a new task: identifying the purpose
behind electoral tweets--why do people post election-oriented tweets? We show
that identifying purpose is correlated with the related phenomenon of sentiment
and emotion detection, but yet significantly different. Detecting purpose has a
number of applications including detecting the mood of the electorate,
estimating the popularity of policies, identifying key issues of contention,
and predicting the course of events. We create a large dataset of electoral
tweets and annotate a few thousand tweets for purpose. We develop a system that
automatically classifies electoral tweets as per their purpose, obtaining an
accuracy of 43.56% on an 11-class task and an accuracy of 73.91% on a 3-class
task (both accuracies well above the most-frequent-class baseline). Finally, we
show that resources developed for emotion detection are also helpful for
detecting purpose.
| 2,013 | Computation and Language |
Category-Theoretic Quantitative Compositional Distributional Models of
Natural Language Semantics | This thesis is about the problem of compositionality in distributional
semantics. Distributional semantics presupposes that the meanings of words are
a function of their occurrences in textual contexts. It models words as
distributions over these contexts and represents them as vectors in high
dimensional spaces. The problem of compositionality for such models concerns
itself with how to produce representations for larger units of text by
composing the representations of smaller units of text.
This thesis focuses on a particular approach to this compositionality
problem, namely using the categorical framework developed by Coecke, Sadrzadeh,
and Clark, which combines syntactic analysis formalisms with distributional
semantic representations of meaning to produce syntactically motivated
composition operations. This thesis shows how this approach can be
theoretically extended and practically implemented to produce concrete
compositional distributional models of natural language semantics. It
furthermore demonstrates that such models can perform on par with, or better
than, other competing approaches in the field of natural language processing.
There are three principal contributions to computational linguistics in this
thesis. The first is to extend the DisCoCat framework on the syntactic front
and semantic front, incorporating a number of syntactic analysis formalisms and
providing learning procedures allowing for the generation of concrete
compositional distributional models. The second contribution is to evaluate the
models developed from the procedures presented here, showing that they
outperform other compositional distributional models present in the literature.
The third contribution is to show how using category theory to solve linguistic
problems forms a sound basis for research, illustrated by examples of work on
this topic, that also suggest directions for future research.
| 2,013 | Computation and Language |
Semantic Sort: A Supervised Approach to Personalized Semantic
Relatedness | We propose and study a novel supervised approach to learning statistical
semantic relatedness models from subjectively annotated training examples. The
proposed semantic model consists of parameterized co-occurrence statistics
associated with textual units of a large background knowledge corpus. We
present an efficient algorithm for learning such semantic models from a
training sample of relatedness preferences. Our method is corpus independent
and can essentially rely on any sufficiently large (unstructured) collection of
coherent texts. Moreover, the approach facilitates the fitting of semantic
models for specific users or groups of users. We present the results of
extensive range of experiments from small to large scale, indicating that the
proposed method is effective and competitive with the state-of-the-art.
| 2,013 | Computation and Language |
Authorship Attribution Using Word Network Features | In this paper, we explore a set of novel features for authorship attribution
of documents. These features are derived from a word network representation of
natural language text. As has been noted in previous studies, natural language
tends to show complex network structure at word level, with low degrees of
separation and scale-free (power law) degree distribution. There has also been
work on authorship attribution that incorporates ideas from complex networks.
The goal of our paper is to explore properties of these complex networks that
are suitable as features for machine-learning-based authorship attribution of
documents. We performed experiments on three different datasets, and obtained
promising results.
| 2,013 | Computation and Language |
Cornell SPF: Cornell Semantic Parsing Framework | The Cornell Semantic Parsing Framework (SPF) is a learning and inference
framework for mapping natural language to formal representation of its meaning.
| 2,016 | Computation and Language |
Architecture of an Ontology-Based Domain-Specific Natural Language
Question Answering System | Question answering (QA) system aims at retrieving precise information from a
large collection of documents against a query. This paper describes the
architecture of a Natural Language Question Answering (NLQA) system for a
specific domain based on the ontological information, a step towards semantic
web question answering. The proposed architecture defines four basic modules
suitable for enhancing current QA capabilities with the ability of processing
complex questions. The first module was the question processing, which analyses
and classifies the question and also reformulates the user query. The second
module allows the process of retrieving the relevant documents. The next module
processes the retrieved documents, and the last module performs the extraction
and generation of a response. Natural language processing techniques are used
for processing the question and documents and also for answer extraction.
Ontology and domain knowledge are used for reformulating queries and
identifying the relations. The aim of the system is to generate short and
specific answer to the question that is asked in the natural language in a
specific domain. We have achieved 94 % accuracy of natural language question
answering in our implementation.
| 2,013 | Computation and Language |
HEVAL: Yet Another Human Evaluation Metric | Machine translation evaluation is a very important activity in machine
translation development. Automatic evaluation metrics proposed in literature
are inadequate as they require one or more human reference translations to
compare them with output produced by machine translation. This does not always
give accurate results as a text can have several different translations. Human
evaluation metrics, on the other hand, lacks inter-annotator agreement and
repeatability. In this paper we have proposed a new human evaluation metric
which addresses these issues. Moreover this metric also provides solid grounds
for making sound assumptions on the quality of the text produced by a machine
translation.
| 2,013 | Computation and Language |
Big Data and Cross-Document Coreference Resolution: Current State and
Future Opportunities | Information Extraction (IE) is the task of automatically extracting
structured information from unstructured/semi-structured machine-readable
documents. Among various IE tasks, extracting actionable intelligence from
ever-increasing amount of data depends critically upon Cross-Document
Coreference Resolution (CDCR) - the task of identifying entity mentions across
multiple documents that refer to the same underlying entity. Recently, document
datasets of the order of peta-/tera-bytes has raised many challenges for
performing effective CDCR such as scaling to large numbers of mentions and
limited representational power. The problem of analysing such datasets is
called "big data". The aim of this paper is to provide readers with an
understanding of the central concepts, subtasks, and the current
state-of-the-art in CDCR process. We provide assessment of existing
tools/techniques for CDCR subtasks and highlight big data challenges in each of
them to help readers identify important and outstanding issues for further
investigation. Finally, we provide concluding remarks and discuss possible
directions for future work.
| 2,013 | Computation and Language |
Clustering and Relational Ambiguity: from Text Data to Natural Data | Text data is often seen as "take-away" materials with little noise and easy
to process information. Main questions are how to get data and transform them
into a good document format. But data can be sensitive to noise oftenly called
ambiguities. Ambiguities are aware from a long time, mainly because polysemy is
obvious in language and context is required to remove uncertainty. I claim in
this paper that syntactic context is not suffisant to improve interpretation.
In this paper I try to explain that firstly noise can come from natural data
themselves, even involving high technology, secondly texts, seen as verified
but meaningless, can spoil content of a corpus; it may lead to contradictions
and background noise.
| 2,014 | Computation and Language |
Complexity measurement of natural and artificial languages | We compared entropy for texts written in natural languages (English, Spanish)
and artificial languages (computer software) based on a simple expression for
the entropy as a function of message length and specific word diversity. Code
text written in artificial languages showed higher entropy than text of similar
length expressed in natural languages. Spanish texts exhibit more symbolic
diversity than English ones. Results showed that algorithms based on complexity
measures differentiate artificial from natural languages, and that text
analysis based on complexity measures allows the unveiling of important aspects
of their nature. We propose specific expressions to examine entropy related
aspects of tests and estimate the values of entropy, emergence,
self-organization and complexity based on specific diversity and message
length.
| 2,015 | Computation and Language |
Automatic Ranking of MT Outputs using Approximations | Since long, research on machine translation has been ongoing. Still, we do
not get good translations from MT engines so developed. Manual ranking of these
outputs tends to be very time consuming and expensive. Identifying which one is
better or worse than the others is a very taxing task. In this paper, we show
an approach which can provide automatic ranks to MT outputs (translations)
taken from different MT Engines and which is based on N-gram approximations. We
provide a solution where no human intervention is required for ranking systems.
Further we also show the evaluations of our results which show equivalent
results as that of human ranking.
| 2,013 | Computation and Language |
Build Electronic Arabic Lexicon | There are many known Arabic lexicons organized on different ways, each of
them has a different number of Arabic words according to its organization way.
This paper has used mathematical relations to count a number of Arabic words,
which proofs the number of Arabic words presented by Al Farahidy. The paper
also presents new way to build an electronic Arabic lexicon by using a hash
function that converts each word (as input) to correspond a unique integer
number (as output), these integer numbers will be used as an index to a lexicon
entry.
| 2,011 | Computation and Language |
NILE: Fast Natural Language Processing for Electronic Health Records | Objective: Narrative text in Electronic health records (EHR) contain rich
information for medical and data science studies. This paper introduces the
design and performance of Narrative Information Linear Extraction (NILE), a
natural language processing (NLP) package for EHR analysis that we share with
the medical informatics community. Methods: NILE uses a modified prefix-tree
search algorithm for named entity recognition, which can detect prefix and
suffix sharing. The semantic analyses are implemented as rule-based finite
state machines. Analyses include negation, location, modification, family
history, and ignoring. Result: The processing speed of NILE is hundreds to
thousands times faster than existing NLP software for medical text. The
accuracy of presence analysis of NILE is on par with the best performing models
on the 2010 i2b2/VA NLP challenge data. Conclusion: The speed, accuracy, and
being able to operate via API make NILE a valuable addition to the NLP software
for medical informatics and data science.
| 2,019 | Computation and Language |
Learning Semantic Representations for the Phrase Translation Model | This paper presents a novel semantic-based phrase translation model. A pair
of source and target phrases are projected into continuous-valued vector
representations in a low-dimensional latent semantic space, where their
translation score is computed by the distance between the pair in this new
space. The projection is performed by a multi-layer neural network whose
weights are learned on parallel training data. The learning is aimed to
directly optimize the quality of end-to-end machine translation results.
Experimental evaluation has been performed on two Europarl translation tasks,
English-French and German-English. The results show that the new semantic-based
phrase translation model significantly improves the performance of a
state-of-the-art phrase-based statistical machine translation sys-tem, leading
to a gain of 0.7-1.0 BLEU points.
| 2,013 | Computation and Language |
Towards Structural Natural Language Formalization: Mapping Discourse to
Controlled Natural Language | The author describes a conceptual study towards mapping grounded natural
language discourse representation structures to instances of controlled
language statements. This can be achieved via a pipeline of preexisting state
of the art technologies, namely natural language syntax to semantic discourse
mapping, and a reduction of the latter to controlled language discourse, given
a set of previously learnt reduction rules. Concludingly a description on
evaluation, potential and limitations for ontology-based reasoning is
presented.
| 2,013 | Computation and Language |
Time-dependent Hierarchical Dirichlet Model for Timeline Generation | Timeline Generation aims at summarizing news from different epochs and
telling readers how an event evolves. It is a new challenge that combines
salience ranking with novelty detection. For long-term public events, the main
topic usually includes various aspects across different epochs and each aspect
has its own evolving pattern. Existing approaches neglect such hierarchical
topic structure involved in the news corpus in timeline generation. In this
paper, we develop a novel time-dependent Hierarchical Dirichlet Model (HDM) for
timeline generation. Our model can aptly detect different levels of topic
information across corpus and such structure is further used for sentence
selection. Based on the topic mined fro HDM, sentences are selected by
considering different aspects such as relevance, coherence and coverage. We
develop experimental systems to evaluate 8 long-term events that public
concern. Performance comparison between different systems demonstrates the
effectiveness of our model in terms of ROUGE metrics.
| 2,015 | Computation and Language |
One Billion Word Benchmark for Measuring Progress in Statistical
Language Modeling | We propose a new benchmark corpus to be used for measuring progress in
statistical language modeling. With almost one billion words of training data,
we hope this benchmark will be useful to quickly evaluate novel language
modeling techniques, and to compare their contribution when combined with other
advanced techniques. We show performance of several well-known types of
language models, with the best results achieved with a recurrent neural network
based language model. The baseline unpruned Kneser-Ney 5-gram model achieves
perplexity 67.6; a combination of techniques leads to 35% reduction in
perplexity, or 10% reduction in cross-entropy (bits), over that baseline.
The benchmark is available as a code.google.com project; besides the scripts
needed to rebuild the training/held-out data, it also makes available
log-probability values for each word in each of ten held-out data sets, for
each of the baseline n-gram models.
| 2,014 | Computation and Language |
Semantic Types, Lexical Sorts and Classifiers | We propose a cognitively and linguistically motivated set of sorts for
lexical semantics in a compositional setting: the classifiers in languages that
do have such pronouns. These sorts are needed to include lexical considerations
in a semantical analyser such as Boxer or Grail. Indeed, all proposed lexical
extensions of usual Montague semantics to model restriction of selection,
felicitous and infelicitous copredication require a rich and refined type
system whose base types are the lexical sorts, the basis of the many-sorted
logic in which semantical representations of sentences are stated. However,
none of those approaches define precisely the actual base types or sorts to be
used in the lexicon. In this article, we shall discuss some of the options
commonly adopted by researchers in formal lexical semantics, and defend the
view that classifiers in the languages which have such pronouns are an
appealing solution, both linguistically and cognitively motivated.
| 2,013 | Computation and Language |
Towards The Development of a Bishnupriya Manipuri Corpus | For any deep computational processing of language we need evidences, and one
such set of evidences is corpus. This paper describes the development of a
text-based corpus for the Bishnupriya Manipuri language. A Corpus is considered
as a building block for any language processing tasks. Due to the lack of
awareness like other Indian languages, it is also studied less frequently. As a
result the language still lacks a good corpus and basic language processing
tools. As per our knowledge this is the first effort to develop a corpus for
Bishnupriya Manipuri language.
| 2,013 | Computation and Language |
Implicit Sensitive Text Summarization based on Data Conveyed by
Connectives | So far and trying to reach human capabilities, research in automatic
summarization has been based on hypothesis that are both enabling and limiting.
Some of these limitations are: how to take into account and reflect (in the
generated summary) the implicit information conveyed in the text, the author
intention, the reader intention, the context influence, the general world
knowledge. Thus, if we want machines to mimic human abilities, then they will
need access to this same large variety of knowledge. The implicit is affecting
the orientation and the argumentation of the text and consequently its summary.
Most of Text Summarizers (TS) are processing as compressing the initial data
and they necessarily suffer from information loss. TS are focusing on features
of the text only, not on what the author intended or why the reader is reading
the text. In this paper, we address this problem and we present a system
focusing on acquiring knowledge that is implicit. We principally spotlight the
implicit information conveyed by the argumentative connectives such as: but,
even, yet and their effect on the summary.
| 2,013 | Computation and Language |
Domain adaptation for sequence labeling using hidden Markov models | Most natural language processing systems based on machine learning are not
robust to domain shift. For example, a state-of-the-art syntactic dependency
parser trained on Wall Street Journal sentences has an absolute drop in
performance of more than ten points when tested on textual data from the Web.
An efficient solution to make these methods more robust to domain shift is to
first learn a word representation using large amounts of unlabeled data from
both domains, and then use this representation as features in a supervised
learning algorithm. In this paper, we propose to use hidden Markov models to
learn word representations for part-of-speech tagging. In particular, we study
the influence of using data from the source, the target or both domains to
learn the representation and the different ways to represent words using an
HMM.
| 2,013 | Computation and Language |
Deep Learning Embeddings for Discontinuous Linguistic Units | Deep learning embeddings have been successfully used for many natural
language processing problems. Embeddings are mostly computed for word forms
although a number of recent papers have extended this to other linguistic units
like morphemes and phrases. In this paper, we argue that learning embeddings
for discontinuous linguistic units should also be considered. In an
experimental evaluation on coreference resolution, we show that such embeddings
perform better than word form embeddings.
| 2,013 | Computation and Language |
Word Emdeddings through Hellinger PCA | Word embeddings resulting from neural language models have been shown to be
successful for a large variety of NLP tasks. However, such architecture might
be difficult to train and time-consuming. Instead, we propose to drastically
simplify the word embeddings computation through a Hellinger PCA of the word
co-occurence matrix. We compare those new word embeddings with some well-known
embeddings on NER and movie review tasks and show that we can reach similar or
even better performance. Although deep learning is not really necessary for
generating good word embeddings, we show that it can provide an easy way to
adapt embeddings to specific tasks.
| 2,014 | Computation and Language |
Distributional Models and Deep Learning Embeddings: Combining the Best
of Both Worlds | There are two main approaches to the distributed representation of words:
low-dimensional deep learning embeddings and high-dimensional distributional
models, in which each dimension corresponds to a context word. In this paper,
we combine these two approaches by learning embeddings based on
distributional-model vectors - as opposed to one-hot vectors as is standardly
done in deep learning. We show that the combined approach has better
performance on a word relatedness judgment task.
| 2,014 | Computation and Language |
Learning Type-Driven Tensor-Based Meaning Representations | This paper investigates the learning of 3rd-order tensors representing the
semantics of transitive verbs. The meaning representations are part of a
type-driven tensor-based semantic framework, from the newly emerging field of
compositional distributional semantics. Standard techniques from the neural
networks literature are used to learn the tensors, which are tested on a
selectional preference-style task with a simple 2-dimensional sentence space.
Promising results are obtained against a competitive corpus-based baseline. We
argue that extending this work beyond transitive verbs, and to
higher-dimensional sentence spaces, is an interesting and challenging problem
for the machine learning community to consider.
| 2,014 | Computation and Language |
Multilingual Distributed Representations without Word Alignment | Distributed representations of meaning are a natural way to encode covariance
relationships between words and phrases in NLP. By overcoming data sparsity
problems, as well as providing information about semantic relatedness which is
not available in discrete representations, distributed representations have
proven useful in many NLP tasks. Recent work has shown how compositional
semantic representations can successfully be applied to a number of monolingual
applications such as sentiment analysis. At the same time, there has been some
initial success in work on learning shared word-level representations across
languages. We combine these two approaches by proposing a method for learning
distributed representations in a multilingual setup. Our model learns to assign
similar embeddings to aligned sentences and dissimilar ones to sentence which
are not aligned while not requiring word alignments. We show that our
representations are semantically informative and apply them to a cross-lingual
document classification task where we outperform the previous state of the art.
Further, by employing parallel corpora of multiple language pairs we find that
our model learns representations that capture semantic relationships across
languages for which no parallel data was used.
| 2,014 | Computation and Language |
Can recursive neural tensor networks learn logical reasoning? | Recursive neural network models and their accompanying vector representations
for words have seen success in an array of increasingly semantically
sophisticated tasks, but almost nothing is known about their ability to
accurately capture the aspects of linguistic meaning that are necessary for
interpretation or reasoning. To evaluate this, I train a recursive model on a
new corpus of constructed examples of logical reasoning in short sentences,
like the inference of "some animal walks" from "some dog walks" or "some cat
walks," given that dogs and cats are animals. This model learns representations
that generalize well to new types of reasoning pattern in all but a few cases,
a result which is promising for the ability of learned representation models to
capture logical reasoning.
| 2,014 | Computation and Language |
Speech Recognition Front End Without Information Loss | Speech representation and modelling in high-dimensional spaces of acoustic
waveforms, or a linear transformation thereof, is investigated with the aim of
improving the robustness of automatic speech recognition to additive noise. The
motivation behind this approach is twofold: (i) the information in acoustic
waveforms that is usually removed in the process of extracting low-dimensional
features might aid robust recognition by virtue of structured redundancy
analogous to channel coding, (ii) linear feature domains allow for exact noise
adaptation, as opposed to representations that involve non-linear processing
which makes noise adaptation challenging. Thus, we develop a generative
framework for phoneme modelling in high-dimensional linear feature domains, and
use it in phoneme classification and recognition tasks. Results show that
classification and recognition in this framework perform better than analogous
PLP and MFCC classifiers below 18 dB SNR. A combination of the high-dimensional
and MFCC features at the likelihood level performs uniformly better than either
of the individual representations across all noise levels.
| 2,015 | Computation and Language |
Formal Ontology Learning on Factual IS-A Corpus in English using
Description Logics | Ontology Learning (OL) is the computational task of generating a knowledge
base in the form of an ontology given an unstructured corpus whose content is
in natural language (NL). Several works can be found in this area most of which
are limited to statistical and lexico-syntactic pattern matching based
techniques Light-Weight OL. These techniques do not lead to very accurate
learning mostly because of several linguistic nuances in NL. Formal OL is an
alternative (less explored) methodology were deep linguistics analysis is made
using theory and tools found in computational linguistics to generate formal
axioms and definitions instead simply inducing a taxonomy. In this paper we
propose "Description Logic (DL)" based formal OL framework for learning factual
IS-A type sentences in English. We claim that semantic construction of IS-A
sentences is non trivial. Hence, we also claim that such sentences requires
special studies in the context of OL before any truly formal OL can be
proposed. We introduce a learner tool, called DLOL_IS-A, that generated such
ontologies in the owl format. We have adopted "Gold Standard" based OL
evaluation on IS-A rich WCL v.1.1 dataset and our own Community representative
IS-A dataset. We observed significant improvement of DLOL_IS-A when compared to
the light-weight OL tool Text2Onto and formal OL tool FRED.
| 2,016 | Computation and Language |
Description Logics based Formalization of Wh-Queries | The problem of Natural Language Query Formalization (NLQF) is to translate a
given user query in natural language (NL) into a formal language so that the
semantic interpretation has equivalence with the NL interpretation.
Formalization of NL queries enables logic based reasoning during information
retrieval, database query, question-answering, etc. Formalization also helps in
Web query normalization and indexing, query intent analysis, etc. In this paper
we are proposing a Description Logics based formal methodology for wh-query
intent (also called desire) identification and corresponding formal
translation. We evaluated the scalability of our proposed formalism using
Microsoft Encarta 98 query dataset and OWL-S TC v.4.0 dataset.
| 2,013 | Computation and Language |
Language Modeling with Power Low Rank Ensembles | We present power low rank ensembles (PLRE), a flexible framework for n-gram
language modeling where ensembles of low rank matrices and tensors are used to
obtain smoothed probability estimates of words in context. Our method can be
understood as a generalization of n-gram modeling to non-integer n, and
includes standard techniques such as absolute discounting and Kneser-Ney
smoothing as special cases. PLRE training is efficient and our approach
outperforms state-of-the-art modified Kneser Ney baselines in terms of
perplexity on large corpora as well as on BLEU score in a downstream machine
translation task.
| 2,014 | Computation and Language |
Quality Estimation of English-Hindi Outputs using Naive Bayes Classifier | In this paper we present an approach for estimating the quality of machine
translation system. There are various methods for estimating the quality of
output sentences, but in this paper we focus on Na\"ive Bayes classifier to
build model using features which are extracted from the input sentences. These
features are used for finding the likelihood of each of the sentences of the
training data which are then further used for determining the scores of the
test data. On the basis of these scores we determine the class labels of the
test data.
| 2,013 | Computation and Language |
Zero-Shot Learning for Semantic Utterance Classification | We propose a novel zero-shot learning method for semantic utterance
classification (SUC). It learns a classifier $f: X \to Y$ for problems where
none of the semantic categories $Y$ are present in the training set. The
framework uncovers the link between categories and utterances using a semantic
space. We show that this semantic space can be learned by deep neural networks
trained on large amounts of search engine query log data. More precisely, we
propose a novel method that can learn discriminative semantic features without
supervision. It uses the zero-shot learning framework to guide the learning of
the semantic features. We demonstrate the effectiveness of the zero-shot
semantic learning algorithm on the SUC dataset collected by (Tur, 2012).
Furthermore, we achieve state-of-the-art results by combining the semantic
features with a supervised method.
| 2,014 | Computation and Language |
Natural Language Processing in Biomedicine: A Unified System
Architecture Overview | In modern electronic medical records (EMR) much of the clinically important
data - signs and symptoms, symptom severity, disease status, etc. - are not
provided in structured data fields, but rather are encoded in clinician
generated narrative text. Natural language processing (NLP) provides a means of
"unlocking" this important data source for applications in clinical decision
support, quality assurance, and public health. This chapter provides an
overview of representative NLP systems in biomedicine based on a unified
architectural view. A general architecture in an NLP system consists of two
main components: background knowledge that includes biomedical knowledge
resources and a framework that integrates NLP tools to process text. Systems
differ in both components, which we will review briefly. Additionally,
challenges facing current research efforts in biomedical NLP include the
paucity of large, publicly available annotated corpora, although initiatives
that facilitate data sharing, system evaluation, and collaborative work between
researchers in clinical NLP are starting to emerge.
| 2,014 | Computation and Language |
Multi-Topic Multi-Document Summarizer | Current multi-document summarization systems can successfully extract summary
sentences, however with many limitations including: low coverage, inaccurate
extraction to important sentences, redundancy and poor coherence among the
selected sentences. The present study introduces a new concept of centroid
approach and reports new techniques for extracting summary sentences for
multi-document. In both techniques keyphrases are used to weigh sentences and
documents. The first summarization technique (Sen-Rich) prefers maximum
richness sentences. While the second (Doc-Rich), prefers sentences from
centroid document. To demonstrate the new summarization system application to
extract summaries of Arabic documents we performed two experiments. First, we
applied Rouge measure to compare the new techniques among systems presented at
TAC2011. The results show that Sen-Rich outperformed all systems in ROUGE-S.
Second, the system was applied to summarize multi-topic documents. Using human
evaluators, the results show that Doc-Rich is the superior, where summary
sentences characterized by extra coverage and more cohesion.
| 2,013 | Computation and Language |
Plurals: individuals and sets in a richly typed semantics | We developed a type-theoretical framework for natural lan- guage semantics
that, in addition to the usual Montagovian treatment of compositional
semantics, includes a treatment of some phenomena of lex- ical semantic:
coercions, meaning, transfers, (in)felicitous co-predication. In this setting
we see how the various readings of plurals (collective, dis- tributive,
coverings,...) can be modelled.
| 2,013 | Computation and Language |
Quantitative methods for Phylogenetic Inference in Historical
Linguistics: An experimental case study of South Central Dravidian | In this paper we examine the usefulness of two classes of algorithms Distance
Methods, Discrete Character Methods (Felsenstein and Felsenstein 2003) widely
used in genetics, for predicting the family relationships among a set of
related languages and therefore, diachronic language change. Applying these
algorithms to the data on the numbers of shared cognates- with-change and
changed as well as unchanged cognates for a group of six languages belonging to
a Dravidian language sub-family given in Krishnamurti et al. (1983), we
observed that the resultant phylogenetic trees are largely in agreement with
the linguistic family tree constructed using the comparative method of
reconstruction with only a few minor differences. Furthermore, we studied these
minor differences and found that they were cases of genuine ambiguity even for
a well-trained historical linguist. We evaluated the trees obtained through our
experiments using a well-defined criterion and report the results here. We
finally conclude that quantitative methods like the ones we examined are quite
useful in predicting family relationships among languages. In addition, we
conclude that a modest degree of confidence attached to the intuition that
there could indeed exist a parallelism between the processes of linguistic and
genetic change is not totally misplaced.
| 2,009 | Computation and Language |
Properties of phoneme N -grams across the world's language families | In this article, we investigate the properties of phoneme N-grams across half
of the world's languages. We investigate if the sizes of three different N-gram
distributions of the world's language families obey a power law. Further, the
N-gram distributions of language families parallel the sizes of the families,
which seem to obey a power law distribution. The correlation between N-gram
distributions and language family sizes improves with increasing values of N.
We applied statistical tests, originally given by physicists, to test the
hypothesis of power law fit to twelve different datasets. The study also raises
some new questions about the use of N-gram distributions in linguistic
research, which we answer by running a statistical test.
| 2,014 | Computation and Language |
Effective Slot Filling Based on Shallow Distant Supervision Methods | Spoken Language Systems at Saarland University (LSV) participated this year
with 5 runs at the TAC KBP English slot filling track. Effective algorithms for
all parts of the pipeline, from document retrieval to relation prediction and
response post-processing, are bundled in a modular end-to-end relation
extraction system called RelationFactory. The main run solely focuses on
shallow techniques and achieved significant improvements over LSV's last year's
system, while using the same training data and patterns. Improvements mainly
have been obtained by a feature representation focusing on surface skip n-grams
and improved scoring for extracted distant supervision patterns. Important
factors for effective extraction are the training and tuning scheme for distant
supervision classifiers, and the query expansion by a translation model based
on Wikipedia links. In the TAC KBP 2013 English Slotfilling evaluation, the
submitted main run of the LSV RelationFactory system achieved the top-ranked
F1-score of 37.3%.
| 2,014 | Computation and Language |
Learning Multilingual Word Representations using a Bag-of-Words
Autoencoder | Recent work on learning multilingual word representations usually relies on
the use of word-level alignements (e.g. infered with the help of GIZA++)
between translated sentences, in order to align the word embeddings in
different languages. In this workshop paper, we investigate an autoencoder
model for learning multilingual word representations that does without such
word-level alignements. The autoencoder is trained to reconstruct the
bag-of-word representation of given sentence from an encoded representation
extracted from its translation. We evaluate our approach on a multilingual
document classification task, where labeled data is available only for one
language (e.g. English) while classification must be performed in a different
language (e.g. French). In our experiments, we observe that our method compares
favorably with a previously proposed method that exploits word-level alignments
to learn word representations.
| 2,014 | Computation and Language |
The semantic similarity ensemble | Computational measures of semantic similarity between geographic terms
provide valuable support across geographic information retrieval, data mining,
and information integration. To date, a wide variety of approaches to
geo-semantic similarity have been devised. A judgment of similarity is not
intrinsically right or wrong, but obtains a certain degree of cognitive
plausibility, depending on how closely it mimics human behavior. Thus selecting
the most appropriate measure for a specific task is a significant challenge. To
address this issue, we make an analogy between computational similarity
measures and soliciting domain expert opinions, which incorporate a subjective
set of beliefs, perceptions, hypotheses, and epistemic biases. Following this
analogy, we define the semantic similarity ensemble (SSE) as a composition of
different similarity measures, acting as a panel of experts having to reach a
decision on the semantic similarity of a set of geographic terms. The approach
is evaluated in comparison to human judgments, and results indicate that an SSE
performs better than the average of its parts. Although the best member tends
to outperform the ensemble, all ensembles outperform the average performance of
each ensemble's member. Hence, in contexts where the best measure is unknown,
the ensemble provides a more cognitively plausible approach.
| 2,013 | Computation and Language |
Towards a Generic Framework for the Development of Unicode Based Digital
Sindhi Dictionaries | Dictionaries are essence of any language providing vital linguistic recourse
for the language learners, researchers and scholars. This paper focuses on the
methodology and techniques used in developing software architecture for a
UBSESD (Unicode Based Sindhi to English and English to Sindhi Dictionary). The
proposed system provides an accurate solution for construction and
representation of Unicode based Sindhi characters in a dictionary implementing
Hash Structure algorithm and a custom java Object as its internal data
structure saved in a file. The System provides facilities for Insertion,
Deletion and Editing of new records of Sindhi. Through this framework any type
of Sindhi to English and English to Sindhi Dictionary (belonging to different
domains of knowledge, e.g. engineering, medicine, computer, biology etc.) could
be developed easily with accurate representation of Unicode Characters in font
independent manner.
| 2,012 | Computation and Language |
Dictionary-Based Concept Mining: An Application for Turkish | In this study, a dictionary-based method is used to extract expressive
concepts from documents. So far, there have been many studies concerning
concept mining in English, but this area of study for Turkish, an agglutinative
language, is still immature. We used dictionary instead of WordNet, a lexical
database grouping words into synsets that is widely used for concept
extraction. The dictionaries are rarely used in the domain of concept mining,
but taking into account that dictionary entries have synonyms, hypernyms,
hyponyms and other relationships in their meaning texts, the success rate has
been high for determining concepts. This concept extraction method is
implemented on documents, that are collected from different corpora.
| 2,014 | Computation and Language |
A survey of methods to ease the development of highly multilingual text
mining applications | Multilingual text processing is useful because the information content found
in different languages is complementary, both regarding facts and opinions.
While Information Extraction and other text mining software can, in principle,
be developed for many languages, most text analysis tools have only been
applied to small sets of languages because the development effort per language
is large. Self-training tools obviously alleviate the problem, but even the
effort of providing training data and of manually tuning the results is usually
considerable. In this paper, we gather insights by various multilingual system
developers on how to minimise the effort of developing natural language
processing applications for many languages. We also explain the main guidelines
underlying our own effort to develop complex text mining software for tens of
languages. While these guidelines - most of all: extreme simplicity - can be
very restrictive and limiting, we believe to have shown the feasibility of the
approach through the development of the Europe Media Monitor (EMM) family of
applications (http://emm.newsbrief.eu/overview.html). EMM is a set of complex
media monitoring tools that process and analyse up to 100,000 online news
articles per day in between twenty and fifty languages. We will also touch upon
the kind of language resources that would make it easier for all to develop
highly multilingual text mining applications. We will argue that - to achieve
this - the most needed resources would be freely available, simple, parallel
and uniform multilingual dictionaries, corpora and software tools.
| 2,012 | Computation and Language |
ONTS: "Optima" News Translation System | We propose a real-time machine translation system that allows users to select
a news category and to translate the related live news articles from Arabic,
Czech, Danish, Farsi, French, German, Italian, Polish, Portuguese, Spanish and
Turkish into English. The Moses-based system was optimised for the news domain
and differs from other available systems in four ways: (1) News items are
automatically categorised on the source side, before translation; (2) Named
entity translation is optimised by recognising and extracting them on the
source side and by re-inserting their translation in the target language,
making use of a separate entity repository; (3) News titles are translated with
a separate translation system which is optimised for the specific style of news
titles; (4) The system was optimised for speed in order to cope with the large
volume of daily news articles.
| 2,012 | Computation and Language |
Optimization Of Cross Domain Sentiment Analysis Using Sentiwordnet | The task of sentiment analysis of reviews is carried out using manually built
/ automatically generated lexicon resources of their own with which terms are
matched with lexicon to compute the term count for positive and negative
polarity. On the other hand the Sentiwordnet, which is quite different from
other lexicon resources that gives scores (weights) of the positive and
negative polarity for each word. The polarity of a word namely positive,
negative and neutral have the score ranging between 0 to 1 indicates the
strength/weight of the word with that sentiment orientation. In this paper, we
show that using the Sentiwordnet, how we could enhance the performance of the
classification at both sentence and document level.
| 2,013 | Computation and Language |
A Subband-Based SVM Front-End for Robust ASR | This work proposes a novel support vector machine (SVM) based robust
automatic speech recognition (ASR) front-end that operates on an ensemble of
the subband components of high-dimensional acoustic waveforms. The key issues
of selecting the appropriate SVM kernels for classification in frequency
subbands and the combination of individual subband classifiers using ensemble
methods are addressed. The proposed front-end is compared with state-of-the-art
ASR front-ends in terms of robustness to additive noise and linear filtering.
Experiments performed on the TIMIT phoneme classification task demonstrate the
benefits of the proposed subband based SVM front-end: it outperforms the
standard cepstral front-end in the presence of noise and linear filtering for
signal-to-noise ratio (SNR) below 12-dB. A combination of the proposed
front-end with a conventional front-end such as MFCC yields further
improvements over the individual front ends across the full range of noise
levels.
| 2,014 | Computation and Language |
Learning Language from a Large (Unannotated) Corpus | A novel approach to the fully automated, unsupervised extraction of
dependency grammars and associated syntax-to-semantic-relationship mappings
from large text corpora is described. The suggested approach builds on the
authors' prior work with the Link Grammar, RelEx and OpenCog systems, as well
as on a number of prior papers and approaches from the statistical language
learning literature. If successful, this approach would enable the mining of
all the information needed to power a natural language comprehension and
generation system, directly from a large, unannotated corpus.
| 2,014 | Computation and Language |
Learning Document-Level Semantic Properties from Free-Text Annotations | This paper presents a new method for inferring the semantic properties of
documents by leveraging free-text keyphrase annotations. Such annotations are
becoming increasingly abundant due to the recent dramatic growth in
semi-structured, user-generated online content. One especially relevant domain
is product reviews, which are often annotated by their authors with pros/cons
keyphrases such as a real bargain or good value. These annotations are
representative of the underlying semantic properties; however, unlike expert
annotations, they are noisy: lay authors may use different labels to denote the
same property, and some labels may be missing. To learn using such noisy
annotations, we find a hidden paraphrase structure which clusters the
keyphrases. The paraphrase structure is linked with a latent topic model of the
review texts, enabling the system to predict the properties of unannotated
documents and to effectively aggregate the semantic properties of multiple
reviews. Our approach is implemented as a hierarchical Bayesian model with
joint inference. We find that joint inference increases the robustness of the
keyphrase clustering and encourages the latent topics to correlate with
semantically meaningful properties. Multiple evaluations demonstrate that our
model substantially outperforms alternative approaches for summarizing single
and multiple documents into a set of semantically salient keyphrases.
| 2,009 | Computation and Language |
Complex Question Answering: Unsupervised Learning Approaches and
Experiments | Complex questions that require inferencing and synthesizing information from
multiple documents can be seen as a kind of topic-oriented, informative
multi-document summarization where the goal is to produce a single text as a
compressed version of a set of documents with a minimum loss of relevant
information. In this paper, we experiment with one empirical method and two
unsupervised statistical machine learning techniques: K-means and Expectation
Maximization (EM), for computing relative importance of the sentences. We
compare the results of these approaches. Our experiments show that the
empirical approach outperforms the other two techniques and EM performs better
than K-means. However, the performance of these approaches depends entirely on
the feature set used and the weighting of these features. In order to measure
the importance and relevance to the user query we extract different kinds of
features (i.e. lexical, lexical semantic, cosine similarity, basic element,
tree kernel based syntactic and shallow-semantic) for each of the document
sentences. We use a local search technique to learn the weights of the
features. To the best of our knowledge, no study has used tree kernel functions
to encode syntactic/semantic information for more complex tasks such as
computing the relatedness between the query sentences and the document
sentences in order to generate query-focused summaries (or answers to complex
questions). For each of our methods of generating summaries (i.e. empirical,
K-means and EM) we show the effects of syntactic and shallow-semantic features
over the bag-of-words (BOW) features.
| 2,009 | Computation and Language |
Enhancing QA Systems with Complex Temporal Question Processing
Capabilities | This paper presents a multilayered architecture that enhances the
capabilities of current QA systems and allows different types of complex
questions or queries to be processed. The answers to these questions need to be
gathered from factual information scattered throughout different documents.
Specifically, we designed a specialized layer to process the different types of
temporal questions. Complex temporal questions are first decomposed into simple
questions, according to the temporal relations expressed in the original
question. In the same way, the answers to the resulting simple questions are
recomposed, fulfilling the temporal restrictions of the original complex
question. A novel aspect of this approach resides in the decomposition which
uses a minimal quantity of resources, with the final aim of obtaining a
portable platform that is easily extensible to other languages. In this paper
we also present a methodology for evaluation of the decomposition of the
questions as well as the ability of the implemented temporal layer to perform
at a multilingual level. The temporal layer was first performed for English,
then evaluated and compared with: a) a general purpose QA system (F-measure
65.47% for QA plus English temporal layer vs. 38.01% for the general QA
system), and b) a well-known QA system. Much better results were obtained for
temporal questions with the multilayered system. This system was therefore
extended to Spanish and very good results were again obtained in the evaluation
(F-measure 40.36% for QA plus Spanish temporal layer vs. 22.94% for the general
QA system).
| 2,009 | Computation and Language |