Titles
stringlengths 6
220
| Abstracts
stringlengths 37
3.26k
| Years
int64 1.99k
2.02k
| Categories
stringclasses 1
value |
---|---|---|---|
CHR Grammars | A grammar formalism based upon CHR is proposed analogously to the way
Definite Clause Grammars are defined and implemented on top of Prolog. These
grammars execute as robust bottom-up parsers with an inherent treatment of
ambiguity and a high flexibility to model various linguistic phenomena. The
formalism extends previous logic programming based grammars with a form of
context-sensitive rules and the possibility to include extra-grammatical
hypotheses in both head and body of grammar rules. Among the applications are
straightforward implementations of Assumption Grammars and abduction under
integrity constraints for language analysis. CHR grammars appear as a powerful
tool for specification and implementation of language processors and may be
proposed as a new standard for bottom-up grammars in logic programming.
To appear in Theory and Practice of Logic Programming (TPLP), 2005
| 2,007 | Computation and Language |
Multi-dimensional Type Theory: Rules, Categories, and Combinators for
Syntax and Semantics | We investigate the possibility of modelling the syntax and semantics of
natural language by constraints, or rules, imposed by the multi-dimensional
type theory Nabla. The only multiplicity we explicitly consider is two, namely
one dimension for the syntax and one dimension for the semantics, but the
general perspective is important. For example, issues of pragmatics could be
handled as additional dimensions.
One of the main problems addressed is the rather complicated repertoire of
operations that exists besides the notion of categories in traditional Montague
grammar. For the syntax we use a categorial grammar along the lines of Lambek.
For the semantics we use so-called lexical and logical combinators inspired by
work in natural logic. Nabla provides a concise interpretation and a sequent
calculus as the basis for implementations.
| 2,007 | Computation and Language |
Fractal geometry of literature: first attempt to Shakespeare's works | It was demonstrated that there is a geometrical order in the structure of
literature. Fractal geometry as a modern mathematical approach and a new
geometrical viewpoint on natural objects including both processes and
structures was employed for analysis of literature. As the first study, the
works of William Shakespeare were chosen as the most important items in western
literature. By counting the number of letters applied in a manuscript, it is
possible to study the whole manuscript statistically. A novel method based on
basic assumption of fractal geometry was proposed for the calculation of
fractal dimensions of the literature. The results were compared with Zipf's
law. Zipf's law was successfully used for letters instead of words. Two new
concepts namely Zipf's dimension and Zipf's order were also introduced. It was
found that changes of both fractal dimension and Zipf's dimension are similar
and dependent on the manuscript length. Interestingly, direct plotting the data
obtained in semi-logarithmic and logarithmic forms also led to a power-law.
| 2,007 | Computation and Language |
Application of the Double Metaphone Algorithm to Amharic Orthography | The Metaphone algorithm applies the phonetic encoding of orthographic
sequences to simplify words prior to comparison. While Metaphone has been
highly successful for the English language, for which it was designed, it may
not be applied directly to Ethiopian languages. The paper details how the
principles of Metaphone can be applied to Ethiopic script and uses Amharic as a
case study. Match results improve as specific considerations are made for
Amharic writing practices. Results are shown to improve further when common
errors from Amharic input methods are considered.
| 2,007 | Computation and Language |
The role of robust semantic analysis in spoken language dialogue systems | In this paper we summarized a framework for designing grammar-based procedure
for the automatic extraction of the semantic content from spoken queries.
Starting with a case study and following an approach which combines the notions
of fuzziness and robustness in sentence parsing, we showed we built practical
domain-dependent rules which can be applied whenever it is possible to
superimpose a sentence-level semantic structure to a text without relying on a
previous deep syntactical analysis. This kind of procedure can be also
profitably used as a pre-processing tool in order to cut out part of the
sentence which have been recognized to have no relevance in the understanding
process. In the case of particular dialogue applications where there is no need
to build a complex semantic structure (e.g. word spotting or excerpting) the
presented methodology may represent an efficient alternative solution to a
sequential composition of deep linguistic analysis modules. Even if the query
generation problem may not seem a critical application it should be held in
mind that the sentence processing must be done on-line. Having this kind of
constraints we cannot design our system without caring for efficiency and thus
provide an immediate response. Another critical issue is related to whole
robustness of the system. In our case study we tried to make experiences on how
it is possible to deal with an unreliable and noisy input without asking the
user for any repetition or clarification. This may correspond to a similar
problem one may have when processing text coming from informal writing such as
e-mails, news and in many cases Web pages where it is often the case to have
irrelevant surrounding information.
| 2,000 | Computation and Language |
Proofing Tools Technology at Neurosoft S.A. | The aim of this paper is to present the R&D activities carried out at
Neurosoft S.A. regarding the development of proofing tools for Modern Greek.
Firstly, we focus on infrastructure issues that we faced during our initial
steps. Subsequently, we describe the most important insights of three proofing
tools developed by Neurosoft, i.e. the spelling checker, the hyphenator and the
thesaurus, outlining their efficiencies and inefficiencies. Finally, we discuss
some improvement ideas and give our future directions.
| 2,007 | Computation and Language |
Verbal chunk extraction in French using limited resources | A way of extracting French verbal chunks, inflected and infinitive, is
explored and tested on effective corpus. Declarative morphological and local
grammar rules specifying chunks and some simple contextual structures are used,
relying on limited lexical information and some simple heuristic/statistic
properties obtained from restricted corpora. The specific goals, the
architecture and the formalism of the system, the linguistic information on
which it relies and the obtained results on effective corpus are presented.
| 2,007 | Computation and Language |
An electronic dictionary as a basis for NLP tools: The Greek case | The existence of a Dictionary in electronic form for Modern Greek (MG) is
mandatory if one is to process MG at the morphological and syntactic levels
since MG is a highly inflectional language with marked stress and a spelling
system with many characteristics carried over from Ancient Greek. Moreover,
such a tool becomes necessary if one is to create efficient and sophisticated
NLP applications with substantial linguistic backing and coverage. The present
paper will focus on the deployment of such an electronic dictionary for Modern
Greek, which was built in two phases: first it was constructed to be the basis
for a spelling correction schema and then it was reconstructed in order to
become the platform for the deployment of a wider spectrum of NLP tools.
| 2,007 | Computation and Language |
A Model for Fine-Grained Alignment of Multilingual Texts | While alignment of texts on the sentential level is often seen as being too
coarse, and word alignment as being too fine-grained, bi- or multilingual texts
which are aligned on a level in-between are a useful resource for many
purposes. Starting from a number of examples of non-literal translations, which
tend to make alignment difficult, we describe an alignment model which copes
with these cases by explicitly coding them. The model is based on
predicate-argument structures and thus covers the middle ground between
sentence and word alignment. The model is currently used in a recently
initiated project of a parallel English-German treebank (FuSe), which can in
principle be extended with additional languages.
| 2,004 | Computation and Language |
A Sentimental Education: Sentiment Analysis Using Subjectivity
Summarization Based on Minimum Cuts | Sentiment analysis seeks to identify the viewpoint(s) underlying a text span;
an example application is classifying a movie review as "thumbs up" or "thumbs
down". To determine this sentiment polarity, we propose a novel
machine-learning method that applies text-categorization techniques to just the
subjective portions of the document. Extracting these portions can be
implemented using efficient techniques for finding minimum cuts in graphs; this
greatly facilitates incorporation of cross-sentence contextual constraints.
| 2,004 | Computation and Language |
Robust Dialogue Understanding in HERALD | We tackle the problem of robust dialogue processing from the perspective of
language engineering. We propose an agent-oriented architecture that allows us
a flexible way of composing robust processors. Our approach is based on
Shoham's Agent Oriented Programming (AOP) paradigm. We will show how the AOP
agent model can be enriched with special features and components that allow us
to deal with classical problems of dialogue understanding.
| 2,001 | Computation and Language |
Semantic filtering by inference on domain knowledge in spoken dialogue
systems | General natural dialogue processing requires large amounts of domain
knowledge as well as linguistic knowledge in order to ensure acceptable
coverage and understanding. There are several ways of integrating lexical
resources (e.g. dictionaries, thesauri) and knowledge bases or ontologies at
different levels of dialogue processing. We concentrate in this paper on how to
exploit domain knowledge for filtering interpretation hypotheses generated by a
robust semantic parser. We use domain knowledge to semantically constrain the
hypothesis space. Moreover, adding an inference mechanism allows us to complete
the interpretation when information is not explicitly available. Further, we
discuss briefly how this can be generalized towards a predictive natural
interactive system.
| 2,000 | Computation and Language |
An argumentative annotation schema for meeting discussions | In this article, we are interested in the annotation of transcriptions of
human-human dialogue taken from meeting records. We first propose a meeting
content model where conversational acts are interpreted with respect to their
argumentative force and their role in building the argumentative structure of
the meeting discussion. Argumentation in dialogue describes the way
participants take part in the discussion and argue their standpoints. Then, we
propose an annotation scheme based on such an argumentative dialogue model as
well as the evaluation of its adequacy. The obtained higher-level semantic
annotations are exploited in the conceptual indexing of the information
contained in meeting discussions.
| 2,004 | Computation and Language |
Automatic Keyword Extraction from Spoken Text. A Comparison of two
Lexical Resources: the EDR and WordNet | Lexical resources such as WordNet and the EDR electronic dictionary have been
used in several NLP tasks. Probably, partly due to the fact that the EDR is not
freely available, WordNet has been used far more often than the EDR. We have
used both resources on the same task in order to make a comparison possible.
The task is automatic assignment of keywords to multi-party dialogue episodes
(i.e. thematically coherent stretches of spoken text). We show that the use of
lexical resources in such a task results in slightly higher performances than
the use of a purely statistically based method.
| 2,004 | Computation and Language |
Building Chinese Lexicons from Scratch by Unsupervised Short Document
Self-Segmentation | Chinese text segmentation is a well-known and difficult problem. On one side,
there is not a simple notion of "word" in Chinese language making really hard
to implement rule-based systems to segment written texts, thus lexicons and
statistical information are usually employed to achieve such a task. On the
other side, any piece of Chinese text usually includes segments present neither
in the lexicons nor in the training data. Even worse, such unseen sequences can
be segmented into a number of totally unrelated words making later processing
phases difficult. For instance, using a lexicon-based system the sequence
???(Baluozuo, Barroso, current president-designate of the European Commission)
can be segmented into ?(ba, to hope, to wish) and ??(luozuo, an undefined word)
changing completely the meaning of the sentence. A new and extremely simple
algorithm specially suited to work over short Chinese documents is introduced.
This new algorithm performs text "self-segmentation" producing results
comparable to those achieved by native speakers without using either lexicons
or any statistical information beyond the obtained from the input text.
Furthermore, it is really robust for finding new "words", especially proper
nouns, and it is well suited to build lexicons from scratch. Some preliminary
results are provided in addition to examples of its employment.
| 2,007 | Computation and Language |
A Tutorial on the Expectation-Maximization Algorithm Including
Maximum-Likelihood Estimation and EM Training of Probabilistic Context-Free
Grammars | The paper gives a brief review of the expectation-maximization algorithm
(Dempster 1977) in the comprehensible framework of discrete mathematics. In
Section 2, two prominent estimation methods, the relative-frequency estimation
and the maximum-likelihood estimation are presented. Section 3 is dedicated to
the expectation-maximization algorithm and a simpler variant, the generalized
expectation-maximization algorithm. In Section 4, two loaded dice are rolled. A
more interesting example is presented in Section 5: The estimation of
probabilistic context-free grammars.
| 2,007 | Computation and Language |
Inside-Outside Estimation Meets Dynamic EM | We briefly review the inside-outside and EM algorithm for probabilistic
context-free grammars. As a result, we formally prove that inside-outside
estimation is a dynamic-programming variant of EM. This is interesting in its
own right, but even more when considered in a theoretical context since the
well-known convergence behavior of inside-outside estimation has been confirmed
by many experiments but apparently has never been formally proved. However,
being a version of EM, inside-outside estimation also inherits the good
convergence behavior of EM. Therefore, the as yet imperfect line of
argumentation can be transformed into a coherent proof.
| 2,001 | Computation and Language |
Human-Level Performance on Word Analogy Questions by Latent Relational
Analysis | This paper introduces Latent Relational Analysis (LRA), a method for
measuring relational similarity. LRA has potential applications in many areas,
including information extraction, word sense disambiguation, machine
translation, and information retrieval. Relational similarity is correspondence
between relations, in contrast with attributional similarity, which is
correspondence between attributes. When two words have a high degree of
attributional similarity, we call them synonyms. When two pairs of words have a
high degree of relational similarity, we say that their relations are
analogous. For example, the word pair mason/stone is analogous to the pair
carpenter/wood. Past work on semantic similarity measures has mainly been
concerned with attributional similarity. Recently the Vector Space Model (VSM)
of information retrieval has been adapted to the task of measuring relational
similarity, achieving a score of 47% on a collection of 374 college-level
multiple-choice word analogy questions. In the VSM approach, the relation
between a pair of words is characterized by a vector of frequencies of
predefined patterns in a large corpus. LRA extends the VSM approach in three
ways: (1) the patterns are derived automatically from the corpus (they are not
predefined), (2) the Singular Value Decomposition (SVD) is used to smooth the
frequency data (it is also used this way in Latent Semantic Analysis), and (3)
automatically generated synonyms are used to explore reformulations of the word
pairs. LRA achieves 56% on the 374 analogy questions, statistically equivalent
to the average human score of 57%. On the related problem of classifying
noun-modifier relations, LRA achieves similar gains over the VSM, while using a
smaller corpus.
| 2,007 | Computation and Language |
A Framework for Creating Natural Language User Interfaces for
Action-Based Applications | In this paper we present a framework for creating natural language interfaces
to action-based applications. Our framework uses a number of reusable
application-independent components, in order to reduce the effort of creating a
natural language interface for a given application. Using a type-logical
grammar, we first translate natural language sentences into expressions in an
extended higher-order logic. These expressions can be seen as executable
specifications corresponding to the original sentences. The executable
specifications are then interpreted by invoking appropriate procedures provided
by the application for which a natural language interface is being created.
| 2,007 | Computation and Language |
The Google Similarity Distance | Words and phrases acquire meaning from the way they are used in society, from
their relative semantics to other words and phrases. For computers the
equivalent of `society' is `database,' and the equivalent of `use' is `way to
search the database.' We present a new theory of similarity between words and
phrases based on information distance and Kolmogorov complexity. To fix
thoughts we use the world-wide-web as database, and Google as search engine.
The method is also applicable to other search engines and databases. This
theory is then applied to construct a method to automatically extract
similarity, the Google similarity distance, of words and phrases from the
world-wide-web using Google page counts. The world-wide-web is the largest
database on earth, and the context information entered by millions of
independent users averages out to provide automatic semantics of useful
quality. We give applications in hierarchical clustering, classification, and
language translation. We give examples to distinguish between colors and
numbers, cluster names of paintings by 17th century Dutch masters and names of
books by English novelists, the ability to understand emergencies, and primes,
and we demonstrate the ability to do a simple automatic English-Spanish
translation. Finally, we use the WordNet database as an objective baseline
against which to judge the performance of our method. We conduct a massive
randomized trial in binary classification using support vector machines to
learn categories based on our Google distance, resulting in an a mean agreement
of 87% with the expert crafted WordNet categories.
| 2,007 | Computation and Language |
State of the Art, Evaluation and Recommendations regarding "Document
Processing and Visualization Techniques" | Several Networks of Excellence have been set up in the framework of the
European FP5 research program. Among these Networks of Excellence, the NEMIS
project focuses on the field of Text Mining.
Within this field, document processing and visualization was identified as
one of the key topics and the WG1 working group was created in the NEMIS
project, to carry out a detailed survey of techniques associated with the text
mining process and to identify the relevant research topics in related research
areas.
In this document we present the results of this comprehensive survey. The
report includes a description of the current state-of-the-art and practice, a
roadmap for follow-up research in the identified areas, and recommendations for
anticipated technological development in the domain of text mining.
| 2,007 | Computation and Language |
Thematic Annotation: extracting concepts out of documents | Contrarily to standard approaches to topic annotation, the technique used in
this work does not centrally rely on some sort of -- possibly statistical --
keyword extraction. In fact, the proposed annotation algorithm uses a large
scale semantic database -- the EDR Electronic Dictionary -- that provides a
concept hierarchy based on hyponym and hypernym relations. This concept
hierarchy is used to generate a synthetic representation of the document by
aggregating the words present in topically homogeneous document segments into a
set of concepts best preserving the document's content.
This new extraction technique uses an unexplored approach to topic selection.
Instead of using semantic similarity measures based on a semantic resource, the
later is processed to extract the part of the conceptual hierarchy relevant to
the document content. Then this conceptual hierarchy is searched to extract the
most relevant set of concepts to represent the topics discussed in the
document. Notice that this algorithm is able to extract generic concepts that
are not directly present in the document.
| 2,007 | Computation and Language |
Multi-document Biography Summarization | In this paper we describe a biography summarization system using sentence
classification and ideas from information retrieval. Although the individual
techniques are not new, assembling and applying them to generate multi-document
biographies is new. Our system was evaluated in DUC2004. It is among the top
performers in task 5-short summaries focused by person questions.
| 2,004 | Computation and Language |
An Introduction to the Summarization of Evolving Events: Linear and
Non-linear Evolution | This paper examines the summarization of events that evolve through time. It
discusses different types of evolution taking into account the time in which
the incidents of an event are happening and the different sources reporting on
the specific event. It proposes an approach for multi-document summarization
which employs ``messages'' for representing the incidents of an event and
cross-document relations that hold between messages according to certain
conditions. The paper also outlines the current version of the summarization
system we are implementing to realize this approach.
| 2,005 | Computation and Language |
Weighted Automata in Text and Speech Processing | Finite-state automata are a very effective tool in natural language
processing. However, in a variety of applications and especially in speech
precessing, it is necessary to consider more general machines in which arcs are
assigned weights or costs. We briefly describe some of the main theoretical and
algorithmic aspects of these machines. In particular, we describe an efficient
composition algorithm for weighted transducers, and give examples illustrating
the value of determinization and minimization algorithms for weighted automata.
| 1,996 | Computation and Language |
A Matter of Opinion: Sentiment Analysis and Business Intelligence
(position paper) | A general-audience introduction to the area of "sentiment analysis", the
computational treatment of subjective, opinion-oriented language (an example
application is determining whether a review is "thumbs up" or "thumbs down").
Some challenges, applications to business-intelligence tasks, and potential
future directions are described.
| 2,004 | Computation and Language |
Summarization from Medical Documents: A Survey | Objective:
The aim of this paper is to survey the recent work in medical documents
summarization.
Background:
During the last decade, documents summarization got increasing attention by
the AI research community. More recently it also attracted the interest of the
medical research community as well, due to the enormous growth of information
that is available to the physicians and researchers in medicine, through the
large and growing number of published journals, conference proceedings, medical
sites and portals on the World Wide Web, electronic medical records, etc.
Methodology:
This survey gives first a general background on documents summarization,
presenting the factors that summarization depends upon, discussing evaluation
issues and describing briefly the various types of summarization techniques. It
then examines the characteristics of the medical domain through the different
types of medical documents. Finally, it presents and discusses the
summarization techniques used so far in the medical domain, referring to the
corresponding systems and their characteristics.
Discussion and conclusions:
The paper discusses thoroughly the promising paths for future research in
medical documents summarization. It mainly focuses on the issue of scaling to
large collections of documents in various languages and from different media,
on personalization issues, on portability to new sub-domains, and on the
integration of summarization technology in practical applications
| 2,005 | Computation and Language |
Metalinguistic Information Extraction for Terminology | This paper describes and evaluates the Metalinguistic Operation Processor
(MOP) system for automatic compilation of metalinguistic information from
technical and scientific documents. This system is designed to extract
non-standard terminological resources that we have called Metalinguistic
Information Databases (or MIDs), in order to help update changing glossaries,
knowledge bases and ontologies, as well as to reflect the metastable dynamics
of special-domain knowledge.
| 2,007 | Computation and Language |
Seeing stars: Exploiting class relationships for sentiment
categorization with respect to rating scales | We address the rating-inference problem, wherein rather than simply decide
whether a review is "thumbs up" or "thumbs down", as in previous sentiment
analysis work, one must determine an author's evaluation with respect to a
multi-point scale (e.g., one to five "stars"). This task represents an
interesting twist on standard multi-class text categorization because there are
several different degrees of similarity between class labels; for example,
"three stars" is intuitively closer to "four stars" than to "one star". We
first evaluate human performance at the task. Then, we apply a meta-algorithm,
based on a metric labeling formulation of the problem, that alters a given
n-ary classifier's output in an explicit attempt to ensure that similar items
receive similar labels. We show that the meta-algorithm can provide significant
improvements over both multi-class and regression versions of SVMs when we
employ a novel similarity measure appropriate to the problem.
| 2,007 | Computation and Language |
On Hilberg's Law and Its Links with Guiraud's Law | Hilberg (1990) supposed that finite-order excess entropy of a random human
text is proportional to the square root of the text length. Assuming that
Hilberg's hypothesis is true, we derive Guiraud's law, which states that the
number of word types in a text is greater than proportional to the square root
of the text length. Our derivation is based on some mathematical conjecture in
coding theory and on several experiments suggesting that words can be defined
approximately as the nonterminals of the shortest context-free grammar for the
text. Such operational definition of words can be applied even to texts
deprived of spaces, which do not allow for Mandelbrot's ``intermittent
silence'' explanation of Zipf's and Guiraud's laws. In contrast to
Mandelbrot's, our model assumes some probabilistic long-memory effects in human
narration and might be capable of explaining Menzerath's law.
| 2,006 | Computation and Language |
Summarizing Reports on Evolving Events; Part I: Linear Evolution | We present an approach for summarization from multiple documents which report
on events that evolve through time, taking into account the different document
sources. We distinguish the evolution of an event into linear and non-linear.
According to our approach, each document is represented by a collection of
messages which are then used in order to instantiate the cross-document
relations that determine the summary content. The paper presents the
summarization system that implements this approach through a case study on
linear evolution.
| 2,005 | Computation and Language |
Automatic extraction of paraphrastic phrases from medium size corpora | This paper presents a versatile system intended to acquire paraphrastic
phrases from a representative corpus. In order to decrease the time spent on
the elaboration of resources for NLP system (for example Information
Extraction, IE hereafter), we suggest to use a machine learning system that
helps defining new templates and associated resources. This knowledge is
automatically derived from the text collection, in interaction with a large
semantic network.
| 2,004 | Computation and Language |
Word sense disambiguation criteria: a systematic study | This article describes the results of a systematic in-depth study of the
criteria used for word sense disambiguation. Our study is based on 60 target
words: 20 nouns, 20 adjectives and 20 verbs. Our results are not always in line
with some practices in the field. For example, we show that omitting
non-content words decreases performance and that bigrams yield better results
than unigrams.
| 2,004 | Computation and Language |
Using phonetic constraints in acoustic-to-articulatory inversion | The goal of this work is to recover articulatory information from the speech
signal by acoustic-to-articulatory inversion. One of the main difficulties with
inversion is that the problem is underdetermined and inversion methods
generally offer no guarantee on the phonetical realism of the inverse
solutions. A way to adress this issue is to use additional phonetic
constraints. Knowledge of the phonetic caracteristics of French vowels enable
the derivation of reasonable articulatory domains in the space of Maeda
parameters: given the formants frequencies (F1,F2,F3) of a speech sample, and
thus the vowel identity, an "ideal" articulatory domain can be derived. The
space of formants frequencies is partitioned into vowels, using either
speaker-specific data or generic information on formants. Then, to each
articulatory vector can be associated a phonetic score varying with the
distance to the "ideal domain" associated with the corresponding vowel.
Inversion experiments were conducted on isolated vowels and vowel-to-vowel
transitions. Articulatory parameters were compared with those obtained without
using these constraints and those measured from X-ray data.
| 2,005 | Computation and Language |
An elitist approach for extracting automatically well-realized speech
sounds with high confidence | This paper presents an "elitist approach" for extracting automatically
well-realized speech sounds with high confidence. The elitist approach uses a
speech recognition system based on Hidden Markov Models (HMM). The HMM are
trained on speech sounds which are systematically well-detected in an iterative
procedure. The results show that, by using the HMM models defined in the
training phase, the speech recognizer detects reliably specific speech sounds
with a small rate of errors.
| 2,007 | Computation and Language |
Statistical Parameters of the Novel "Perekhresni stezhky" ("The
Cross-Paths") by Ivan Franko | In the paper, a complex statistical characteristics of a Ukrainian novel is
given for the first time. The distribution of word-forms with respect to their
size is studied. The linguistic laws by Zipf-Mandelbrot and Altmann-Menzerath
are analyzed.
| 2,007 | Computation and Language |
Analyzing language development from a network approach | In this paper we propose some new measures of language development using
network analyses, which is inspired by the recent surge of interests in network
studies of many real-world systems. Children's and care-takers' speech data
from a longitudinal study are represented as a series of networks, word forms
being taken as nodes and collocation of words as links. Measures on the
properties of the networks, such as size, connectivity, hub and authority
analyses, etc., allow us to make quantitative comparison so as to reveal
different paths of development. For example, the asynchrony of development in
network size and average degree suggests that children cannot be simply
classified as early talkers or late talkers by one or two measures. Children
follow different paths in a multi-dimensional space. They may develop faster in
one dimension but slower in another dimension. The network approach requires
little preprocessing of words and analyses on sentence structures, and the
characteristics of words and their usage emerge from the network and are
independent of any grammatical presumptions. We show that the change of the two
articles "the" and "a" in their roles as important nodes in the network
reflects the progress of children's syntactic development: the two articles
often start in children's networks as hubs and later shift to authorities,
while they are authorities constantly in the adult's networks. The network
analyses provide a new approach to study language development, and at the same
time language development also presents a rich area for network theories to
explore.
| 2,007 | Computation and Language |
Constraint-based verification of abstract models of multitreaded
programs | We present a technique for the automated verification of abstract models of
multithreaded programs providing fresh name generation, name mobility, and
unbounded control.
As high level specification language we adopt here an extension of
communication finite-state machines with local variables ranging over an
infinite name domain, called TDL programs. Communication machines have been
proved very effective for representing communication protocols as well as for
representing abstractions of multithreaded software.
The verification method that we propose is based on the encoding of TDL
programs into a low level language based on multiset rewriting and constraints
that can be viewed as an extension of Petri Nets. By means of this encoding,
the symbolic verification procedure developed for the low level language in our
previous work can now be applied to TDL programs. Furthermore, the encoding
allows us to isolate a decidable class of verification problems for TDL
programs that still provide fresh name generation, name mobility, and unbounded
control. Our syntactic restrictions are in fact defined on the internal
structure of threads: In order to obtain a complete and terminating method,
threads are only allowed to have at most one local variable (ranging over an
infinite domain of names).
| 2,007 | Computation and Language |
Unification of multi-lingual scientific terminological resources using
the ISO 16642 standard. The TermSciences initiative | This paper presents the TermSciences portal, which deals with the
implementation of a conceptual model that uses the recent ISO 16642 standard
(Terminological Markup Framework). This standard turns out to be suitable for
concept modeling since it allowed for organizing the original resources by
concepts and to associate the various terms for a given concept. Additional
structuring is produced by sharing conceptual relationships, that is,
cross-linking of resource results through the introduction of semantic
relations which may have initially be missing.
| 2,009 | Computation and Language |
Numeration-automatic sequences | We present a base class of automata that induce a numeration system and we
give an algorithm to give the n-th word in the language of the automaton when
the expansion of n in the induced numeration system is feeded to the automaton.
Furthermore we give some algorithms for reverse reading of this expansion and a
way to combine automata to other automata having the same properties.
| 2,007 | Computation and Language |
Foundations of Modern Language Resource Archives | A number of serious reasons will convince an increasing amount of researchers
to store their relevant material in centers which we will call "language
resource archives". They combine the duty of taking care of long-term
preservation as well as the task to give access to their material to different
user groups. Access here is meant in the sense that an active interaction with
the data will be made possible to support the integration of new data, new
versions or commentaries of all sort. Modern Language Resource Archives will
have to adhere to a number of basic principles to fulfill all requirements and
they will have to be involved in federations to create joint language resource
domains making it even more simple for the researchers to access the data. This
paper makes an attempt to formulate the essential pillars language resource
archives have to adhere to.
| 2,009 | Computation and Language |
Building a resource for studying translation shifts | This paper describes an interdisciplinary approach which brings together the
fields of corpus linguistics and translation studies. It presents ongoing work
on the creation of a corpus resource in which translation shifts are explicitly
annotated. Translation shifts denote departures from formal correspondence
between source and target text, i.e. deviations that have occurred during the
translation process. A resource in which such shifts are annotated in a
systematic way will make it possible to study those phenomena that need to be
addressed if machine translation output is to resemble human translation. The
resource described in this paper contains English source texts (parliamentary
proceedings) and their German translations. The shift annotation is based on
predicate-argument structures and proceeds in two steps: first, predicates and
their arguments are annotated monolingually in a straightforward manner. Then,
the corresponding English and German predicates and arguments are aligned with
each other. Whenever a shift - mainly grammatical or semantic -has occurred,
the alignment is tagged accordingly.
| 2,006 | Computation and Language |
Adapting a general parser to a sublanguage | In this paper, we propose a method to adapt a general parser (Link Parser) to
sublanguages, focusing on the parsing of texts in biology. Our main proposal is
the use of terminology (identication and analysis of terms) in order to reduce
the complexity of the text to be parsed. Several other strategies are explored
and finally combined among which text normalization, lexicon and
morpho-guessing module extensions and grammar rules adaptation. We compare the
parsing results before and after these adaptations.
| 2,005 | Computation and Language |
Lexical Adaptation of Link Grammar to the Biomedical Sublanguage: a
Comparative Evaluation of Three Approaches | We study the adaptation of Link Grammar Parser to the biomedical sublanguage
with a focus on domain terms not found in a general parser lexicon. Using two
biomedical corpora, we implement and evaluate three approaches to addressing
unknown words: automatic lexicon expansion, the use of morphological clues, and
disambiguation using a part-of-speech tagger. We evaluate each approach
separately for its effect on parsing performance and consider combinations of
these approaches. In addition to a 45% increase in parsing efficiency, we find
that the best approach, incorporating information from a domain part-of-speech
tagger, offers a statistically signicant 10% relative decrease in error. The
adapted parser is available under an open-source license at
http://www.it.utu.fi/biolg.
| 2,006 | Computation and Language |
Raisonner avec des diagrammes : perspectives cognitives et
computationnelles | Diagrammatic, analogical or iconic representations are often contrasted with
linguistic or logical representations, in which the shape of the symbols is
arbitrary. The aim of this paper is to make a case for the usefulness of
diagrams in inferential knowledge representation systems. Although commonly
used, diagrams have for a long time suffered from the reputation of being only
a heuristic tool or a mere support for intuition. The first part of this paper
is an historical background paying tribute to the logicians, psychologists and
computer scientists who put an end to this formal prejudice against diagrams.
The second part is a discussion of their characteristics as opposed to those of
linguistic forms. The last part is aimed at reviving the interest for
heterogeneous representation systems including both linguistic and diagrammatic
representations.
| 2,005 | Computation and Language |
Get out the vote: Determining support or opposition from Congressional
floor-debate transcripts | We investigate whether one can determine from the transcripts of U.S.
Congressional floor debates whether the speeches represent support of or
opposition to proposed legislation. To address this problem, we exploit the
fact that these speeches occur as part of a discussion; this allows us to use
sources of information regarding relationships between discourse segments, such
as whether a given utterance indicates agreement with the opinion expressed by
another. We find that the incorporation of such information yields substantial
improvements over classifying speeches in isolation.
| 2,012 | Computation and Language |
Using Answer Set Programming in an Inference-Based approach to Natural
Language Semantics | Using Answer Set Programming in an Inference-Based approach to Natural
Language Semantics
| 2,006 | Computation and Language |
Expressing Implicit Semantic Relations without Supervision | We present an unsupervised learning algorithm that mines large text corpora
for patterns that express implicit semantic relations. For a given input word
pair X:Y with some unspecified semantic relations, the corresponding output
list of patterns <P1,...,Pm> is ranked according to how well each pattern Pi
expresses the relations between X and Y. For example, given X=ostrich and
Y=bird, the two highest ranking output patterns are "X is the largest Y" and "Y
such as the X". The output patterns are intended to be useful for finding
further pairs with the same relations, to support the construction of lexicons,
ontologies, and semantic networks. The patterns are sorted by pertinence, where
the pertinence of a pattern Pi for a word pair X:Y is the expected relational
similarity between the given pair and typical pairs for Pi. The algorithm is
empirically evaluated on two tasks, solving multiple-choice SAT word analogy
questions and classifying semantic relations in noun-modifier pairs. On both
tasks, the algorithm achieves state-of-the-art results, performing
significantly better than several alternative pattern ranking algorithms, based
on tf-idf.
| 2,006 | Computation and Language |
Similarity of Semantic Relations | There are at least two kinds of similarity. Relational similarity is
correspondence between relations, in contrast with attributional similarity,
which is correspondence between attributes. When two words have a high degree
of attributional similarity, we call them synonyms. When two pairs of words
have a high degree of relational similarity, we say that their relations are
analogous. For example, the word pair mason:stone is analogous to the pair
carpenter:wood. This paper introduces Latent Relational Analysis (LRA), a
method for measuring relational similarity. LRA has potential applications in
many areas, including information extraction, word sense disambiguation, and
information retrieval. Recently the Vector Space Model (VSM) of information
retrieval has been adapted to measuring relational similarity, achieving a
score of 47% on a collection of 374 college-level multiple-choice word analogy
questions. In the VSM approach, the relation between a pair of words is
characterized by a vector of frequencies of predefined patterns in a large
corpus. LRA extends the VSM approach in three ways: (1) the patterns are
derived automatically from the corpus, (2) the Singular Value Decomposition
(SVD) is used to smooth the frequency data, and (3) automatically generated
synonyms are used to explore variations of the word pairs. LRA achieves 56% on
the 374 analogy questions, statistically equivalent to the average human score
of 57%. On the related problem of classifying semantic relations, LRA achieves
similar gains over the VSM.
| 2,006 | Computation and Language |
Improving Term Extraction with Terminological Resources | Studies of different term extractors on a corpus of the biomedical domain
revealed decreasing performances when applied to highly technical texts. The
difficulty or impossibility of customising them to new domains is an additional
limitation. In this paper, we propose to use external terminologies to
influence generic linguistic data in order to augment the quality of the
extraction. The tool we implemented exploits testified terms at different steps
of the process: chunking, parsing and extraction of term candidates.
Experiments reported here show that, using this method, more term candidates
can be acquired with a higher level of reliability. We further describe the
extraction process involving endogenous disambiguation implemented in the term
extractor YaTeA.
| 2,006 | Computation and Language |
Challenging the principle of compositionality in interpreting natural
language texts | The paper aims at emphasizing that, even relaxed, the hypothesis of
compositionality has to face many problems when used for interpreting natural
language texts. Rather than fixing these problems within the compositional
framework, we believe that a more radical change is necessary, and propose
another approach.
| 2,004 | Computation and Language |
The role of time in considering collections | The paper concerns the understanding of plurals in the framework of
Artificial Intelligence and emphasizes the role of time. The construction of
collection(s) and their evolution across time is often crucial and has to be
accounted for. The paper contrasts a "de dicto" collection where the collection
can be considered as persisting over these situations even if its members
change with a "de re" collection whose composition does not vary through time.
It expresses different criteria of choice between the two interpretations (de
re and de dicto) depending on the context of enunciation.
| 2,004 | Computation and Language |
Multilingual person name recognition and transliteration | We present an exploratory tool that extracts person names from multilingual
news collections, matches name variants referring to the same person, and
infers relationships between people based on the co-occurrence of their names
in related news. A novel feature is the matching of name variants across
languages and writing systems, including names written with the Greek, Cyrillic
and Arabic writing system. Due to our highly multilingual setting, we use an
internal standard representation for name representation and matching, instead
of adopting the traditional bilingual approach to transliteration. This work is
part of the news analysis system NewsExplorer that clusters an average of
25,000 news articles per day to detect related news within the same and across
different languages.
| 2,005 | Computation and Language |
Navigating multilingual news collections using automatically extracted
information | We are presenting a text analysis tool set that allows analysts in various
fields to sieve through large collections of multilingual news items quickly
and to find information that is of relevance to them. For a given document
collection, the tool set automatically clusters the texts into groups of
similar articles, extracts names of places, people and organisations, lists the
user-defined specialist terms found, links clusters and entities, and generates
hyperlinks. Through its daily news analysis operating on thousands of articles
per day, the tool also learns relationships between people and other entities.
The fully functional prototype system allows users to explore and navigate
multilingual document collections across languages and time.
| 2,005 | Computation and Language |
The JRC-Acquis: A multilingual aligned parallel corpus with 20+
languages | We present a new, unique and freely available parallel corpus containing
European Union (EU) documents of mostly legal nature. It is available in all 20
official EUanguages, with additional documents being available in the languages
of the EU candidate countries. The corpus consists of almost 8,000 documents
per language, with an average size of nearly 9 million words per language.
Pair-wise paragraph alignment information produced by two different aligners
(Vanilla and HunAlign) is available for all 190+ language pair combinations.
Most texts have been manually classified according to the EUROVOC subject
domains so that the collection can also be used to train and test multi-label
classification algorithms and keyword-assignment software. The corpus is
encoded in XML, according to the Text Encoding Initiative Guidelines. Due to
the large number of parallel texts in many languages, the JRC-Acquis is
particularly suitable to carry out all types of cross-language research, as
well as to test and benchmark text analysis software across different languages
(for instance for alignment, sentence splitting and term extraction).
| 2,006 | Computation and Language |
Automatic annotation of multilingual text collections with a conceptual
thesaurus | Automatic annotation of documents with controlled vocabulary terms
(descriptors) from a conceptual thesaurus is not only useful for document
indexing and retrieval. The mapping of texts onto the same thesaurus
furthermore allows to establish links between similar documents. This is also a
substantial requirement of the Semantic Web. This paper presents an almost
language-independent system that maps documents written in different languages
onto the same multilingual conceptual thesaurus, EUROVOC. Conceptual thesauri
differ from Natural Language Thesauri in that they consist of relatively small
controlled lists of words or phrases with a rather abstract meaning. To
automatically identify which thesaurus descriptors describe the contents of a
document best, we developed a statistical, associative system that is trained
on texts that have previously been indexed manually. In addition to describing
the large number of empirically optimised parameters of the fully functional
application, we present the performance of the software according to a human
evaluation by professional indexers.
| 2,003 | Computation and Language |
Automatic Identification of Document Translations in Large Multilingual
Document Collections | Texts and their translations are a rich linguistic resource that can be used
to train and test statistics-based Machine Translation systems and many other
applications. In this paper, we present a working system that can identify
translations and other very similar documents among a large number of
candidates, by representing the document contents with a vector of thesaurus
terms from a multilingual thesaurus, and by then measuring the semantic
similarity between the vectors. Tests on different text types have shown that
the system can detect translations with over 96% precision in a large search
space of 820 documents or more. The system was tuned to ignore
language-specific similarities and to give similar documents in a second
language the same similarity score as equivalent documents in the same
language. The application can also be used to detect cross-lingual document
plagiarism.
| 2,003 | Computation and Language |
Cross-lingual keyword assignment | This paper presents a language-independent approach to controlled vocabulary
keyword assignment using the EUROVOC thesaurus. Due to the multilingual nature
of EUROVOC, the keywords for a document written in one language can be
displayed in all eleven official European Union languages. The mapping of
documents written in different languages to the same multilingual thesaurus
furthermore allows cross-language document comparison. The assignment of the
controlled vocabulary thesaurus descriptors is achieved by applying a
statistical method that uses a collection of manually indexed documents to
identify, for each thesaurus descriptor, a large number of lemmas that are
statistically associated to the descriptor. These associated words are then
used during the assignment procedure to identify a ranked list of those EUROVOC
terms that are most likely to be good keywords for a given document. The paper
also describes the challenges of this task and discusses the achieved results
of the fully functional prototype.
| 2,001 | Computation and Language |
Extending an Information Extraction tool set to Central and Eastern
European languages | In a highly multilingual and multicultural environment such as in the
European Commission with soon over twenty official languages, there is an
urgent need for text analysis tools that use minimal linguistic knowledge so
that they can be adapted to many languages without much human effort. We are
presenting two such Information Extraction tools that have already been adapted
to various Western and Eastern European languages: one for the recognition of
date expressions in text, and one for the detection of geographical place names
and the visualisation of the results in geographical maps. An evaluation of the
performance has produced very satisfying results.
| 2,003 | Computation and Language |
Exploiting multilingual nomenclatures and language-independent text
features as an interlingua for cross-lingual text analysis applications | We are proposing a simple, but efficient basic approach for a number of
multilingual and cross-lingual language technology applications that are not
limited to the usual two or three languages, but that can be applied with
relatively little effort to larger sets of languages. The approach consists of
using existing multilingual linguistic resources such as thesauri,
nomenclatures and gazetteers, as well as exploiting the existence of additional
more or less language-independent text items such as dates, currency
expressions, numbers, names and cognates. Mapping texts onto the multilingual
resources and identifying word token links between texts in different languages
are basic ingredients for applications such as cross-lingual document
similarity calculation, multilingual clustering and categorisation,
cross-lingual document retrieval, and tools to provide cross-lingual
information access.
| 2,004 | Computation and Language |
Geocoding multilingual texts: Recognition, disambiguation and
visualisation | We are presenting a method to recognise geographical references in free text.
Our tool must work on various languages with a minimum of language-dependent
resources, except a gazetteer. The main difficulty is to disambiguate these
place names by distinguishing places from persons and by selecting the most
likely place out of a list of homographic place names world-wide. The system
uses a number of language-independent clues and heuristics to disambiguate
place name homographs. The final aim is to index texts with the countries and
cities they mention and to automatically visualise this information on
geographical maps using various tools.
| 2,006 | Computation and Language |
Building and displaying name relations using automatic unsupervised
analysis of newspaper articles | We present a tool that, from automatically recognised names, tries to infer
inter-person relations in order to present associated people on maps. Based on
an in-house Named Entity Recognition tool, applied on clusters of an average of
15,000 news articles per day, in 15 different languages, we build a knowledge
base that allows extracting statistical co-occurrences of persons and
visualising them on a per-person page or in various graphs.
| 2,006 | Computation and Language |
A tool set for the quick and efficient exploration of large document
collections | We are presenting a set of multilingual text analysis tools that can help
analysts in any field to explore large document collections quickly in order to
determine whether the documents contain information of interest, and to find
the relevant text passages. The automatic tool, which currently exists as a
fully functional prototype, is expected to be particularly useful when users
repeatedly have to sieve through large collections of documents such as those
downloaded automatically from the internet. The proposed system takes a whole
document collection as input. It first carries out some automatic analysis
tasks (named entity recognition, geo-coding, clustering, term extraction),
annotates the texts with the generated meta-information and stores the
meta-information in a database. The system then generates a zoomable and
hyperlinked geographic map enhanced with information on entities and terms
found. When the system is used on a regular basis, it builds up a historical
database that contains information on which names have been mentioned together
with which other names or places, and users can query this database to retrieve
information extracted in the past.
| 2,005 | Computation and Language |
Rapport technique du projet OGRE | This repport concerns automatic understanding of (french) iterative
sentences, i.e. sentences where one single verb has to be interpreted by a more
or less regular plurality of events. A linguistic analysis is proposed along an
extension of Reichenbach's theory, several formal representations are
considered and a corpus of 18000 newspaper extracts is described.
| 2,016 | Computation and Language |
DepAnn - An Annotation Tool for Dependency Treebanks | DepAnn is an interactive annotation tool for dependency treebanks, providing
both graphical and text-based annotation interfaces. The tool is aimed for
semi-automatic creation of treebanks. It aids the manual inspection and
correction of automatically created parses, making the annotation process
faster and less error-prone. A novel feature of the tool is that it enables the
user to view outputs from several parsers as the basis for creating the final
tree to be saved to the treebank. DepAnn uses TIGER-XML, an XML-based general
encoding format for both, representing the parser outputs and saving the
annotated treebank. The tool includes an automatic consistency checker for
sentence structures. In addition, the tool enables users to build structures
manually, add comments on the annotations, modify the tagsets, and mark
sentences for further revision.
| 2,006 | Computation and Language |
Dependency Treebanks: Methods, Annotation Schemes and Tools | In this paper, current dependencybased treebanks are introduced and analyzed.
The methods used for building the resources, the annotation schemes applied,
and the tools used (such as POS taggers, parsers and annotation software) are
discussed.
| 2,005 | Computation and Language |
Un mod\`ele g\'en\'erique d'organisation de corpus en ligne: application
\`a la FReeBank | The few available French resources for evaluating linguistic models or
algorithms on other linguistic levels than morpho-syntax are either
insufficient from quantitative as well as qualitative point of view or not
freely accessible. Based on this fact, the FREEBANK project intends to create
French corpora constructed using manually revised output from a hybrid
Constraint Grammar parser and annotated on several linguistic levels
(structure, morpho-syntax, syntax, coreference), with the objective to make
them available on-line for research purposes. Therefore, we will focus on using
standard annotation schemes, integration of existing resources and maintenance
allowing for continuous enrichment of the annotations. Prior to the actual
presentation of the prototype that has been implemented, this paper describes a
generic model for the organization and deployment of a linguistic resource
archive, in compliance with the various works currently conducted within
international standardization initiatives (TEI and ISO/TC 37/SC 4).
| 2,006 | Computation and Language |
Scaling Construction Grammar up to Production Systems: the SCIM | While a great effort has concerned the development of fully integrated
modular understanding systems, few researches have focused on the problem of
unifying existing linguistic formalisms with cognitive processing models. The
Situated Constructional Interpretation Model is one of these attempts. In this
model, the notion of "construction" has been adapted in order to be able to
mimic the behavior of Production Systems. The Construction Grammar approach
establishes a model of the relations between linguistic forms and meaning, by
the mean of constructions. The latter can be considered as pairings from a
topologically structured space to an unstructured space, in some way a special
kind of production rules.
| 2,006 | Computation and Language |
An Anthological Review of Research Utilizing MontyLingua, a Python-Based
End-to-End Text Processor | MontyLingua, an integral part of ConceptNet which is currently the largest
commonsense knowledge base, is an English text processor developed using Python
programming language in MIT Media Lab. The main feature of MontyLingua is the
coverage for all aspects of English text processing from raw input text to
semantic meanings and summary generation, yet each component in MontyLingua is
loosely-coupled to each other at the architectural and code level, which
enabled individual components to be used independently or substituted. However,
there has been no review exploring the role of MontyLingua in recent research
work utilizing it. This paper aims to review the use of and roles played by
MontyLingua and its components in research work published in 19 articles
between October 2004 and August 2006. We had observed a diversified use of
MontyLingua in many different areas, both generic and domain-specific. Although
the use of text summarizing component had not been observe, we are optimistic
that it will have a crucial role in managing the current trend of information
overload in future research.
| 2,006 | Computation and Language |
Acronym-Meaning Extraction from Corpora Using Multi-Tape Weighted
Finite-State Machines | The automatic extraction of acronyms and their meaning from corpora is an
important sub-task of text mining. It can be seen as a special case of string
alignment, where a text chunk is aligned with an acronym. Alternative
alignments have different cost, and ideally the least costly one should give
the correct meaning of the acronym. We show how this approach can be
implemented by means of a 3-tape weighted finite-state machine (3-WFSM) which
reads a text chunk on tape 1 and an acronym on tape 2, and generates all
alternative alignments on tape 3. The 3-WFSM can be automatically generated
from a simple regular expression. No additional algorithms are required at any
stage. Our 3-WFSM has a size of 27 states and 64 transitions, and finds the
best analysis of an acronym in a few milliseconds.
| 2,009 | Computation and Language |
Viterbi Algorithm Generalized for n-Tape Best-Path Search | We present a generalization of the Viterbi algorithm for identifying the path
with minimal (resp. maximal) weight in a n-tape weighted finite-state machine
(n-WFSM), that accepts a given n-tuple of input strings (s_1,... s_n). It also
allows us to compile the best transduction of a given input n-tuple by a
weighted (n+m)-WFSM (transducer) with n input and m output tapes. Our algorithm
has a worst-case time complexity of O(|s|^n |E| log (|s|^n |Q|)), where n and
|s| are the number and average length of the strings in the n-tuple, and |Q|
and |E| the number of states and transitions in the n-WFSM, respectively. A
straight forward alternative, consisting in intersection followed by classical
shortest-distance search, operates in O(|s|^n (|E|+|Q|) log (|s|^n |Q|)) time.
| 2,009 | Computation and Language |
Statistical keyword detection in literary corpora | Understanding the complexity of human language requires an appropriate
analysis of the statistical distribution of words in texts. We consider the
information retrieval problem of detecting and ranking the relevant words of a
text by means of statistical information referring to the "spatial" use of the
words. Shannon's entropy of information is used as a tool for automatic keyword
extraction. By using The Origin of Species by Charles Darwin as a
representative text sample, we show the performance of our detector and compare
it with another proposals in the literature. The random shuffled text receives
special attention as a tool for calibrating the ranking indices.
| 2,008 | Computation and Language |
Complex networks and human language | This paper introduces how human languages can be studied in light of recent
development of network theories. There are two directions of exploration. One
is to study networks existing in the language system. Various lexical networks
can be built based on different relationships between words, being semantic or
syntactic. Recent studies have shown that these lexical networks exhibit
small-world and scale-free features. The other direction of exploration is to
study networks of language users (i.e. social networks of people in the
linguistic community), and their role in language evolution. Social networks
also show small-world and scale-free features, which cannot be captured by
random or regular network models. In the past, computational models of language
change and language emergence often assume a population to have a random or
regular structure, and there has been little discussion how network structures
may affect the dynamics. In the second part of the paper, a series of
simulation models of diffusion of linguistic innovation are used to illustrate
the importance of choosing realistic conditions of population structure for
modeling language change. Four types of social networks are compared, which
exhibit two categories of diffusion dynamics. While the questions about which
type of networks are more appropriate for modeling still remains, we give some
preliminary suggestions for choosing the type of social networks for modeling.
| 2,007 | Computation and Language |
A Note on Local Ultrametricity in Text | High dimensional, sparsely populated data spaces have been characterized in
terms of ultrametric topology. This implies that there are natural, not
necessarily unique, tree or hierarchy structures defined by the ultrametric
topology. In this note we study the extent of local ultrametric topology in
texts, with the aim of finding unique ``fingerprints'' for a text or corpus,
discriminating between texts from different domains, and opening up the
possibility of exploiting hierarchical structures in the data. We use coherent
and meaningful collections of over 1000 texts, comprising over 1.3 million
words.
| 2,007 | Computation and Language |
Menzerath-Altmann Law for Syntactic Structures in Ukrainian | In the paper, the definition of clause suitable for an automated processing
of a Ukrainian text is proposed. The Menzerath-Altmann law is verified on the
sentence level and the parameters for the dependences of the clause length
counted in words and syllables on the sentence length counted in clauses are
calculated for "Perekhresni Stezhky" ("The Cross-Paths"), a novel by Ivan
Franko.
| 2,008 | Computation and Language |
Random Sentences from a Generalized Phrase-Structure Grammar Interpreter | In numerous domains in cognitive science it is often useful to have a source
for randomly generated corpora. These corpora may serve as a foundation for
artificial stimuli in a learning experiment (e.g., Ellefson & Christiansen,
2000), or as input into computational models (e.g., Christiansen & Dale, 2001).
The following compact and general C program interprets a phrase-structure
grammar specified in a text file. It follows parameters set at a Unix or
Unix-based command-line and generates a corpus of random sentences from that
grammar.
| 2,007 | Computation and Language |
Interroger un corpus par le sens | In textual knowledge management, statistical methods prevail. Nonetheless,
some difficulties cannot be overcome by these methodologies. I propose a
symbolic approach using a complete textual analysis to identify which analysis
level can improve the the answers provided by a system. The approach identifies
word senses and relation between words and generates as many rephrasings as
possible. Using synonyms and derivative, the system provides new utterances
without changing the original meaning of the sentences. Such a way, an
information can be retrieved whatever the question or answer's wording may be.
| 2,005 | Computation and Language |
Dependency Parsing with Dynamic Bayesian Network | Exact parsing with finite state automata is deemed inappropriate because of
the unbounded non-locality languages overwhelmingly exhibit. We propose a way
to structure the parsing task in order to make it amenable to local
classification methods. This allows us to build a Dynamic Bayesian Network
which uncovers the syntactic dependency structure of English sentences.
Experiments with the Wall Street Journal demonstrate that the model
successfully learns from labeled data.
| 2,005 | Computation and Language |
Linear Segmentation and Segment Significance | We present a new method for discovering a segmental discourse structure of a
document while categorizing segment function. We demonstrate how retrieval of
noun phrases and pronominal forms, along with a zero-sum weighting scheme,
determines topicalized segmentation. Futhermore, we use term distribution to
aid in identifying the role that the segment performs in the document. Finally,
we present results of evaluation in terms of precision and recall which surpass
earlier approaches.
| 1,998 | Computation and Language |
Producing NLP-based On-line Contentware | For its internal needs as well as for commercial purposes, CDC Group has
produced several NLP-based on-line contentware applications for years. The
development process of such applications is subject to numerous constraints
such as quality of service, integration of new advances in NLP, direct
reactions from users, continuous versioning, short delivery deadlines and cost
control. Following this industrial and commercial experience, malleability of
the applications, their openness towards foreign components, efficiency of
applications and their ease of exploitation have appeared to be key points. In
this paper, we describe TalLab, a powerful architecture for on-line contentware
which fulfils these requirements.
| 1,998 | Computation and Language |
Modelling Users, Intentions, and Structure in Spoken Dialog | We outline how utterances in dialogs can be interpreted using a partial first
order logic. We exploit the capability of this logic to talk about the truth
status of formulae to define a notion of coherence between utterances and
explain how this coherence relation can serve for the construction of AND/OR
trees that represent the segmentation of the dialog. In a BDI model we
formalize basic assumptions about dialog and cooperative behaviour of
participants. These assumptions provide a basis for inferring speech acts from
coherence relations between utterances and attitudes of dialog participants.
Speech acts prove to be useful for determining dialog segments defined on the
notion of completing expectations of dialog participants. Finally, we sketch
how explicit segmentation signalled by cue phrases and performatives is covered
by our dialog model.
| 2,007 | Computation and Language |
A Lexicalized Tree Adjoining Grammar for English | This document describes a sizable grammar of English written in the TAG
formalism and implemented for use with the XTAG system. This report and the
grammar described herein supersedes the TAG grammar described in an earlier
1995 XTAG technical report. The English grammar described in this report is
based on the TAG formalism which has been extended to include lexicalization,
and unification-based feature structures. The range of syntactic phenomena that
can be handled is large and includes auxiliaries (including inversion), copula,
raising and small clause constructions, topicalization, relative clauses,
infinitives, gerunds, passives, adjuncts, it-clefts, wh-clefts, PRO
constructions, noun-noun modifications, extraposition, determiner sequences,
genitives, negation, noun-verb contractions, sentential adjuncts and
imperatives. This technical report corresponds to the XTAG Release 8/31/98. The
XTAG grammar is continuously updated with the addition of new analyses and
modification of old ones, and an online version of this report can be found at
the XTAG web page at http://www.cis.upenn.edu/~xtag/
| 2,012 | Computation and Language |
Separating Dependency from Constituency in a Tree Rewriting System | In this paper we present a new tree-rewriting formalism called Link-Sharing
Tree Adjoining Grammar (LSTAG) which is a variant of synchronous TAGs. Using
LSTAG we define an approach towards coordination where linguistic dependency is
distinguished from the notion of constituency. Such an approach towards
coordination that explicitly distinguishes dependencies from constituency gives
a better formal understanding of its representation when compared to previous
approaches that use tree-rewriting systems which conflate the two issues.
| 1,997 | Computation and Language |
Incremental Parser Generation for Tree Adjoining Grammars | This paper describes the incremental generation of parse tables for the
LR-type parsing of Tree Adjoining Languages (TALs). The algorithm presented
handles modifications to the input grammar by updating the parser generated so
far. In this paper, a lazy generation of LR-type parsers for TALs is defined in
which parse tables are created by need while parsing. We then describe an
incremental parser generator for TALs which responds to modification of the
input grammar by updating parse tables built so far.
| 1,996 | Computation and Language |
A Freely Available Morphological Analyzer, Disambiguator and Context
Sensitive Lemmatizer for German | In this paper we present Morphy, an integrated tool for German morphology,
part-of-speech tagging and context-sensitive lemmatization. Its large lexicon
of more than 320,000 word forms plus its ability to process German compound
nouns guarantee a wide morphological coverage. Syntactic ambiguities can be
resolved with a standard statistical part-of-speech tagger. By using the output
of the tagger, the lemmatizer can determine the correct root even for ambiguous
word forms. The complete package is freely available and can be downloaded from
the World Wide Web.
| 1,998 | Computation and Language |
Spoken Language Dialogue Systems and Components: Best practice in
development and evaluation (DISC 24823) - Periodic Progress Report 1: Basic
Details of the Action | The DISC project aims to (a) build an in-depth understanding of the
state-of-the-art in spoken language dialogue systems (SLDSs) and components
development and evaluation with the purpose of (b) developing a first best
practice methodology in the field. The methodology will be accompanied by (c) a
series of development and evaluation support tools. To the limited extent
possible within the duration of the project, the draft versions of the
methodology and the tools will be (d) tested by SLDS developers from industry
and research, and will be (e) packaged to best suit their needs. In the first
year of DISC, (a) has been accomplished, and (b) and (c) have started. A
proposal to complete the work proposed above by adding 12 months to the 18
months of the present project, has been submitted to Esprit Long-Term Research
in March 1998.
| 2,007 | Computation and Language |
Similarity-Based Models of Word Cooccurrence Probabilities | In many applications of natural language processing (NLP) it is necessary to
determine the likelihood of a given word combination. For example, a speech
recognizer may need to determine which of the two word combinations ``eat a
peach'' and ``eat a beach'' is more likely. Statistical NLP methods determine
the likelihood of a word combination from its frequency in a training corpus.
However, the nature of language is such that many word combinations are
infrequent and do not occur in any given corpus. In this work we propose a
method for estimating the probability of such previously unseen word
combinations using available information on ``most similar'' words.
We describe probabilistic word association models based on distributional
word similarity, and apply them to two tasks, language modeling and pseudo-word
disambiguation. In the language modeling task, a similarity-based model is used
to improve probability estimates for unseen bigrams in a back-off language
model. The similarity-based method yields a 20% perplexity improvement in the
prediction of unseen bigrams and statistically significant reductions in
speech-recognition error.
We also compare four similarity-based estimation methods against back-off and
maximum-likelihood estimation methods on a pseudo-word sense disambiguation
task in which we controlled for both unigram and bigram frequency to avoid
giving too much weight to easy-to-disambiguate high-frequency configurations.
The similarity-based methods perform up to 40% better on this particular task.
| 1,999 | Computation and Language |
On the Evaluation and Comparison of Taggers: The Effect of Noise in
Testing Corpora | This paper addresses the issue of {\sc pos} tagger evaluation. Such
evaluation is usually performed by comparing the tagger output with a reference
test corpus, which is assumed to be error-free. Currently used corpora contain
noise which causes the obtained performance to be a distortion of the real
value. We analyze to what extent this distortion may invalidate the comparison
between taggers or the measure of the improvement given by a new system. The
main conclusion is that a more rigorous testing experimentation
setting/designing is needed to reliably evaluate and compare tagger accuracies.
| 2,007 | Computation and Language |
Improving Tagging Performance by Using Voting Taggers | We present a bootstrapping method to develop an annotated corpus, which is
specially useful for languages with few available resources. The method is
being applied to develop a corpus of Spanish of over 5Mw. The method consists
on taking advantage of the collaboration of two different POS taggers. The
cases in which both taggers agree present a higher accuracy and are used to
retrain the taggers.
| 2,007 | Computation and Language |
Ultrametric Distance in Syntax | Phrase structure trees have a hierarchical structure. In many subjects, most
notably in Taxonomy such tree structures have been studied using ultrametrics.
Here syntactical hierarchical phrase trees are subject to a similar analysis,
which is much siompler as the branching structure is more readily discernible
and switched. The occurence of hierarchical structure elsewhere in linguistics
is mentioned. The phrase tree can be represented by a matrix and the elements
of the matrix can be represented by triangles. The height at which branching
occurs is not prescribed in previous syntatic models, but it is by using the
ultrametric matrix. The ambiguity of which branching height to choose is
resolved by postulating that branching occurs at the lowest height available.
An ultrametric produces a measure of the complexity of sentences: presumably
the complexity of sentence increases as a language is aquired so that this can
be tested. A All ultrametric triangles are equilateral or isocles, here it is
shown that X structur implies that there are no equilateral triangles.
Restricting attention to simple syntax a minium ultrametric distance between
lexical categories is calculatex. This ultrametric distance is shown to be
different than the matrix obtasined from feaures. It is shown that the
definition of c-commabnd can be replaced by an equivalent ultrametric
definition. The new definition invokes a minimum distance between nodes and
this is more aesthetically satisfing than previouv varieties of definitions.
From the new definition of c-command follows a new definition of government.
| 2,015 | Computation and Language |
Resources for Evaluation of Summarization Techniques | We report on two corpora to be used in the evaluation of component systems
for the tasks of (1) linear segmentation of text and (2) summary-directed
sentence extraction. We present characteristics of the corpora, methods used in
the collection of user judgments, and an overview of the application of the
corpora to evaluating the component system. Finally, we discuss the problems
and issues with construction of the test set which apply broadly to the
construction of evaluation resources for language technologies.
| 1,998 | Computation and Language |
Does Meaning Evolve? | A common method of making a theory more understandable, is by comparing it to
another theory which has been better developed. Radical interpretation is a
theory which attempts to explain how communication has meaning. Radical
interpretation is treated as another time-dependent theory and compared to the
time dependent theory of biological evolution. The main reason for doing this
is to find the nature of the time dependence; producing analogs between the two
theories is a necessary prerequisite to this and brings up many problems. Once
the nature of the time dependence is better known it might allow the underlying
mechanism to be uncovered. Several similarities and differences are uncovered,
there appear to be more differences than similarities.
| 2,007 | Computation and Language |
Machine Learning of Generic and User-Focused Summarization | A key problem in text summarization is finding a salience function which
determines what information in the source should be included in the summary.
This paper describes the use of machine learning on a training corpus of
documents and their abstracts to discover salience functions which describe
what combination of features is optimal for a given summarization task. The
method addresses both "generic" and user-focused summaries.
| 2,007 | Computation and Language |
Translating near-synonyms: Possibilities and preferences in the
interlingua | This paper argues that an interlingual representation must explicitly
represent some parts of the meaning of a situation as possibilities (or
preferences), not as necessary or definite components of meaning (or
constraints). Possibilities enable the analysis and generation of nuance,
something required for faithful translation. Furthermore, the representation of
the meaning of words, especially of near-synonyms, is crucial, because it
specifies which nuances words can convey in which contexts.
| 1,998 | Computation and Language |
Comparing a statistical and a rule-based tagger for German | In this paper we present the results of comparing a statistical tagger for
German based on decision trees and a rule-based Brill-Tagger for German. We
used the same training corpus (and therefore the same tag-set) to train both
taggers. We then applied the taggers to the same test corpus and compared their
respective behavior and in particular their error rates. Both taggers perform
similarly with an error rate of around 5%. From the detailed error analysis it
can be seen that the rule-based tagger has more problems with unknown words
than the statistical tagger. But the results are opposite for tokens that are
many-ways ambiguous. If the unknown words are fed into the taggers with the
help of an external lexicon (such as the Gertwol system) the error rate of the
rule-based tagger drops to 4.7%, and the respective rate of the statistical
taggers drops to around 3.7%. Combining the taggers by using the output of one
tagger to help the other did not lead to any further improvement.
| 2,007 | Computation and Language |
P-model Alternative to the T-model | Standard linguistic analysis of syntax uses the T-model. This model requires
the ordering: D-structure $>$ S-structure $>$ LF. Between each of these
representations there is movement which alters the order of the constituent
words; movement is achieved using the principles and parameters of syntactic
theory. Psychological serial models do not accommodate the T-model immediately
so that here a new model called the P-model is introduced. Here it is argued
that the LF representation should be replaced by a variant of Frege's three
qualities. In the F-representation the order of elements is not necessarily the
same as that in LF and it is suggested that the correct ordering is:
F-representation $>$ D-structure $>$ S-structure. Within this framework
movement originates as the outcome of emphasis applied to the sentence.
| 2,007 | Computation and Language |
A Structured Language Model | The paper presents a language model that develops syntactic structure and
uses it to extract meaningful information from the word history, thus enabling
the use of long distance dependencies. The model assigns probability to every
joint sequence of words - binary-parse-structure with headword annotation. The
model, its probabilistic parametrization, and a set of experiments meant to
evaluate its predictive power are presented.
| 2,007 | Computation and Language |
A Probabilistic Approach to Lexical Semantic Knowledge Acquisition and S
tructural Disambiguation | In this thesis, I address the problem of automatically acquiring lexical
semantic knowledge, especially that of case frame patterns, from large corpus
data and using the acquired knowledge in structural disambiguation. The
approach I adopt has the following characteristics: (1) dividing the problem
into three subproblems: case slot generalization, case dependency learning, and
word clustering (thesaurus construction). (2) viewing each subproblem as that
of statistical estimation and defining probability models for each subproblem,
(3) adopting the Minimum Description Length (MDL) principle as learning
strategy, (4) employing efficient learning algorithms, and (5) viewing the
disambiguation problem as that of statistical prediction. Major contributions
of this thesis include: (1) formalization of the lexical knowledge acquisition
problem, (2) development of a number of learning methods for lexical knowledge
acquisition, and (3) development of a high-performance disambiguation method.
| 2,007 | Computation and Language |
Name Strategy: Its Existence and Implications | It is argued that colour name strategy, object name strategy, and chunking
strategy in memory are all aspects of the same general phenomena, called
stereotyping. It is pointed out that the Berlin-Kay universal partial ordering
of colours and the frequency of traffic accidents classified by colour are
surprisingly similar. Some consequences of the existence of a name strategy for
the philosophy of language and mathematics are discussed. It is argued that
real valued quantities occur {\it ab initio}. The implication of real valued
truth quantities is that the {\bf Continuum Hypothesis} of pure mathematics is
side-stepped. The existence of name strategy shows that thought/sememes and
talk/phonemes can be separate, and this vindicates the assumption of thought
occurring before talk used in psycholinguistic speech production models.
| 2,005 | Computation and Language |
A Flexible Shallow Approach to Text Generation | In order to support the efficient development of NL generation systems, two
orthogonal methods are currently pursued with emphasis: (1) reusable, general,
and linguistically motivated surface realization components, and (2) simple,
task-oriented template-based techniques. In this paper we argue that, from an
application-oriented perspective, the benefits of both are still limited. In
order to improve this situation, we suggest and evaluate shallow generation
methods associated with increased flexibility. We advise a close connection
between domain-motivated and linguistic ontologies that supports the quick
adaptation to new tasks and domains, rather than the reuse of general
resources. Our method is especially designed for generating reports with
limited linguistic variations.
| 1,998 | Computation and Language |