arxiv_id
stringlengths 10
10
| published
stringlengths 20
20
| titles
stringlengths 9
243
| authors
sequencelengths 1
389
| abstract
stringlengths 96
3.09k
| categories
sequencelengths 1
10
| selected
bool 2
classes |
---|---|---|---|---|---|---|
2305.08173 | 2023-05-14T14:46:12Z | Croatian Film Review Dataset (Cro-FiReDa): A Sentiment Annotated Dataset
of Film Reviews | [
"Gaurish Thakkar",
"Nives Mikelic Preradovic",
"Marko Tadić"
] | This paper introduces Cro-FiReDa, a sentiment-annotated dataset for Croatian
in the domain of movie reviews. The dataset, which contains over 10,000
sentences, has been annotated at the sentence level. In addition to presenting
the overall annotation process, we also present benchmark results based on the
transformer-based fine-tuning approach | [
"cs.CL"
] | false |
2305.08187 | 2023-05-14T15:53:54Z | CroSentiNews 2.0: A Sentence-Level News Sentiment Corpus | [
"Gaurish Thakkar",
"Nives Mikelic Preradović",
"Marko Tadić"
] | This article presents a sentence-level sentiment dataset for the Croatian
news domain. In addition to the 3K annotated texts already present, our dataset
contains 14.5K annotated sentence occurrences that have been tagged with 5
classes. We provide baseline scores in addition to the annotation process and
inter-annotator agreement. | [
"cs.CL"
] | false |
2305.08200 | 2023-05-14T16:52:20Z | A Cognitive Stimulation Dialogue System with Multi-source Knowledge
Fusion for Elders with Cognitive Impairment | [
"Jiyue Jiang",
"Sheng Wang",
"Qintong Li",
"Lingpeng Kong",
"Chuan Wu"
] | When communicating with elders with cognitive impairment, cognitive
stimulation (CS) help to maintain the cognitive health of elders. Data sparsity
is the main challenge in building CS-based dialogue systems, particularly in
the Chinese language. To fill this gap, we construct a Chinese CS conversation
(CSConv) dataset, which contains about 2.6K groups of dialogues with CS
principles and emotional support strategy labels. Making chit chat while
providing emotional support is overlooked by the majority of existing cognitive
dialogue systems. In this paper, we propose a multi-source knowledge fusion
method for CS dialogue (CSD), to generate open-ended responses guided by the CS
principle and emotional support strategy. We first use a progressive mask
method based on external knowledge to learn encoders as effective classifiers,
which is the prerequisite to predict the CS principle and emotional support
strategy of the target response. Then a decoder interacts with the perceived CS
principle and emotional support strategy to generate responses. Extensive
experiments conducted on the CSConv dataset demonstrate the effectiveness of
the proposed method, while there is still a large space for improvement
compared to human performance. | [
"cs.CL"
] | false |
2305.08271 | 2023-05-14T22:36:08Z | $SmartProbe$: A Virtual Moderator for Market Research Surveys | [
"Josh Seltzer",
"Jiahua Pan",
"Kathy Cheng",
"Yuxiao Sun",
"Santosh Kolagati",
"Jimmy Lin",
"Shi Zong"
] | Market research surveys are a powerful methodology for understanding consumer
perspectives at scale, but are limited by depth of understanding and insights.
A virtual moderator can introduce elements of qualitative research into
surveys, developing a rapport with survey participants and dynamically asking
probing questions, ultimately to elicit more useful information for market
researchers. In this work, we introduce ${\tt SmartProbe}$, an API which
leverages the adaptive capabilities of large language models (LLMs), and
incorporates domain knowledge from market research, in order to generate
effective probing questions in any market research survey. We outline the
modular processing flow of $\tt SmartProbe$, and evaluate the quality and
effectiveness of its generated probing questions. We believe our efforts will
inspire industry practitioners to build real-world applications based on the
latest advances in LLMs. Our demo is publicly available at
https://nexxt.in/smartprobe-demo | [
"cs.CL"
] | false |
2305.08088 | 2023-05-14T07:33:59Z | Make Prompt-based Black-Box Tuning Colorful: Boosting Model
Generalization from Three Orthogonal Perspectives | [
"Qiushi Sun",
"Chengcheng Han",
"Nuo Chen",
"Renyu Zhu",
"Jingyang Gong",
"Xiang Li",
"Ming Gao"
] | Large language models (LLMs) have shown increasing power on various natural
language processing (NLP) tasks. However, tuning these models for downstream
tasks usually needs exorbitant costs or is unavailable due to commercial
considerations. Recently, black-box tuning has been proposed to address this
problem by optimizing task-specific prompts without accessing the gradients and
hidden representations. However, most existing works have yet fully exploited
the potential of gradient-free optimization under the scenario of few-shot
learning. In this paper, we describe BBT-RGB, a suite of straightforward and
complementary techniques for enhancing the efficiency and performance of
black-box optimization. Specifically, our method includes three plug-and-play
components: (1) Two-stage derivative-free optimization strategy that
facilitates fast convergence and mitigates overfitting; (2) Automatic
verbalizer construction with its novel usage under few-shot settings; (3)
Better prompt initialization policy based on instruction search and
auto-selected demonstration. Extensive experiments across various tasks on
natural language understanding and inference demonstrate the effectiveness of
our method. Our codes are publicly available at
https://github.com/QiushiSun/BBT-RGB. | [
"cs.CL",
"cs.AI"
] | false |
2305.08096 | 2023-05-14T08:23:03Z | Towards Understanding and Improving Knowledge Distillation for Neural
Machine Translation | [
"Songming Zhang",
"Yunlong Liang",
"Shuaibo Wang",
"Wenjuan Han",
"Jian Liu",
"Jinan Xu",
"Yufeng Chen"
] | Knowledge distillation (KD) is a promising technique for model compression in
neural machine translation. However, where the knowledge hides in KD is still
not clear, which may hinder the development of KD. In this work, we first
unravel this mystery from an empirical perspective and show that the knowledge
comes from the top-1 predictions of teachers, which also helps us build a
potential connection between word- and sequence-level KD. Further, we point out
two inherent issues in vanilla word-level KD based on this finding. Firstly,
the current objective of KD spreads its focus to whole distributions to learn
the knowledge, yet lacks special treatment on the most crucial top-1
information. Secondly, the knowledge is largely covered by the golden
information due to the fact that most top-1 predictions of teachers overlap
with ground-truth tokens, which further restricts the potential of KD. To
address these issues, we propose a novel method named \textbf{T}op-1
\textbf{I}nformation \textbf{E}nhanced \textbf{K}nowledge \textbf{D}istillation
(TIE-KD). Specifically, we design a hierarchical ranking loss to enforce the
learning of the top-1 information from the teacher. Additionally, we develop an
iterative KD procedure to infuse more additional knowledge by distilling on the
data without ground-truth targets. Experiments on WMT'14 English-German, WMT'14
English-French and WMT'16 English-Romanian demonstrate that our method can
respectively boost Transformer$_{base}$ students by +1.04, +0.60 and +1.11 BLEU
scores and significantly outperform the vanilla word-level KD baseline.
Besides, our method shows higher generalizability on different teacher-student
capacity gaps than existing KD techniques. | [
"cs.CL",
"cs.AI"
] | false |
2305.08146 | 2023-05-14T12:49:16Z | ParaLS: Lexical Substitution via Pretrained Paraphraser | [
"Jipeng Qiang",
"Kang Liu",
"Yun Li",
"Yunhao Yuan",
"Yi Zhu"
] | Lexical substitution (LS) aims at finding appropriate substitutes for a
target word in a sentence. Recently, LS methods based on pretrained language
models have made remarkable progress, generating potential substitutes for a
target word through analysis of its contextual surroundings. However, these
methods tend to overlook the preservation of the sentence's meaning when
generating the substitutes. This study explores how to generate the substitute
candidates from a paraphraser, as the generated paraphrases from a paraphraser
contain variations in word choice and preserve the sentence's meaning. Since we
cannot directly generate the substitutes via commonly used decoding strategies,
we propose two simple decoding strategies that focus on the variations of the
target word during decoding. Experimental results show that our methods
outperform state-of-the-art LS methods based on pre-trained language models on
three benchmarks. | [
"cs.CL",
"cs.AI"
] | false |
2305.08883 | 2023-05-14T07:37:33Z | Watermarking Text Generated by Black-Box Language Models | [
"Xi Yang",
"Kejiang Chen",
"Weiming Zhang",
"Chang Liu",
"Yuang Qi",
"Jie Zhang",
"Han Fang",
"Nenghai Yu"
] | LLMs now exhibit human-like skills in various fields, leading to worries
about misuse. Thus, detecting generated text is crucial. However, passive
detection methods are stuck in domain specificity and limited adversarial
robustness. To achieve reliable detection, a watermark-based method was
proposed for white-box LLMs, allowing them to embed watermarks during text
generation. The method involves randomly dividing the model vocabulary to
obtain a special list and adjusting the probability distribution to promote the
selection of words in the list. A detection algorithm aware of the list can
identify the watermarked text. However, this method is not applicable in many
real-world scenarios where only black-box language models are available. For
instance, third-parties that develop API-based vertical applications cannot
watermark text themselves because API providers only supply generated text and
withhold probability distributions to shield their commercial interests. To
allow third-parties to autonomously inject watermarks into generated text, we
develop a watermarking framework for black-box language model usage scenarios.
Specifically, we first define a binary encoding function to compute a random
binary encoding corresponding to a word. The encodings computed for
non-watermarked text conform to a Bernoulli distribution, wherein the
probability of a word representing bit-1 being approximately 0.5. To inject a
watermark, we alter the distribution by selectively replacing words
representing bit-0 with context-based synonyms that represent bit-1. A
statistical test is then used to identify the watermark. Experiments
demonstrate the effectiveness of our method on both Chinese and English
datasets. Furthermore, results under re-translation, polishing, word deletion,
and synonym substitution attacks reveal that it is arduous to remove the
watermark without compromising the original semantics. | [
"cs.CL",
"cs.AI"
] | false |
2305.08067 | 2023-05-14T04:38:20Z | Improving End-to-End SLU performance with Prosodic Attention and
Distillation | [
"Shangeth Rajaa"
] | Most End-to-End SLU methods depend on the pretrained ASR or language model
features for intent prediction. However, other essential information in speech,
such as prosody, is often ignored. Recent research has shown improved results
in classifying dialogue acts by incorporating prosodic information. The margins
of improvement in these methods are minimal as the neural models ignore
prosodic features. In this work, we propose prosody-attention, which uses the
prosodic features differently to generate attention maps across time frames of
the utterance. Then we propose prosody-distillation to explicitly learn the
prosodic information in the acoustic encoder rather than concatenating the
implicit prosodic features. Both the proposed methods improve the baseline
results, and the prosody-distillation method gives an intent classification
accuracy improvement of 8\% and 2\% on SLURP and STOP datasets over the prosody
baseline. | [
"cs.CL",
"cs.LG",
"cs.SD",
"eess.AS"
] | false |
2305.08227 | 2023-05-14T19:09:35Z | DeepFilterNet: Perceptually Motivated Real-Time Speech Enhancement | [
"Hendrik Schröter",
"Tobias Rosenkranz",
"Alberto N. Escalante-B.",
"Andreas Maier"
] | Multi-frame algorithms for single-channel speech enhancement are able to take
advantage from short-time correlations within the speech signal. Deep Filtering
(DF) was proposed to directly estimate a complex filter in frequency domain to
take advantage of these correlations. In this work, we present a real-time
speech enhancement demo using DeepFilterNet. DeepFilterNet's efficiency is
enabled by exploiting domain knowledge of speech production and psychoacoustic
perception. Our model is able to match state-of-the-art speech enhancement
benchmarks while achieving a real-time-factor of 0.19 on a single threaded
notebook CPU. The framework as well as pretrained weights have been published
under an open source license. | [
"eess.AS",
"cs.CL",
"cs.SD"
] | false |
2305.08246 | 2023-05-14T20:57:11Z | Learning Non-linguistic Skills without Sacrificing Linguistic
Proficiency | [
"Mandar Sharma",
"Nikhil Muralidhar",
"Naren Ramakrishnan"
] | The field of Math-NLP has witnessed significant growth in recent years,
motivated by the desire to expand LLM performance to the learning of
non-linguistic notions (numerals, and subsequently, arithmetic reasoning).
However, non-linguistic skill injection typically comes at a cost for LLMs: it
leads to catastrophic forgetting of core linguistic skills, a consequence that
often remains unaddressed in the literature. As Math-NLP has been able to
create LLMs that can closely approximate the mathematical skills of a
grade-schooler or the arithmetic reasoning skills of a calculator, the
practicality of these models fail if they concomitantly shed their linguistic
capabilities. In this work, we take a closer look into the phenomena of
catastrophic forgetting as it pertains to LLMs and subsequently offer a novel
framework for non-linguistic skill injection for LLMs based on information
theoretic interventions and skill-specific losses that enable the learning of
strict arithmetic reasoning. Our model outperforms the state-of-the-art both on
injected non-linguistic skills and on linguistic knowledge retention, and does
so with a fraction of the non-linguistic training data (1/4) and zero
additional synthetic linguistic training data. | [
"cs.CL",
"cs.AI",
"cs.LG"
] | false |
2305.08264 | 2023-05-14T22:01:24Z | MatSci-NLP: Evaluating Scientific Language Models on Materials Science
Language Tasks Using Text-to-Schema Modeling | [
"Yu Song",
"Santiago Miret",
"Bang Liu"
] | We present MatSci-NLP, a natural language benchmark for evaluating the
performance of natural language processing (NLP) models on materials science
text. We construct the benchmark from publicly available materials science text
data to encompass seven different NLP tasks, including conventional NLP tasks
like named entity recognition and relation classification, as well as NLP tasks
specific to materials science, such as synthesis action retrieval which relates
to creating synthesis procedures for materials. We study various BERT-based
models pretrained on different scientific text corpora on MatSci-NLP to
understand the impact of pretraining strategies on understanding materials
science text. Given the scarcity of high-quality annotated data in the
materials science domain, we perform our fine-tuning experiments with limited
training data to encourage the generalize across MatSci-NLP tasks. Our
experiments in this low-resource training setting show that language models
pretrained on scientific text outperform BERT trained on general text. MatBERT,
a model pretrained specifically on materials science journals, generally
performs best for most tasks. Moreover, we propose a unified text-to-schema for
multitask learning on \benchmark and compare its performance with traditional
fine-tuning methods. In our analysis of different training methods, we find
that our proposed text-to-schema methods inspired by question-answering
consistently outperform single and multitask NLP fine-tuning methods. The code
and datasets are publicly available at
\url{https://github.com/BangLab-UdeM-Mila/NLP4MatSci-ACL23}. | [
"cs.CL",
"cond-mat.mtrl-sci",
"cs.AI"
] | false |
2305.08073 | 2023-05-14T05:11:52Z | HiPerformer: Hierarchically Permutation-Equivariant Transformer for Time
Series Forecasting | [
"Ryo Umagami",
"Yu Ono",
"Yusuke Mukuta",
"Tatsuya Harada"
] | It is imperative to discern the relationships between multiple time series
for accurate forecasting. In particular, for stock prices, components are often
divided into groups with the same characteristics, and a model that extracts
relationships consistent with this group structure should be effective. Thus,
we propose the concept of hierarchical permutation-equivariance, focusing on
index swapping of components within and among groups, to design a model that
considers this group structure. When the prediction model has hierarchical
permutation-equivariance, the prediction is consistent with the group
relationships of the components. Therefore, we propose a hierarchically
permutation-equivariant model that considers both the relationship among
components in the same group and the relationship among groups. The experiments
conducted on real-world data demonstrate that the proposed method outperforms
existing state-of-the-art methods. | [
"cs.LG"
] | false |
2305.08105 | 2023-05-14T08:51:44Z | Blockchain Transaction Fee Forecasting: A Comparison of Machine Learning
Methods | [
"Conall Butler",
"Martin Crane"
] | Gas is the transaction-fee metering system of the Ethereum network. Users of
the network are required to select a gas price for submission with their
transaction, creating a risk of overpaying or delayed/unprocessed transactions
in this selection. In this work, we investigate data in the aftermath of the
London Hard Fork and shed insight into the transaction dynamics of the net-work
after this major fork. As such, this paper provides an update on work previous
to 2019 on the link between EthUSD BitUSD and gas price. For forecasting, we
compare a novel combination of machine learning methods such as Direct
Recursive Hybrid LSTM, CNNLSTM, and Attention LSTM. These are combined with
wavelet threshold denoising and matrix profile data processing toward the
forecasting of block minimum gas price, on a 5-min timescale, over multiple
lookaheads. As the first application of the matrix profile being applied to gas
price data and forecasting we are aware of, this study demonstrates that matrix
profile data can enhance attention-based models however, given the hardware
constraints, hybrid models outperformed attention and CNNLSTM models. The
wavelet coherence of inputs demonstrates correlation in multiple variables on a
1 day timescale, which is a deviation of base free from gas price. A
Direct-Recursive Hybrid LSTM strategy outperforms other models. Hybrid models
have favourable performance up to a 20 min lookahead with performance being
comparable to attention models when forecasting 25/50-min ahead. Forecasts over
a range of lookaheads allow users to make an informed decision on gas price
selection and the optimal window to submit their transaction in without fear of
their transaction being rejected. This, in turn, gives more detailed insight
into gas price dynamics than existing recommenders, oracles and forecasting
approaches, which provide simple heuristics or limited lookahead horizons. | [
"cs.LG",
"62H20, 65C20, 68T01, 91B84"
] | false |
2305.08120 | 2023-05-14T10:46:20Z | Unraveling Cold Start Enigmas in Predictive Analytics for OTT Media:
Synergistic Meta-Insights and Multimodal Ensemble Mastery | [
"K. Ganguly",
"A. Patra"
] | The cold start problem is a common challenge in various domains, including
media use cases such as predicting viewership for newly launched shows on
Over-The-Top (OTT) platforms. In this study, we propose a generic approach to
tackle cold start problems by leveraging metadata and employing multi-model
ensemble techniques. Our methodology includes feature engineering, model
selection, and an ensemble approach based on a weighted average of predictions.
The performance of our proposed method is evaluated using various performance
metrics. Our results indicate that the multi-model ensemble approach
significantly improves prediction accuracy compared to individual models. | [
"cs.LG"
] | false |
2305.08139 | 2023-05-14T12:20:13Z | Predicting Unplanned Readmissions in the Intensive Care Unit: A
Multimodality Evaluation | [
"Eitam Sheetrit",
"Menachem Brief",
"Oren Elisha"
] | A hospital readmission is when a patient who was discharged from the hospital
is admitted again for the same or related care within a certain period.
Hospital readmissions are a significant problem in the healthcare domain, as
they lead to increased hospitalization costs, decreased patient satisfaction,
and increased risk of adverse outcomes such as infections, medication errors,
and even death. The problem of hospital readmissions is particularly acute in
intensive care units (ICUs), due to the severity of the patients' conditions,
and the substantial risk of complications. Predicting Unplanned Readmissions in
ICUs is a challenging task, as it involves analyzing different data modalities,
such as static data, unstructured free text, sequences of diagnoses and
procedures, and multivariate time-series. Here, we investigate the
effectiveness of each data modality separately, then alongside with others,
using state-of-the-art machine learning approaches in time-series analysis and
natural language processing. Using our evaluation process, we are able to
determine the contribution of each data modality, and for the first time in the
context of readmission, establish a hierarchy of their predictive value.
Additionally, we demonstrate the impact of Temporal Abstractions in enhancing
the performance of time-series approaches to readmission prediction. Due to
conflicting definitions in the literature, we also provide a clear definition
of the term Unplanned Readmission to enhance reproducibility and consistency of
future research and to prevent any potential misunderstandings that could
result from diverse interpretations of the term. Our experimental results on a
large benchmark clinical data set show that Discharge Notes written by
physicians, have better capabilities for readmission prediction than all other
modalities. | [
"cs.LG"
] | false |
2305.08164 | 2023-05-14T14:21:58Z | Latent Processes Identification From Multi-View Time Series | [
"Zenan Huang",
"Haobo Wang",
"Junbo Zhao",
"Nenggan Zheng"
] | Understanding the dynamics of time series data typically requires identifying
the unique latent factors for data generation, \textit{a.k.a.}, latent
processes identification. Driven by the independent assumption, existing works
have made great progress in handling single-view data. However, it is a
non-trivial problem that extends them to multi-view time series data because of
two main challenges: (i) the complex data structure, such as temporal
dependency, can result in violation of the independent assumption; (ii) the
factors from different views are generally overlapped and are hard to be
aggregated to a complete set. In this work, we propose a novel framework MuLTI
that employs the contrastive learning technique to invert the data generative
process for enhanced identifiability. Additionally, MuLTI integrates a
permutation mechanism that merges corresponding overlapped variables by the
establishment of an optimal transport formula. Extensive experimental results
on synthetic and real-world datasets demonstrate the superiority of our method
in recovering identifiable latent variables on multi-view time series. | [
"cs.LG"
] | false |
2305.08036 | 2023-05-14T00:42:24Z | Small-data Reduced Order Modeling of Chaotic Dynamics through SyCo-AE:
Synthetically Constrained Autoencoders | [
"Andrey A. Popov",
"Renato Zanetti"
] | Data-driven reduced order modeling of chaotic dynamics can result in systems
that either dissipate or diverge catastrophically. Leveraging non-linear
dimensionality reduction of autoencoders and the freedom of non-linear operator
inference with neural-networks, we aim to solve this problem by imposing a
synthetic constraint in the reduced order space. The synthetic constraint
allows our reduced order model both the freedom to remain fully non-linear and
highly unstable while preventing divergence. We illustrate the methodology with
the classical 40-variable Lorenz '96 equations, showing that our methodology is
capable of producing medium-to-long range forecasts with lower error using less
data. | [
"cs.LG",
"math.DS"
] | false |
2305.08044 | 2023-05-14T02:01:54Z | Using EEG Signals to Assess Workload during Memory Retrieval in a
Real-world Scenario | [
"Kuan-Jung Chiang",
"Steven Dong",
"Chung-Kuan Cheng",
"Tzyy-Ping Jung"
] | Objective: The Electroencephalogram (EEG) is gaining popularity as a
physiological measure for neuroergonomics in human factor studies because it is
objective, less prone to bias, and capable of assessing the dynamics of
cognitive states. This study investigated the associations between memory
workload and EEG during participants' typical office tasks on a single-monitor
and dual-monitor arrangement. We expect a higher memory workload for the
single-monitor arrangement. Approach: We designed an experiment that mimics the
scenario of a subject performing some office work and examined whether the
subjects experienced various levels of memory workload in two different office
setups: 1) a single-monitor setup and 2) a dual-monitor setup. We used EEG band
power, mutual information, and coherence as features to train machine learning
models to classify high versus low memory workload states. Main results: The
study results showed that these characteristics exhibited significant
differences that were consistent across all participants. We also verified the
robustness and consistency of these EEG signatures in a different data set
collected during a Sternberg task in a prior study. Significance: The study
found the EEG correlates of memory workload across individuals, demonstrating
the effectiveness of using EEG analysis in conducting real-world neuroergonomic
studies. | [
"cs.HC",
"cs.LG"
] | false |
2305.08048 | 2023-05-14T03:05:14Z | Towards Understanding the Generalization of Graph Neural Networks | [
"Huayi Tang",
"Yong Liu"
] | Graph neural networks (GNNs) are the most widely adopted model in
graph-structured data oriented learning and representation. Despite their
extraordinary success in real-world applications, understanding their working
mechanism by theory is still on primary stage. In this paper, we move towards
this goal from the perspective of generalization. To be specific, we first
establish high probability bounds of generalization gap and gradients in
transductive learning with consideration of stochastic optimization. After
that, we provide high probability bounds of generalization gap for popular
GNNs. The theoretical results reveal the architecture specific factors
affecting the generalization gap. Experimental results on benchmark datasets
show the consistency between theoretical results and empirical evidence. Our
results provide new insights in understanding the generalization of GNNs. | [
"cs.LG",
"cs.AI"
] | false |
2305.08100 | 2023-05-14T08:29:15Z | Conditional mean embeddings and optimal feature selection via positive
definite kernels | [
"Palle E. T. Jorgensen",
"Myung-Sin Song",
"James Tian"
] | Motivated by applications, we consider here new operator theoretic approaches
to Conditional mean embeddings (CME). Our present results combine a spectral
analysis-based optimization scheme with the use of kernels, stochastic
processes, and constructive learning algorithms. For initially given non-linear
data, we consider optimization-based feature selections. This entails the use
of convex sets of positive definite (p.d.) kernels in a construction of optimal
feature selection via regression algorithms from learning models. Thus, with
initial inputs of training data (for a suitable learning algorithm,) each
choice of p.d. kernel $K$ in turn yields a variety of Hilbert spaces and
realizations of features. A novel idea here is that we shall allow an
optimization over selected sets of kernels $K$ from a convex set $C$ of
positive definite kernels $K$. Hence our \textquotedblleft
optimal\textquotedblright{} choices of feature representations will depend on a
secondary optimization over p.d. kernels $K$ within a specified convex set $C$. | [
"cs.LG",
"math.FA",
"Primary: 47N10, 47A52, 47B32. Secondary: 42A82, 42C15, 62H12, 62J07,\n 65J20, 68T07, 90C20"
] | false |
2305.08102 | 2023-05-14T08:33:11Z | A machine learning-based viscoelastic-viscoplastic model for epoxy
nanocomposites with moisture content | [
"Betim Bahtiri",
"Behrouz Arash",
"Sven Scheffler",
"Maximilian Jux",
"Raimund Rolfes"
] | In this work, we propose a deep learning (DL)-based constitutive model for
investigating the cyclic viscoelastic-viscoplastic-damage behavior of
nanoparticle/epoxy nanocomposites with moisture content. For this, a long
short-term memory network is trained using a combined framework of a sampling
technique and a perturbation method. The training framework, along with the
training data generated by an experimentally validated
viscoelastic-viscoplastic model, enables the DL model to accurately capture the
rate-dependent stress-strain relationship and consistent tangent moduli. In
addition, the DL-based constitutive model is implemented into finite element
analysis. Finite element simulations are performed to study the effect of load
rate and moisture content on the force-displacement response of nanoparticle/
epoxy samples. Numerical examples show that the computational efficiency of the
DL model depends on the loading condition and is significantly higher than the
conventional constitutive model. Furthermore, comparing numerical results and
experimental data demonstrates good agreement with different nanoparticle and
moisture contents. | [
"cs.LG",
"cs.CE"
] | false |
2305.08115 | 2023-05-14T10:15:35Z | Automatic Generation of Attention Rules For Containment of Machine
Learning Model Errors | [
"Samuel Ackerman",
"Axel Bendavid",
"Eitan Farchi",
"Orna Raz"
] | Machine learning (ML) solutions are prevalent in many applications. However,
many challenges exist in making these solutions business-grade. For instance,
maintaining the error rate of the underlying ML models at an acceptably low
level. Typically, the true relationship between feature inputs and the target
feature to be predicted is uncertain, and hence statistical in nature. The
approach we propose is to separate the observations that are the most likely to
be predicted incorrectly into 'attention sets'. These can directly aid model
diagnosis and improvement, and be used to decide on alternative courses of
action for these problematic observations. We present several algorithms
(`strategies') for determining optimal rules to separate these observations. In
particular, we prefer strategies that use feature-based slicing because they
are human-interpretable, model-agnostic, and require minimal supplementary
inputs or knowledge. In addition, we show that these strategies outperform
several common baselines, such as selecting observations with prediction
confidence below a threshold. To evaluate strategies, we introduce metrics to
measure various desired qualities, such as their performance, stability, and
generalizability to unseen data; the strategies are evaluated on several
publicly-available datasets. We use TOPSIS, a Multiple Criteria Decision Making
method, to aggregate these metrics into a single quality score for each
strategy, to allow comparison. | [
"cs.LG",
"stat.AP"
] | false |
2305.08130 | 2023-05-14T11:49:37Z | Inverse Reinforcement Learning With Constraint Recovery | [
"Nirjhar Das",
"Arpan Chattopadhyay"
] | In this work, we propose a novel inverse reinforcement learning (IRL)
algorithm for constrained Markov decision process (CMDP) problems. In standard
IRL problems, the inverse learner or agent seeks to recover the reward function
of the MDP, given a set of trajectory demonstrations for the optimal policy. In
this work, we seek to infer not only the reward functions of the CMDP, but also
the constraints. Using the principle of maximum entropy, we show that the IRL
with constraint recovery (IRL-CR) problem can be cast as a constrained
non-convex optimization problem. We reduce it to an alternating constrained
optimization problem whose sub-problems are convex. We use exponentiated
gradient descent algorithm to solve it. Finally, we demonstrate the efficacy of
our algorithm for the grid world environment. | [
"cs.LG",
"cs.AI"
] | false |
2305.08273 | 2023-05-14T23:00:10Z | Decoupled Graph Neural Networks for Large Dynamic Graphs | [
"Yanping Zheng",
"Zhewei Wei",
"Jiajun Liu"
] | Real-world graphs, such as social networks, financial transactions, and
recommendation systems, often demonstrate dynamic behavior. This phenomenon,
known as graph stream, involves the dynamic changes of nodes and the emergence
and disappearance of edges. To effectively capture both the structural and
temporal aspects of these dynamic graphs, dynamic graph neural networks have
been developed. However, existing methods are usually tailored to process
either continuous-time or discrete-time dynamic graphs, and cannot be
generalized from one to the other. In this paper, we propose a decoupled graph
neural network for large dynamic graphs, including a unified dynamic
propagation that supports efficient computation for both continuous and
discrete dynamic graphs. Since graph structure-related computations are only
performed during the propagation process, the prediction process for the
downstream task can be trained separately without expensive graph computations,
and therefore any sequence model can be plugged-in and used. As a result, our
algorithm achieves exceptional scalability and expressiveness. We evaluate our
algorithm on seven real-world datasets of both continuous-time and
discrete-time dynamic graphs. The experimental results demonstrate that our
algorithm achieves state-of-the-art performance in both kinds of dynamic
graphs. Most notably, the scalability of our algorithm is well illustrated by
its successful application to large graphs with up to over a billion temporal
edges and over a hundred million nodes. | [
"cs.LG",
"cs.SI"
] | false |
2305.08057 | 2023-05-14T03:50:46Z | CREMP: Conformer-Rotamer Ensembles of Macrocyclic Peptides for Machine
Learning | [
"Colin A. Grambow",
"Hayley Weir",
"Christian N. Cunningham",
"Tommaso Biancalani",
"Kangway V. Chuang"
] | Computational and machine learning approaches to model the conformational
landscape of macrocyclic peptides have the potential to enable rational design
and optimization. However, accurate, fast, and scalable methods for modeling
macrocycle geometries remain elusive. Recent deep learning approaches have
significantly accelerated protein structure prediction and the generation of
small-molecule conformational ensembles, yet similar progress has not been made
for macrocyclic peptides due to their unique properties. Here, we introduce
CREMP, a resource generated for the rapid development and evaluation of machine
learning models for macrocyclic peptides. CREMP contains 36,198 unique
macrocyclic peptides and their high-quality structural ensembles generated
using the Conformer-Rotamer Ensemble Sampling Tool (CREST). Altogether, this
new dataset contains nearly 31.3 million unique macrocycle geometries, each
annotated with energies derived from semi-empirical extended tight-binding
(xTB) DFT calculations. We anticipate that this dataset will enable the
development of machine learning models that can improve peptide design and
optimization for novel therapeutics. | [
"q-bio.BM",
"cs.LG",
"physics.chem-ph"
] | false |
2305.08077 | 2023-05-14T05:53:39Z | Optimization of Residential Demand Response Program Cost with
Consideration for Occupants Thermal Comfort and Privacy | [
"Reza Nematirad",
"M. M. Ardehali",
"Amir Khorsandi"
] | Residential consumers can use the demand response program (DRP) if they can
utilize the home energy management system (HEMS), which reduces consumer costs
by automatically adjusting air conditioning (AC) setpoints and shifting some
appliances to off-peak hours. If HEMS knows occupancy status, consumers can
gain more economic benefits and thermal comfort. However, for the building
occupancy status, direct sensing is costly, inaccurate, and intrusive for
residents. So, forecasting algorithms could serve as an effective alternative.
The goal of this study is to present a non-intrusive, accurate, and
cost-effective approach, to develop a multi-objective simulation model for the
application of DRPs in a smart residential house, where (a) electrical load
demand reduction, (b) adjustment in thermal comfort (AC) temperature setpoints,
and (c) , worst cases scenario approach is very conservative. Because that is
unlikely all uncertain parameters take their worst values at all times. So, the
flexible robust counterpart optimization along with uncertainty budgets is
developed to consider uncertainty realistically. Simulated results indicate
that considering uncertainty increases the costs by 36 percent and decreases
the AC temperature setpoints. Besides, using DRPs reduces demand by shifting
some appliance operations to off-peak hours and lowers costs by 13.2 percent. | [
"eess.SY",
"cs.LG",
"cs.SY"
] | false |
2305.08126 | 2023-05-14T11:16:17Z | Semantic Communication of Learnable Concepts | [
"Francesco Pase",
"Szymon Kobus",
"Deniz Gunduz",
"Michele Zorzi"
] | We consider the problem of communicating a sequence of concepts, i.e.,
unknown and potentially stochastic maps, which can be observed only through
examples, i.e., the mapping rules are unknown. The transmitter applies a
learning algorithm to the available examples, and extracts knowledge from the
data by optimizing a probability distribution over a set of models, i.e., known
functions, which can better describe the observed data, and so potentially the
underlying concepts. The transmitter then needs to communicate the learned
models to a remote receiver through a rate-limited channel, to allow the
receiver to decode the models that can describe the underlying sampled concepts
as accurately as possible in their semantic space. After motivating our
analysis, we propose the formal problem of communicating concepts, and provide
its rate-distortion characterization, pointing out its connection with the
concepts of empirical and strong coordination in a network. We also provide a
bound for the distortion-rate function. | [
"cs.IT",
"cs.LG",
"math.IT"
] | false |
2305.08169 | 2023-05-14T14:37:33Z | Can Learning Deteriorate Control? Analyzing Computational Delays in
Gaussian Process-Based Event-Triggered Online Learning | [
"Xiaobing Dai",
"Armin Lederer",
"Zewen Yang",
"Sandra Hirche"
] | When the dynamics of systems are unknown, supervised machine learning
techniques are commonly employed to infer models from data. Gaussian process
(GP) regression is a particularly popular learning method for this purpose due
to the existence of prediction error bounds. Moreover, GP models can be
efficiently updated online, such that event-triggered online learning
strategies can be pursued to ensure specified tracking accuracies. However,
existing trigger conditions must be able to be evaluated at arbitrary times,
which cannot be achieved in practice due to non-negligible computation times.
Therefore, we first derive a delay-aware tracking error bound, which reveals an
accuracy-delay trade-off. Based on this result, we propose a novel event
trigger for GP-based online learning with computational delays, which we show
to offer advantages over offline trained GP models for sufficiently small
computation times. Finally, we demonstrate the effectiveness of the proposed
event trigger for online learning in simulations. | [
"eess.SY",
"cs.LG",
"cs.SY"
] | false |
2305.08197 | 2023-05-14T16:24:09Z | A Dataset Fusion Algorithm for Generalised Anomaly Detection in
Homogeneous Periodic Time Series Datasets | [
"Ayman Elhalwagy",
"Tatiana Kalganova"
] | The generalisation of Neural Networks (NN) to multiple datasets is often
overlooked in literature due to NNs typically being optimised for specific data
sources. This becomes especially challenging in time-series-based multi-dataset
models due to difficulties in fusing sequential data from different sensors and
collection specifications. In a commercial environment, however, generalisation
can effectively utilise available data and computational power, which is
essential in the context of Green AI, the sustainable development of AI models.
This paper introduces "Dataset Fusion," a novel dataset composition algorithm
for fusing periodic signals from multiple homogeneous datasets into a single
dataset while retaining unique features for generalised anomaly detection. The
proposed approach, tested on a case study of 3-phase current data from 2
different homogeneous Induction Motor (IM) fault datasets using an unsupervised
LSTMCaps NN, significantly outperforms conventional training approaches with an
Average F1 score of 0.879 and effectively generalises across all datasets. The
proposed approach was also tested with varying percentages of the training
data, in line with the principles of Green AI. Results show that using only
6.25\% of the training data, translating to a 93.7\% reduction in computational
power, results in a mere 4.04\% decrease in performance, demonstrating the
advantages of the proposed approach in terms of both performance and
computational efficiency. Moreover, the algorithm's effectiveness under
non-ideal conditions highlights its potential for practical use in real-world
applications. | [
"cs.LG",
"cs.AI",
"eess.SP"
] | false |
2305.08226 | 2023-05-14T19:07:21Z | NLP-based Cross-Layer 5G Vulnerabilities Detection via Fuzzing Generated
Run-Time Profiling | [
"Zhuzhu Wang",
"Ying Wang"
] | The effectiveness and efficiency of 5G software stack vulnerability and
unintended behavior detection are essential for 5G assurance, especially for
its applications in critical infrastructures. Scalability and automation are
the main challenges in testing approaches and cybersecurity research. In this
paper, we propose an innovative approach for automatically detecting
vulnerabilities, unintended emergent behaviors, and performance degradation in
5G stacks via run-time profiling documents corresponding to fuzz testing in
code repositories. Piloting on srsRAN, we map the run-time profiling via
Logging Information (LogInfo) generated by fuzzing test to a high dimensional
metric space first and then construct feature spaces based on their timestamp
information. Lastly, we further leverage machine learning-based classification
algorithms, including Logistic Regression, K-Nearest Neighbors, and Random
Forest to categorize the impacts on performance and security attributes. The
performance of the proposed approach has high accuracy, ranging from $ 93.4 \%
$ to $ 95.9 \% $, in detecting the fuzzing impacts. In addition, the proof of
concept could identify and prioritize real-time vulnerabilities on 5G
infrastructures and critical applications in various verticals. | [
"cs.CR",
"cs.LG",
"cs.SE"
] | false |
2305.08885 | 2023-05-14T22:22:16Z | Smart Home Energy Management: VAE-GAN synthetic dataset generator and
Q-learning | [
"Mina Razghandi",
"Hao Zhou",
"Melike Erol-Kantarci",
"Damla Turgut"
] | Recent years have noticed an increasing interest among academia and industry
towards analyzing the electrical consumption of residential buildings and
employing smart home energy management systems (HEMS) to reduce household
energy consumption and costs. HEMS has been developed to simulate the
statistical and functional properties of actual smart grids. Access to publicly
available datasets is a major challenge in this type of research. The potential
of artificial HEMS applications will be further enhanced with the development
of time series that represent different operating conditions of the synthetic
systems. In this paper, we propose a novel variational auto-encoder-generative
adversarial network (VAE-GAN) technique for generating time-series data on
energy consumption in smart homes. We also explore how the generative model
performs when combined with a Q-learning-based HEMS. We tested the online
performance of Q-learning-based HEMS with real-world smart home data. To test
the generated dataset, we measure the Kullback-Leibler (KL) divergence, maximum
mean discrepancy (MMD), and the Wasserstein distance between the probability
distributions of the real and synthetic data. Our experiments show that
VAE-GAN-generated synthetic data closely matches the real data distribution.
Finally, we show that the generated data allows for the training of a
higher-performance Q-learning-based HEMS compared to datasets generated with
baseline approaches. | [
"cs.LG",
"cs.SY",
"eess.SY"
] | false |
2305.18306 | 2023-05-14T20:31:37Z | Multi-View Interactive Collaborative Filtering | [
"Maria Lentini",
"Umashanger Thayasivam"
] | In many scenarios, recommender system user interaction data such as clicks or
ratings is sparse, and item turnover rates (e.g., new articles, job postings)
high. Given this, the integration of contextual "side" information in addition
to user-item ratings is highly desirable. Whilst there are algorithms that can
handle both rating and contextual data simultaneously, these algorithms are
typically limited to making only in-sample recommendations, suffer from the
curse of dimensionality, and do not incorporate multi-armed bandit (MAB)
policies for long-term cumulative reward optimization. We propose multi-view
interactive topic regression (MV-ICTR) a novel partially online latent factor
recommender algorithm that incorporates both rating and contextual information
to model item-specific feature dependencies and users' personal preferences
simultaneously, with multi-armed bandit policies for continued online
personalization. The result is significantly increased performance on datasets
with high percentages of cold-start users and items. | [
"cs.IR",
"cs.AI",
"cs.LG"
] | false |
2305.08104 | 2023-05-14T08:48:02Z | Federated TD Learning over Finite-Rate Erasure Channels: Linear Speedup
under Markovian Sampling | [
"Nicolò Dal Fabbro",
"Aritra Mitra",
"George J. Pappas"
] | Federated learning (FL) has recently gained much attention due to its
effectiveness in speeding up supervised learning tasks under communication and
privacy constraints. However, whether similar speedups can be established for
reinforcement learning remains much less understood theoretically. Towards this
direction, we study a federated policy evaluation problem where agents
communicate via a central aggregator to expedite the evaluation of a common
policy. To capture typical communication constraints in FL, we consider finite
capacity up-link channels that can drop packets based on a Bernoulli erasure
model. Given this setting, we propose and analyze QFedTD - a quantized
federated temporal difference learning algorithm with linear function
approximation. Our main technical contribution is to provide a finite-sample
analysis of QFedTD that (i) highlights the effect of quantization and erasures
on the convergence rate; and (ii) establishes a linear speedup w.r.t. the
number of agents under Markovian sampling. Notably, while different
quantization mechanisms and packet drop models have been extensively studied in
the federated learning, distributed optimization, and networked control systems
literature, our work is the first to provide a non-asymptotic analysis of their
effects in multi-agent and federated reinforcement learning. | [
"cs.LG",
"cs.AI",
"cs.MA",
"cs.SY",
"eess.SY",
"math.OC"
] | false |
2305.08336 | 2023-05-15T04:03:11Z | Inverse Rendering of Translucent Objects using Physical and Neural
Renderers | [
"Chenhao Li",
"Trung Thanh Ngo",
"Hajime Nagahara"
] | In this work, we propose an inverse rendering model that estimates 3D shape,
spatially-varying reflectance, homogeneous subsurface scattering parameters,
and an environment illumination jointly from only a pair of captured images of
a translucent object. In order to solve the ambiguity problem of inverse
rendering, we use a physically-based renderer and a neural renderer for scene
reconstruction and material editing. Because two renderers are differentiable,
we can compute a reconstruction loss to assist parameter estimation. To enhance
the supervision of the proposed neural renderer, we also propose an augmented
loss. In addition, we use a flash and no-flash image pair as the input. To
supervise the training, we constructed a large-scale synthetic dataset of
translucent objects, which consists of 117K scenes. Qualitative and
quantitative results on both synthetic and real-world datasets demonstrated the
effectiveness of the proposed model. | [
"cs.CV"
] | false |
2305.08366 | 2023-05-15T05:59:35Z | CLRerNet: Improving Confidence of Lane Detection with LaneIoU | [
"Hiroto Honda",
"Yusuke Uchida"
] | Lane marker detection is a crucial component of the autonomous driving and
driver assistance systems. Modern deep lane detection methods with row-based
lane representation exhibit excellent performance on lane detection benchmarks.
Through preliminary oracle experiments, we firstly disentangle the lane
representation components to determine the direction of our approach. We show
that correct lane positions are already among the predictions of an existing
row-based detector, and the confidence scores that accurately represent
intersection-over-union (IoU) with ground truths are the most beneficial. Based
on the finding, we propose LaneIoU that better correlates with the metric, by
taking the local lane angles into consideration. We develop a novel detector
coined CLRerNet featuring LaneIoU for the target assignment cost and loss
functions aiming at the improved quality of confidence scores. Through careful
and fair benchmark including cross validation, we demonstrate that CLRerNet
outperforms the state-of-the-art by a large margin - enjoying F1 score of
81.43% compared with 80.47% of the existing method on CULane, and 86.47%
compared with 86.10% on CurveLanes. | [
"cs.CV"
] | false |
2305.08386 | 2023-05-15T06:49:00Z | PLIP: Language-Image Pre-training for Person Representation Learning | [
"Jialong Zuo",
"Changqian Yu",
"Nong Sang",
"Changxin Gao"
] | Pre-training has emerged as an effective technique for learning powerful
person representations. Most existing methods have shown that pre-training on
pure-vision large-scale datasets like ImageNet and LUPerson has achieved
remarkable performance. However, solely relying on visual information, the
absence of robust explicit indicators poses a challenge for these methods to
learn discriminative person representations. Drawing inspiration from the
intrinsic fine-grained attribute indicators of person descriptions, we explore
introducing the language modality into person representation learning. To this
end, we propose a novel language-image pre-training framework for person
representation learning, termed PLIP. To explicitly build fine-grained
cross-modal associations, we specifically design three pretext tasks, \ie
semantic-fused image colorization, visual-fused attributes prediction, and
vision-language matching. In addition, due to the lack of an appropriate
dataset, we present a large-scale person dataset named SYNTH-PEDES, where the
Stylish Pedestrian Attributes-union Captioning method is proposed to synthesize
diverse textual descriptions. We pre-train PLIP on SYNTH-PEDES and evaluate our
model by spanning downstream tasks such as text-based Re-ID, image-based Re-ID,
and person attribute recognition. Extensive experiments demonstrate that our
model not only significantly improves existing methods on all these tasks, but
also shows great ability in the few-shot and domain generalization settings.
The code, dataset and weights will be released
at~\url{https://github.com/Zplusdragon/PLIP} | [
"cs.CV"
] | false |
2305.08418 | 2023-05-15T07:57:55Z | Online Sequence Clustering Algorithm for Video Trajectory Analysis | [
"Aximu Yuemaier",
"Xiaogang Chen",
"Xingyu Qian",
"Longfei Liang",
"Shunfeng Li",
"Zhitang Song"
] | Target tracking and trajectory modeling have important applications in
surveillance video analysis and have received great attention in the fields of
road safety and community security. In this work, we propose a lightweight
real-time video analysis scheme that uses a model learned from motion patterns
to monitor the behavior of objects, which can be used for applications such as
real-time representation and prediction. The proposed sequence clustering
algorithm based on discrete sequences makes the system have continuous online
learning ability. The intrinsic repeatability of the target object trajectory
is used to automatically construct the behavioral model in the three processes
of feature extraction, cluster learning, and model application. In addition to
the discretization of trajectory features and simple model applications, this
paper focuses on online clustering algorithms and their incremental learning
processes. Finally, through the learning of the trajectory model of the actual
surveillance video image, the feasibility of the algorithm is verified. And the
characteristics and performance of the clustering algorithm are discussed in
the analysis. This scheme has real-time online learning and processing of
motion models while avoiding a large number of arithmetic operations, which is
more in line with the application scenarios of front-end intelligent
perception. | [
"cs.CV"
] | false |
2305.08439 | 2023-05-15T08:36:32Z | Exploiting Frequency Spectrum of Adversarial Images for General
Robustness | [
"Chun Yang Tan",
"Kazuhiko Kawamoto",
"Hiroshi Kera"
] | In recent years, there has been growing concern over the vulnerability of
convolutional neural networks (CNNs) to image perturbations. However, achieving
general robustness against different types of perturbations remains
challenging, in which enhancing robustness to some perturbations (e.g.,
adversarial perturbations) may degrade others (e.g., common corruptions). In
this paper, we demonstrate that adversarial training with an emphasis on phase
components significantly improves model performance on clean, adversarial, and
common corruption accuracies. We propose a frequency-based data augmentation
method, Adversarial Amplitude Swap, that swaps the amplitude spectrum between
clean and adversarial images to generate two novel training images: adversarial
amplitude and adversarial phase images. These images act as substitutes for
adversarial images and can be implemented in various adversarial training
setups. Through extensive experiments, we demonstrate that our method enables
the CNNs to gain general robustness against different types of perturbations
and results in a uniform performance against all types of common corruptions. | [
"cs.CV"
] | false |
2305.08509 | 2023-05-15T10:18:52Z | Component-aware anomaly detection framework for adjustable and logical
industrial visual inspection | [
"Tongkun Liu",
"Bing Li",
"Xiao Du",
"Bingke Jiang",
"Xiao Jin",
"Liuyi Jin",
"Zhuo Zhao"
] | Industrial visual inspection aims at detecting surface defects in products
during the manufacturing process. Although existing anomaly detection models
have shown great performance on many public benchmarks, their limited
adjustability and ability to detect logical anomalies hinder their broader use
in real-world settings. To this end, in this paper, we propose a novel
component-aware anomaly detection framework (ComAD) which can simultaneously
achieve adjustable and logical anomaly detection for industrial scenarios.
Specifically, we propose to segment images into multiple components based on a
lightweight and nearly training-free unsupervised semantic segmentation model.
Then, we design an interpretable logical anomaly detection model through
modeling the metrological features of each component and their relationships.
Despite its simplicity, our framework achieves state-of-the-art performance on
image-level logical anomaly detection. Meanwhile, segmenting a product image
into multiple components provides a novel perspective for industrial visual
inspection, demonstrating great potential in model customization, noise
resistance, and anomaly classification. The code will be available at
https://github.com/liutongkun/ComAD. | [
"cs.CV"
] | false |
2305.08522 | 2023-05-15T10:30:38Z | Cross-Modality Time-Variant Relation Learning for Generating Dynamic
Scene Graphs | [
"Jingyi Wang",
"Jinfa Huang",
"Can Zhang",
"Zhidong Deng"
] | Dynamic scene graphs generated from video clips could help enhance the
semantic visual understanding in a wide range of challenging tasks such as
environmental perception, autonomous navigation, and task planning of
self-driving vehicles and mobile robots. In the process of temporal and spatial
modeling during dynamic scene graph generation, it is particularly intractable
to learn time-variant relations in dynamic scene graphs among frames. In this
paper, we propose a Time-variant Relation-aware TRansformer (TR$^2$), which
aims to model the temporal change of relations in dynamic scene graphs.
Explicitly, we leverage the difference of text embeddings of prompted sentences
about relation labels as the supervision signal for relations. In this way,
cross-modality feature guidance is realized for the learning of time-variant
relations. Implicitly, we design a relation feature fusion module with a
transformer and an additional message token that describes the difference
between adjacent frames. Extensive experiments on the Action Genome dataset
prove that our TR$^2$ can effectively model the time-variant relations. TR$^2$
significantly outperforms previous state-of-the-art methods under two different
settings by 2.1% and 2.6% respectively. | [
"cs.CV"
] | false |
2305.08540 | 2023-05-15T11:11:11Z | SRRM: Semantic Region Relation Model for Indoor Scene Recognition | [
"Chuanxin Song",
"Xin Ma"
] | Despite the remarkable success of convolutional neural networks in various
computer vision tasks, recognizing indoor scenes still presents a significant
challenge due to their complex composition. Consequently, effectively
leveraging semantic information in the scene has been a key issue in advancing
indoor scene recognition. Unfortunately, the accuracy of semantic segmentation
has limited the effectiveness of existing approaches for leveraging semantic
information. As a result, many of these approaches remain at the stage of
auxiliary labeling or co-occurrence statistics, with few exploring the
contextual relationships between the semantic elements directly within the
scene. In this paper, we propose the Semantic Region Relationship Model (SRRM),
which starts directly from the semantic information inside the scene.
Specifically, SRRM adopts an adaptive and efficient approach to mitigate the
negative impact of semantic ambiguity and then models the semantic region
relationship to perform scene recognition. Additionally, to more
comprehensively exploit the information contained in the scene, we combine the
proposed SRRM with the PlacesCNN module to create the Combined Semantic Region
Relation Model (CSRRM), and propose a novel information combining approach to
effectively explore the complementary contents between them. CSRRM
significantly outperforms the SOTA methods on the MIT Indoor 67, reduced
Places365 dataset, and SUN RGB-D without retraining. The code is available at:
https://github.com/ChuanxinSong/SRRM | [
"cs.CV"
] | false |
2305.08552 | 2023-05-15T11:26:32Z | Curvature-Aware Training for Coordinate Networks | [
"Hemanth Saratchandran",
"Shin-Fang Chng",
"Sameera Ramasinghe",
"Lachlan MacDonald",
"Simon Lucey"
] | Coordinate networks are widely used in computer vision due to their ability
to represent signals as compressed, continuous entities. However, training
these networks with first-order optimizers can be slow, hindering their use in
real-time applications. Recent works have opted for shallow voxel-based
representations to achieve faster training, but this sacrifices memory
efficiency. This work proposes a solution that leverages second-order
optimization methods to significantly reduce training times for coordinate
networks while maintaining their compressibility. Experiments demonstrate the
effectiveness of this approach on various signal modalities, such as audio,
images, videos, shape reconstruction, and neural radiance fields. | [
"cs.CV"
] | false |
2305.08590 | 2023-05-15T12:13:24Z | NIKI: Neural Inverse Kinematics with Invertible Neural Networks for 3D
Human Pose and Shape Estimation | [
"Jiefeng Li",
"Siyuan Bian",
"Qi Liu",
"Jiasheng Tang",
"Fan Wang",
"Cewu Lu"
] | With the progress of 3D human pose and shape estimation, state-of-the-art
methods can either be robust to occlusions or obtain pixel-aligned accuracy in
non-occlusion cases. However, they cannot obtain robustness and mesh-image
alignment at the same time. In this work, we present NIKI (Neural Inverse
Kinematics with Invertible Neural Network), which models bi-directional errors
to improve the robustness to occlusions and obtain pixel-aligned accuracy. NIKI
can learn from both the forward and inverse processes with invertible networks.
In the inverse process, the model separates the error from the plausible 3D
pose manifold for a robust 3D human pose estimation. In the forward process, we
enforce the zero-error boundary conditions to improve the sensitivity to
reliable joint positions for better mesh-image alignment. Furthermore, NIKI
emulates the analytical inverse kinematics algorithms with the twist-and-swing
decomposition for better interpretability. Experiments on standard and
occlusion-specific benchmarks demonstrate the effectiveness of NIKI, where we
exhibit robust and well-aligned results simultaneously. Code is available at
https://github.com/Jeff-sjtu/NIKI | [
"cs.CV"
] | false |
2305.08628 | 2023-05-15T13:21:44Z | Non-Separable Multi-Dimensional Network Flows for Visual Computing | [
"Viktoria Ehm",
"Daniel Cremers",
"Florian Bernard"
] | Flows in networks (or graphs) play a significant role in numerous computer
vision tasks. The scalar-valued edges in these graphs often lead to a loss of
information and thereby to limitations in terms of expressiveness. For example,
oftentimes high-dimensional data (e.g. feature descriptors) are mapped to a
single scalar value (e.g. the similarity between two feature descriptors). To
overcome this limitation, we propose a novel formalism for non-separable
multi-dimensional network flows. By doing so, we enable an automatic and
adaptive feature selection strategy - since the flow is defined on a
per-dimension basis, the maximizing flow automatically chooses the best
matching feature dimensions. As a proof of concept, we apply our formalism to
the multi-object tracking problem and demonstrate that our approach outperforms
scalar formulations on the MOT16 benchmark in terms of robustness to noise. | [
"cs.CV"
] | false |
2305.08721 | 2023-05-15T15:33:55Z | Learning More Discriminative Local Descriptors for Few-shot Learning | [
"Qijun Song",
"Siyun Zhou",
"Liwei Xu"
] | Few-shot learning for image classification comes up as a hot topic in
computer vision, which aims at fast learning from a limited number of labeled
images and generalize over the new tasks. In this paper, motivated by the idea
of Fisher Score, we propose a Discriminative Local Descriptors Attention (DLDA)
model that adaptively selects the representative local descriptors and does not
introduce any additional parameters, while most of the existing local
descriptors based methods utilize the neural networks that inevitably involve
the tedious parameter tuning. Moreover, we modify the traditional $k$-NN
classification model by adjusting the weights of the $k$ nearest neighbors
according to their distances from the query point. Experiments on four
benchmark datasets show that our method not only achieves higher accuracy
compared with the state-of-art approaches for few-shot learning, but also
possesses lower sensitivity to the choices of $k$. | [
"cs.CV"
] | false |
2305.08779 | 2023-05-15T16:38:55Z | TAA-GCN: A Temporally Aware Adaptive Graph Convolutional Network for Age
Estimation | [
"Matthew Korban",
"Peter Young",
"Scott T. Acton"
] | This paper proposes a novel age estimation algorithm, the Temporally-Aware
Adaptive Graph Convolutional Network (TAA-GCN). Using a new representation
based on graphs, the TAA-GCN utilizes skeletal, posture, clothing, and facial
information to enrich the feature set associated with various ages. Such a
novel graph representation has several advantages: First, reduced sensitivity
to facial expression and other appearance variances; Second, robustness to
partial occlusion and non-frontal-planar viewpoint, which is commonplace in
real-world applications such as video surveillance. The TAA-GCN employs two
novel components, (1) the Temporal Memory Module (TMM) to compute temporal
dependencies in age; (2) Adaptive Graph Convolutional Layer (AGCL) to refine
the graphs and accommodate the variance in appearance. The TAA-GCN outperforms
the state-of-the-art methods on four public benchmarks, UTKFace, MORPHII, CACD,
and FG-NET. Moreover, the TAA-GCN showed reliability in different camera
viewpoints and reduced quality images. | [
"cs.CV"
] | false |
2305.08808 | 2023-05-15T17:14:55Z | GeoMAE: Masked Geometric Target Prediction for Self-supervised Point
Cloud Pre-Training | [
"Xiaoyu Tian",
"Haoxi Ran",
"Yue Wang",
"Hang Zhao"
] | This paper tries to address a fundamental question in point cloud
self-supervised learning: what is a good signal we should leverage to learn
features from point clouds without annotations? To answer that, we introduce a
point cloud representation learning framework, based on geometric feature
reconstruction. In contrast to recent papers that directly adopt masked
autoencoder (MAE) and only predict original coordinates or occupancy from
masked point clouds, our method revisits differences between images and point
clouds and identifies three self-supervised learning objectives peculiar to
point clouds, namely centroid prediction, normal estimation, and curvature
prediction. Combined with occupancy prediction, these four objectives yield an
nontrivial self-supervised learning task and mutually facilitate models to
better reason fine-grained geometry of point clouds. Our pipeline is
conceptually simple and it consists of two major steps: first, it randomly
masks out groups of points, followed by a Transformer-based point cloud
encoder; second, a lightweight Transformer decoder predicts centroid, normal,
and curvature for points in each voxel. We transfer the pre-trained Transformer
encoder to a downstream peception model. On the nuScene Datset, our model
achieves 3.38 mAP improvment for object detection, 2.1 mIoU gain for
segmentation, and 1.7 AMOTA gain for multi-object tracking. We also conduct
experiments on the Waymo Open Dataset and achieve significant performance
improvements over baselines as well. | [
"cs.CV"
] | false |
2305.08810 | 2023-05-15T17:16:46Z | AutoRecon: Automated 3D Object Discovery and Reconstruction | [
"Yuang Wang",
"Xingyi He",
"Sida Peng",
"Haotong Lin",
"Hujun Bao",
"Xiaowei Zhou"
] | A fully automated object reconstruction pipeline is crucial for digital
content creation. While the area of 3D reconstruction has witnessed profound
developments, the removal of background to obtain a clean object model still
relies on different forms of manual labor, such as bounding box labeling, mask
annotations, and mesh manipulations. In this paper, we propose a novel
framework named AutoRecon for the automated discovery and reconstruction of an
object from multi-view images. We demonstrate that foreground objects can be
robustly located and segmented from SfM point clouds by leveraging
self-supervised 2D vision transformer features. Then, we reconstruct decomposed
neural scene representations with dense supervision provided by the decomposed
point clouds, resulting in accurate object reconstruction and segmentation.
Experiments on the DTU, BlendedMVS and CO3D-V2 datasets demonstrate the
effectiveness and robustness of AutoRecon. | [
"cs.CV"
] | true |
2305.08824 | 2023-05-15T17:33:29Z | Five A$^{+}$ Network: You Only Need 9K Parameters for Underwater Image
Enhancement | [
"Jingxia Jiang",
"Tian Ye",
"Jinbin Bai",
"Sixiang Chen",
"Wenhao Chai",
"Shi Jun",
"Yun Liu",
"Erkang Chen"
] | A lightweight underwater image enhancement network is of great significance
for resource-constrained platforms, but balancing model size, computational
efficiency, and enhancement performance has proven difficult for previous
approaches. In this work, we propose the Five A$^{+}$ Network (FA$^{+}$Net), a
highly efficient and lightweight real-time underwater image enhancement network
with only $\sim$ 9k parameters and $\sim$ 0.01s processing time. The
FA$^{+}$Net employs a two-stage enhancement structure. The strong prior stage
aims to decompose challenging underwater degradations into sub-problems, while
the fine-grained stage incorporates multi-branch color enhancement module and
pixel attention module to amplify the network's perception of details. To the
best of our knowledge, FA$^{+}$Net is the only network with the capability of
real-time enhancement of 1080P images. Thorough extensive experiments and
comprehensive visual comparison, we show that FA$^{+}$Net outperforms previous
approaches by obtaining state-of-the-art performance on multiple datasets while
significantly reducing both parameter count and computational complexity. The
code is open source at https://github.com/Owen718/FiveAPlus-Network. | [
"cs.CV"
] | false |
2305.08826 | 2023-05-15T17:34:49Z | Learning Better Contrastive View from Radiologist's Gaze | [
"Sheng Wang",
"Zixu Zhuang",
"Xi Ouyang",
"Lichi Zhang",
"Zheren Li",
"Chong Ma",
"Tianming Liu",
"Dinggang Shen",
"Qian Wang"
] | Recent self-supervised contrastive learning methods greatly benefit from the
Siamese structure that aims to minimizing distances between positive pairs.
These methods usually apply random data augmentation to input images, expecting
the augmented views of the same images to be similar and positively paired.
However, random augmentation may overlook image semantic information and
degrade the quality of augmented views in contrastive learning. This issue
becomes more challenging in medical images since the abnormalities related to
diseases can be tiny, and are easy to be corrupted (e.g., being cropped out) in
the current scheme of random augmentation. In this work, we first demonstrate
that, for widely-used X-ray images, the conventional augmentation prevalent in
contrastive pre-training can affect the performance of the downstream diagnosis
or classification tasks. Then, we propose a novel augmentation method, i.e.,
FocusContrast, to learn from radiologists' gaze in diagnosis and generate
contrastive views for medical images with guidance from radiologists' visual
attention. Specifically, we track the gaze movement of radiologists and model
their visual attention when reading to diagnose X-ray images. The learned model
can predict visual attention of the radiologists given a new input image, and
further guide the attention-aware augmentation that hardly neglects the
disease-related abnormalities. As a plug-and-play and framework-agnostic
module, FocusContrast consistently improves state-of-the-art contrastive
learning methods of SimCLR, MoCo, and BYOL by 4.0~7.0% in classification
accuracy on a knee X-ray dataset. | [
"cs.CV"
] | false |
2305.09072 | 2023-05-15T23:55:56Z | Skin Deep: Investigating Subjectivity in Skin Tone Annotations for
Computer Vision Benchmark Datasets | [
"Teanna Barrett",
"Quan Ze Chen",
"Amy X. Zhang"
] | To investigate the well-observed racial disparities in computer vision
systems that analyze images of humans, researchers have turned to skin tone as
more objective annotation than race metadata for fairness performance
evaluations. However, the current state of skin tone annotation procedures is
highly varied. For instance, researchers use a range of untested scales and
skin tone categories, have unclear annotation procedures, and provide
inadequate analyses of uncertainty. In addition, little attention is paid to
the positionality of the humans involved in the annotation process--both
designers and annotators alike--and the historical and sociological context of
skin tone in the United States. Our work is the first to investigate the skin
tone annotation process as a sociotechnical project. We surveyed recent skin
tone annotation procedures and conducted annotation experiments to examine how
subjective understandings of skin tone are embedded in skin tone annotation
procedures. Our systematic literature review revealed the uninterrogated
association between skin tone and race and the limited effort to analyze
annotator uncertainty in current procedures for skin tone annotation in
computer vision evaluation. Our experiments demonstrated that design decisions
in the annotation procedure such as the order in which the skin tone scale is
presented or additional context in the image (i.e., presence of a face)
significantly affected the resulting inter-annotator agreement and individual
uncertainty of skin tone annotations. We call for greater reflexivity in the
design, analysis, and documentation of procedures for evaluation using skin
tone. | [
"cs.CV"
] | false |
2305.08293 | 2023-05-15T01:31:32Z | Identity-Preserving Talking Face Generation with Landmark and Appearance
Priors | [
"Weizhi Zhong",
"Chaowei Fang",
"Yinqi Cai",
"Pengxu Wei",
"Gangming Zhao",
"Liang Lin",
"Guanbin Li"
] | Generating talking face videos from audio attracts lots of research interest.
A few person-specific methods can generate vivid videos but require the target
speaker's videos for training or fine-tuning. Existing person-generic methods
have difficulty in generating realistic and lip-synced videos while preserving
identity information. To tackle this problem, we propose a two-stage framework
consisting of audio-to-landmark generation and landmark-to-video rendering
procedures. First, we devise a novel Transformer-based landmark generator to
infer lip and jaw landmarks from the audio. Prior landmark characteristics of
the speaker's face are employed to make the generated landmarks coincide with
the facial outline of the speaker. Then, a video rendering model is built to
translate the generated landmarks into face images. During this stage, prior
appearance information is extracted from the lower-half occluded target face
and static reference images, which helps generate realistic and
identity-preserving visual content. For effectively exploring the prior
information of static reference images, we align static reference images with
the target face's pose and expression based on motion fields. Moreover,
auditory features are reused to guarantee that the generated face images are
well synchronized with the audio. Extensive experiments demonstrate that our
method can produce more realistic, lip-synced, and identity-preserving videos
than existing person-generic talking face generation methods. | [
"cs.CV",
"cs.MM"
] | false |
2305.08302 | 2023-05-15T02:05:56Z | t-RAIN: Robust generalization under weather-aliasing label shift attacks | [
"Aboli Marathe",
"Sanjana Prabhu"
] | In the classical supervised learning settings, classifiers are fit with the
assumption of balanced label distributions and produce remarkable results on
the same. In the real world, however, these assumptions often bend and in turn
adversely impact model performance. Identifying bad learners in skewed target
distributions is even more challenging. Thus achieving model robustness under
these "label shift" settings is an important task in autonomous perception. In
this paper, we analyze the impact of label shift on the task of multi-weather
classification for autonomous vehicles. We use this information as a prior to
better assess pedestrian detection in adverse weather. We model the
classification performance as an indicator of robustness under 4 label shift
scenarios and study the behavior of multiple classes of models. We propose
t-RAIN a similarity mapping technique for synthetic data augmentation using
large scale generative models and evaluate the performance on DAWN dataset.
This mapping boosts model test accuracy by 2.1, 4.4, 1.9, 2.7 % in no-shift,
fog, snow, dust shifts respectively. We present state-of-the-art pedestrian
detection results on real and synthetic weather domains with best performing
82.69 AP (snow) and 62.31 AP (fog) respectively. | [
"cs.CV",
"cs.AI"
] | false |
2305.08318 | 2023-05-15T03:08:10Z | CMSG Cross-Media Semantic-Graph Feature Matching Algorithm for
Autonomous Vehicle Relocalization | [
"Shuhang Tan",
"Hengyu Liu",
"Zhiling Wang"
] | Relocalization is the basis of map-based localization algorithms. Camera and
LiDAR map-based methods are pervasive since their robustness under different
scenarios. Generally, mapping and localization using the same sensor have
better accuracy since matching features between the same type of data is
easier. However, due to the camera's lack of 3D information and the high cost
of LiDAR, cross-media methods are developing, which combined live image data
and Lidar map. Although matching features between different media is
challenging, we believe cross-media is the tendency for AV relocalization since
its low cost and accuracy can be comparable to the same-sensor-based methods.
In this paper, we propose CMSG, a novel cross-media algorithm for AV
relocalization tasks. Semantic features are utilized for better interpretation
the correlation between point clouds and image features. What's more,
abstracted semantic graph nodes are introduced, and a graph network
architecture is integrated to better extract the similarity of semantic
features. Validation experiments are conducted on the KITTI odometry dataset.
Our results show that CMSG can have comparable or even better accuracy compared
to current single-sensor-based methods at a speed of 25 FPS on NVIDIA 1080 Ti
GPU. | [
"cs.CV",
"cs.AI"
] | false |
2305.08325 | 2023-05-15T03:24:36Z | Screentone-Aware Manga Super-Resolution Using DeepLearning | [
"Chih-Yuan Yao",
"Husan-Ting Chou",
"Yu-Sheng Lin",
"Kuo-wei Chen"
] | Manga, as a widely beloved form of entertainment around the world, have
shifted from paper to electronic screens with the proliferation of handheld
devices. However, as the demand for image quality increases with screen
development, high-quality images can hinder transmission and affect the viewing
experience. Traditional vectorization methods require a significant amount of
manual parameter adjustment to process screentone. Using deep learning, lines
and screentone can be automatically extracted and image resolution can be
enhanced. Super-resolution can convert low-resolution images to high-resolution
images while maintaining low transmission rates and providing high-quality
results. However, traditional Super Resolution methods for improving manga
resolution do not consider the meaning of screentone density, resulting in
changes to screentone density and loss of meaning. In this paper, we aims to
address this issue by first classifying the regions and lines of different
screentone in the manga using deep learning algorithm, then using corresponding
super-resolution models for quality enhancement based on the different
classifications of each block, and finally combining them to obtain images that
maintain the meaning of screentone and lines in the manga while improving image
resolution. | [
"cs.CV",
"eess.IV"
] | false |
2305.08389 | 2023-05-15T07:12:19Z | Edit As You Wish: Video Description Editing with Multi-grained Commands | [
"Linli Yao",
"Yuanmeng Zhang",
"Ziheng Wang",
"Xinglin Hou",
"Tiezheng Ge",
"Yuning Jiang",
"Qin Jin"
] | Automatically narrating a video with natural language can assist people in
grasping and managing massive videos on the Internet. From the perspective of
video uploaders, they may have varied preferences for writing the desired video
description to attract more potential followers, e.g. catching customers'
attention for product videos. The Controllable Video Captioning task is
therefore proposed to generate a description conditioned on the user demand and
video content. However, existing works suffer from two shortcomings: 1) the
control signal is fixed and can only express single-grained control; 2) the
video description can not be further edited to meet dynamic user demands. In
this paper, we propose a novel Video Description Editing (VDEdit) task to
automatically revise an existing video description guided by flexible user
requests. Inspired by human writing-revision habits, we design the user command
as a {operation, position, attribute} triplet to cover multi-grained use
requirements, which can express coarse-grained control (e.g. expand the
description) as well as fine-grained control (e.g. add specified details in
specified position) in a unified format. To facilitate the VDEdit task, we
first automatically construct a large-scale benchmark dataset namely VATEX-EDIT
in the open domain describing diverse human activities. Considering the
real-life application scenario, we further manually collect an e-commerce
benchmark dataset called EMMAD-EDIT. We propose a unified framework to convert
the {operation, position, attribute} triplet into a textual control sequence to
handle multi-grained editing commands. For VDEdit evaluation, we adopt
comprehensive metrics to measure three aspects of model performance, including
caption quality, caption-command consistency, and caption-video alignment. | [
"cs.CV",
"cs.MM"
] | false |
2305.08408 | 2023-05-15T07:44:10Z | SB-VQA: A Stack-Based Video Quality Assessment Framework for Video
Enhancement | [
"Ding-Jiun Huang",
"Yu-Ting Kao",
"Tieh-Hung Chuang",
"Ya-Chun Tsai",
"Jing-Kai Lou",
"Shuen-Huei Guan"
] | In recent years, several video quality assessment (VQA) methods have been
developed, achieving high performance. However, these methods were not
specifically trained for enhanced videos, which limits their ability to predict
video quality accurately based on human subjective perception. To address this
issue, we propose a stack-based framework for VQA that outperforms existing
state-of-the-art methods on VDPVE, a dataset consisting of enhanced videos. In
addition to proposing the VQA framework for enhanced videos, we also
investigate its application on professionally generated content (PGC). To
address copyright issues with premium content, we create the PGCVQ dataset,
which consists of videos from YouTube. We evaluate our proposed approach and
state-of-the-art methods on PGCVQ, and provide new insights on the results. Our
experiments demonstrate that existing VQA algorithms can be applied to PGC
videos, and we find that VQA performance for PGC videos can be improved by
considering the plot of a play, which highlights the importance of video
semantic understanding. | [
"cs.CV",
"eess.IV"
] | false |
2305.08462 | 2023-05-15T09:04:46Z | Not All Pixels Are Equal: Learning Pixel Hardness for Semantic
Segmentation | [
"Xin Xiao",
"Daiguo Zhou",
"Jiagao Hu",
"Yi Hu",
"Yongchao Xu"
] | Semantic segmentation has recently witnessed great progress. Despite the
impressive overall results, the segmentation performance in some hard areas
(e.g., small objects or thin parts) is still not promising. A straightforward
solution is hard sample mining, which is widely used in object detection. Yet,
most existing hard pixel mining strategies for semantic segmentation often rely
on pixel's loss value, which tends to decrease during training. Intuitively,
the pixel hardness for segmentation mainly depends on image structure and is
expected to be stable. In this paper, we propose to learn pixel hardness for
semantic segmentation, leveraging hardness information contained in global and
historical loss values. More precisely, we add a gradient-independent branch
for learning a hardness level (HL) map by maximizing hardness-weighted
segmentation loss, which is minimized for the segmentation head. This
encourages large hardness values in difficult areas, leading to appropriate and
stable HL map. Despite its simplicity, the proposed method can be applied to
most segmentation methods with no and marginal extra cost during inference and
training, respectively. Without bells and whistles, the proposed method
achieves consistent/significant improvement (1.37% mIoU on average) over most
popular semantic segmentation methods on Cityscapes dataset, and demonstrates
good generalization ability across domains. The source codes are available at
https://github.com/Menoly-xin/Hardness-Level-Learning . | [
"cs.CV",
"cs.AI"
] | false |
2305.08551 | 2023-05-15T11:23:18Z | Enhancing Performance of Vision Transformers on Small Datasets through
Local Inductive Bias Incorporation | [
"Ibrahim Batuhan Akkaya",
"Senthilkumar S. Kathiresan",
"Elahe Arani",
"Bahram Zonooz"
] | Vision transformers (ViTs) achieve remarkable performance on large datasets,
but tend to perform worse than convolutional neural networks (CNNs) when
trained from scratch on smaller datasets, possibly due to a lack of local
inductive bias in the architecture. Recent studies have therefore added
locality to the architecture and demonstrated that it can help ViTs achieve
performance comparable to CNNs in the small-size dataset regime. Existing
methods, however, are architecture-specific or have higher computational and
memory costs. Thus, we propose a module called Local InFormation Enhancer
(LIFE) that extracts patch-level local information and incorporates it into the
embeddings used in the self-attention block of ViTs. Our proposed module is
memory and computation efficient, as well as flexible enough to process
auxiliary tokens such as the classification and distillation tokens. Empirical
results show that the addition of the LIFE module improves the performance of
ViTs on small image classification datasets. We further demonstrate how the
effect can be extended to downstream tasks, such as object detection and
semantic segmentation. In addition, we introduce a new visualization method,
Dense Attention Roll-Out, specifically designed for dense prediction tasks,
allowing the generation of class-specific attention maps utilizing the
attention maps of all tokens. | [
"cs.CV",
"cs.AI"
] | false |
2305.08585 | 2023-05-15T12:12:29Z | Toward Moiré-Free and Detail-Preserving Demosaicking | [
"Xuanchen Li",
"Yan Niu",
"Bo Zhao",
"Haoyuan Shi",
"Zitong An"
] | 3D convolutions are commonly employed by demosaicking neural models, in the
same way as solving other image restoration problems. Counter-intuitively, we
show that 3D convolutions implicitly impede the RGB color spectra from
exchanging complementary information, resulting in spectral-inconsistent
inference of the local spatial high frequency components. As a consequence,
shallow 3D convolution networks suffer the Moir\'e artifacts, but deep 3D
convolutions cause over-smoothness. We analyze the fundamental difference
between demosaicking and other problems that predict lost pixels between
available ones (e.g., super-resolution reconstruction), and present the
underlying reasons for the confliction between Moir\'e-free and
detail-preserving. From the new perspective, our work decouples the common
standard convolution procedure to spectral and spatial feature aggregations,
which allow strengthening global communication in the spectral dimension while
respecting local contrast in the spatial dimension. We apply our demosaicking
model to two tasks: Joint Demosaicking-Denoising and Independently
Demosaicking. In both applications, our model substantially alleviates
artifacts such as Moir\'e and over-smoothness at similar or lower computational
cost to currently top-performing models, as validated by diverse evaluations.
Source code will be released along with paper publication. | [
"cs.CV",
"eess.IV"
] | false |
2305.08660 | 2023-05-15T14:07:22Z | Towards Automated COVID-19 Presence and Severity Classification | [
"Dominik Müller",
"Niklas Schröter",
"Silvan Mertes",
"Fabio Hellmann",
"Miriam Elia",
"Wolfgang Reif",
"Bernhard Bauer",
"Elisabeth André",
"Frank Kramer"
] | COVID-19 presence classification and severity prediction via (3D) thorax
computed tomography scans have become important tasks in recent times.
Especially for capacity planning of intensive care units, predicting the future
severity of a COVID-19 patient is crucial. The presented approach follows
state-of-theart techniques to aid medical professionals in these situations. It
comprises an ensemble learning strategy via 5-fold cross-validation that
includes transfer learning and combines pre-trained 3D-versions of ResNet34 and
DenseNet121 for COVID19 classification and severity prediction respectively.
Further, domain-specific preprocessing was applied to optimize model
performance. In addition, medical information like the infection-lung-ratio,
patient age, and sex were included. The presented model achieves an AUC of
79.0% to predict COVID-19 severity, and 83.7% AUC to classify the presence of
an infection, which is comparable with other currently popular methods. This
approach is implemented using the AUCMEDI framework and relies on well-known
network architectures to ensure robustness and reproducibility. | [
"eess.IV",
"cs.CV"
] | false |
2305.08661 | 2023-05-15T14:09:09Z | Global and Local Mixture Consistency Cumulative Learning for Long-tailed
Visual Recognitions | [
"Fei Du",
"Peng Yang",
"Qi Jia",
"Fengtao Nan",
"Xiaoting Chen",
"Yun Yang"
] | In this paper, our goal is to design a simple learning paradigm for long-tail
visual recognition, which not only improves the robustness of the feature
extractor but also alleviates the bias of the classifier towards head classes
while reducing the training skills and overhead. We propose an efficient
one-stage training strategy for long-tailed visual recognition called Global
and Local Mixture Consistency cumulative learning (GLMC). Our core ideas are
twofold: (1) a global and local mixture consistency loss improves the
robustness of the feature extractor. Specifically, we generate two augmented
batches by the global MixUp and local CutMix from the same batch data,
respectively, and then use cosine similarity to minimize the difference. (2) A
cumulative head tail soft label reweighted loss mitigates the head class bias
problem. We use empirical class frequencies to reweight the mixed label of the
head-tail class for long-tailed data and then balance the conventional loss and
the rebalanced loss with a coefficient accumulated by epochs. Our approach
achieves state-of-the-art accuracy on CIFAR10-LT, CIFAR100-LT, and ImageNet-LT
datasets. Additional experiments on balanced ImageNet and CIFAR demonstrate
that GLMC can significantly improve the generalization of backbones. Code is
made publicly available at https://github.com/ynu-yangpeng/GLMC. | [
"cs.CV",
"cs.AI"
] | false |
2305.08694 | 2023-05-15T14:56:12Z | A Reproducible Extraction of Training Images from Diffusion Models | [
"Ryan Webster"
] | Recently, Carlini et al. demonstrated the widely used model Stable Diffusion
can regurgitate real training samples, which is troublesome from a copyright
perspective. In this work, we provide an efficient extraction attack on par
with the recent attack, with several order of magnitudes less network
evaluations. In the process, we expose a new phenomena, which we dub template
verbatims, wherein a diffusion model will regurgitate a training sample largely
in tact. Template verbatims are harder to detect as they require retrieval and
masking to correctly label. Furthermore, they are still generated by newer
systems, even those which de-duplicate their training set, and we give insight
into why they still appear during generation. We extract training images from
several state of the art systems, including Stable Diffusion 2.0, Deep Image
Floyd, and finally Midjourney v4. We release code to verify our extraction
attack, perform the attack, as well as all extracted prompts at
\url{https://github.com/ryanwebster90/onestep-extraction}. | [
"cs.CV",
"cs.AI"
] | false |
2305.08840 | 2023-05-15T17:55:04Z | Attacking Perceptual Similarity Metrics | [
"Abhijay Ghildyal",
"Feng Liu"
] | Perceptual similarity metrics have progressively become more correlated with
human judgments on perceptual similarity; however, despite recent advances, the
addition of an imperceptible distortion can still compromise these metrics. In
our study, we systematically examine the robustness of these metrics to
imperceptible adversarial perturbations. Following the two-alternative
forced-choice experimental design with two distorted images and one reference
image, we perturb the distorted image closer to the reference via an
adversarial attack until the metric flips its judgment. We first show that all
metrics in our study are susceptible to perturbations generated via common
adversarial attacks such as FGSM, PGD, and the One-pixel attack. Next, we
attack the widely adopted LPIPS metric using spatial-transformation-based
adversarial perturbations (stAdv) in a white-box setting to craft adversarial
examples that can effectively transfer to other similarity metrics in a
black-box setting. We also combine the spatial attack stAdv with PGD
($\ell_\infty$-bounded) attack to increase transferability and use these
adversarial examples to benchmark the robustness of both traditional and
recently developed metrics. Our benchmark provides a good starting point for
discussion and further research on the robustness of metrics to imperceptible
adversarial perturbations. | [
"cs.CV",
"eess.IV"
] | false |
2305.08962 | 2023-05-15T19:03:45Z | Event Camera-based Visual Odometry for Dynamic Motion Tracking of a
Legged Robot Using Adaptive Time Surface | [
"Shifan Zhu",
"Zhipeng Tang",
"Michael Yang",
"Erik Learned-Miller",
"Donghyun Kim"
] | Our paper proposes a direct sparse visual odometry method that combines event
and RGB-D data to estimate the pose of agile-legged robots during dynamic
locomotion and acrobatic behaviors. Event cameras offer high temporal
resolution and dynamic range, which can eliminate the issue of blurred RGB
images during fast movements. This unique strength holds a potential for
accurate pose estimation of agile-legged robots, which has been a challenging
problem to tackle. Our framework leverages the benefits of both RGB-D and event
cameras to achieve robust and accurate pose estimation, even during dynamic
maneuvers such as jumping and landing a quadruped robot, the Mini-Cheetah. Our
major contributions are threefold: Firstly, we introduce an adaptive time
surface (ATS) method that addresses the whiteout and blackout issue in
conventional time surfaces by formulating pixel-wise decay rates based on scene
complexity and motion speed. Secondly, we develop an effective pixel selection
method that directly samples from event data and applies sample filtering
through ATS, enabling us to pick pixels on distinct features. Lastly, we
propose a nonlinear pose optimization formula that simultaneously performs
3D-2D alignment on both RGB-based and event-based maps and images, allowing the
algorithm to fully exploit the benefits of both data streams. We extensively
evaluate the performance of our framework on both public datasets and our own
quadruped robot dataset, demonstrating its effectiveness in accurately
estimating the pose of agile robots during dynamic movements. | [
"cs.RO",
"cs.CV"
] | false |
2305.08995 | 2023-05-15T20:24:38Z | Denoising Diffusion Models for Plug-and-Play Image Restoration | [
"Yuanzhi Zhu",
"Kai Zhang",
"Jingyun Liang",
"Jiezhang Cao",
"Bihan Wen",
"Radu Timofte",
"Luc Van Gool"
] | Plug-and-play Image Restoration (IR) has been widely recognized as a flexible
and interpretable method for solving various inverse problems by utilizing any
off-the-shelf denoiser as the implicit image prior. However, most existing
methods focus on discriminative Gaussian denoisers. Although diffusion models
have shown impressive performance for high-quality image synthesis, their
potential to serve as a generative denoiser prior to the plug-and-play IR
methods remains to be further explored. While several other attempts have been
made to adopt diffusion models for image restoration, they either fail to
achieve satisfactory results or typically require an unacceptable number of
Neural Function Evaluations (NFEs) during inference. This paper proposes
DiffPIR, which integrates the traditional plug-and-play method into the
diffusion sampling framework. Compared to plug-and-play IR methods that rely on
discriminative Gaussian denoisers, DiffPIR is expected to inherit the
generative ability of diffusion models. Experimental results on three
representative IR tasks, including super-resolution, image deblurring, and
inpainting, demonstrate that DiffPIR achieves state-of-the-art performance on
both the FFHQ and ImageNet datasets in terms of reconstruction faithfulness and
perceptual quality with no more than 100 NFEs. The source code is available at
{\url{https://github.com/yuanzhi-zhu/DiffPIR}} | [
"cs.CV",
"eess.IV"
] | false |
2305.11898 | 2023-05-15T09:05:32Z | Neural information coding for efficient spike-based image denoising | [
"Andrea Castagnetti",
"Alain Pegatoquet",
"Benoît Miramond"
] | In recent years, Deep Convolutional Neural Networks (DCNNs) have outreached
the performance of classical algorithms for image restoration tasks. However
most of these methods are not suited for computational efficiency and are
therefore too expensive to be executed on embedded and mobile devices. In this
work we investigate Spiking Neural Networks (SNNs) for Gaussian denoising, with
the goal of approaching the performance of conventional DCNN while reducing the
computational load. We propose a formal analysis of the information conversion
processing carried out by the Leaky Integrate and Fire (LIF) neurons and we
compare its performance with the classical rate-coding mechanism. The neural
coding schemes are then evaluated through experiments in terms of denoising
performance and computation efficiency for a state-of-the-art deep
convolutional neural network. Our results show that SNNs with LIF neurons can
provide competitive denoising performance but at a reduced computational cost. | [
"cs.CV",
"cs.AI"
] | false |
2305.15420 | 2023-05-15T20:08:43Z | A Hybrid Semantic-Geometric Approach for Clutter-Resistant Floorplan
Generation from Building Point Clouds | [
"Seongyong Kim",
"Yosuke Yajima",
"Jisoo Park",
"Jingdao Chen",
"Yong K. Cho"
] | Building Information Modeling (BIM) technology is a key component of modern
construction engineering and project management workflows. As-is BIM models
that represent the spatial reality of a project site can offer crucial
information to stakeholders for construction progress monitoring, error
checking, and building maintenance purposes. Geometric methods for
automatically converting raw scan data into BIM models (Scan-to-BIM) often fail
to make use of higher-level semantic information in the data. Whereas, semantic
segmentation methods only output labels at the point level without creating
object level models that is necessary for BIM. To address these issues, this
research proposes a hybrid semantic-geometric approach for clutter-resistant
floorplan generation from laser-scanned building point clouds. The input point
clouds are first pre-processed by normalizing the coordinate system and
removing outliers. Then, a semantic segmentation network based on PointNet++ is
used to label each point as ceiling, floor, wall, door, stair, and clutter. The
clutter points are removed whereas the wall, door, and stair points are used
for 2D floorplan generation. A region-growing segmentation algorithm paired
with geometric reasoning rules is applied to group the points together into
individual building elements. Finally, a 2-fold Random Sample Consensus
(RANSAC) algorithm is applied to parameterize the building elements into 2D
lines which are used to create the output floorplan. The proposed method is
evaluated using the metrics of precision, recall, Intersection-over-Union
(IOU), Betti error, and warping error. | [
"cs.CV",
"cs.AI"
] | false |
2306.05182 | 2023-05-15T18:38:25Z | Interactive Fashion Content Generation Using LLMs and Latent Diffusion
Models | [
"Krishna Sri Ipsit Mantri",
"Nevasini Sasikumar"
] | Fashionable image generation aims to synthesize images of diverse fashion
prevalent around the globe, helping fashion designers in real-time
visualization by giving them a basic customized structure of how a specific
design preference would look in real life and what further improvements can be
made for enhanced customer satisfaction. Moreover, users can alone interact and
generate fashionable images by just giving a few simple prompts. Recently,
diffusion models have gained popularity as generative models owing to their
flexibility and generation of realistic images from Gaussian noise. Latent
diffusion models are a type of generative model that use diffusion processes to
model the generation of complex data, such as images, audio, or text. They are
called "latent" because they learn a hidden representation, or latent variable,
of the data that captures its underlying structure. We propose a method
exploiting the equivalence between diffusion models and energy-based models
(EBMs) and suggesting ways to compose multiple probability distributions. We
describe a pipeline on how our method can be used specifically for new
fashionable outfit generation and virtual try-on using LLM-guided text-to-image
generation. Our results indicate that using an LLM to refine the prompts to the
latent diffusion model assists in generating globally creative and culturally
diversified fashion styles and reducing bias. | [
"cs.CV",
"cs.LG"
] | false |
2306.06195 | 2023-05-15T20:59:44Z | Risk stratification of malignant melanoma using neural networks | [
"Julian Burghoff",
"Leonhard Ackermann",
"Younes Salahdine",
"Veronika Bram",
"Katharina Wunderlich",
"Julius Balkenhol",
"Thomas Dirschka",
"Hanno Gottschalk"
] | In order to improve the detection and classification of malignant melanoma,
this paper describes an image-based method that can achieve AUROC values of up
to 0.78 without additional clinical information. Furthermore, the importance of
the domain gap between two different image sources is considered, as it is
important to create usability independent of hardware components such as the
high-resolution scanner used. Since for the application of machine learning
methods, alterations of scanner-specific properties such as brightness,
contrast or sharpness can have strong (negative) effects on the quality of the
prediction methods, two ways to overcome this domain gap are discussed in this
paper. | [
"eess.IV",
"cs.CV"
] | false |
2305.08413 | 2023-05-15T07:47:24Z | Artificial intelligence to advance Earth observation: a perspective | [
"Devis Tuia",
"Konrad Schindler",
"Begüm Demir",
"Gustau Camps-Valls",
"Xiao Xiang Zhu",
"Mrinalini Kochupillai",
"Sašo Džeroski",
"Jan N. van Rijn",
"Holger H. Hoos",
"Fabio Del Frate",
"Mihai Datcu",
"Jorge-Arnulfo Quiané-Ruiz",
"Volker Markl",
"Bertrand Le Saux",
"Rochelle Schneider"
] | Earth observation (EO) is a prime instrument for monitoring land and ocean
processes, studying the dynamics at work, and taking the pulse of our planet.
This article gives a bird's eye view of the essential scientific tools and
approaches informing and supporting the transition from raw EO data to usable
EO-based information. The promises, as well as the current challenges of these
developments, are highlighted under dedicated sections. Specifically, we cover
the impact of (i) Computer vision; (ii) Machine learning; (iii) Advanced
processing and computing; (iv) Knowledge-based AI; (v) Explainable AI and
causal inference; (vi) Physics-aware models; (vii) User-centric approaches; and
(viii) the much-needed discussion of ethical and societal issues related to the
massive use of ML technologies in EO. | [
"cs.CV",
"eess.IV",
"stat.AP"
] | false |
2305.08938 | 2023-05-15T18:19:29Z | DopUS-Net: Quality-Aware Robotic Ultrasound Imaging based on Doppler
Signal | [
"Zhongliang Jiang",
"Felix Duelmer",
"Nassir Navab"
] | Medical ultrasound (US) is widely used to evaluate and stage vascular
diseases, in particular for the preliminary screening program, due to the
advantage of being radiation-free. However, automatic segmentation of small
tubular structures (e.g., the ulnar artery) from cross-sectional US images is
still challenging. To address this challenge, this paper proposes the DopUS-Net
and a vessel re-identification module that leverage the Doppler effect to
enhance the final segmentation result. Firstly, the DopUS-Net combines the
Doppler images with B-mode images to increase the segmentation accuracy and
robustness of small blood vessels. It incorporates two encoders to exploit the
maximum potential of the Doppler signal and recurrent neural network modules to
preserve sequential information. Input to the first encoder is a two-channel
duplex image representing the combination of the grey-scale Doppler and B-mode
images to ensure anatomical spatial correctness. The second encoder operates on
the pure Doppler images to provide a region proposal. Secondly, benefiting from
the Doppler signal, this work first introduces an online artery
re-identification module to qualitatively evaluate the real-time segmentation
results and automatically optimize the probe pose for enhanced Doppler images.
This quality-aware module enables the closed-loop control of robotic screening
to further improve the confidence and robustness of image segmentation. The
experimental results demonstrate that the proposed approach with the
re-identification process can significantly improve the accuracy and robustness
of the segmentation results (dice score: from 0:54 to 0:86; intersection over
union: from 0:47 to 0:78). | [
"cs.CV",
"cs.RO",
"eess.IV"
] | false |
2305.09031 | 2023-05-15T21:35:23Z | AI in the Loop -- Functionalizing Fold Performance Disagreement to
Monitor Automated Medical Image Segmentation Pipelines | [
"Harrison C. Gottlich",
"Panagiotis Korfiatis",
"Adriana V. Gregory",
"Timothy L. Kline"
] | Methods for automatically flag poor performing-predictions are essential for
safely implementing machine learning workflows into clinical practice and for
identifying difficult cases during model training. We present a readily
adoptable method using sub-models trained on different dataset folds, where
their disagreement serves as a surrogate for model confidence. Thresholds
informed by human interobserver values were used to determine whether a final
ensemble model prediction would require manual review. In two different
datasets (abdominal CT and MR predicting kidney tumors), our framework
effectively identified low performing automated segmentations. Flagging images
with a minimum Interfold test Dice score below human interobserver variability
maximized the number of flagged images while ensuring maximum ensemble test
Dice. When our internally trained model was applied to an external publicly
available dataset (KiTS21), flagged images included smaller tumors than those
observed in our internally trained dataset, demonstrating the methods
robustness to flagging poor performing out-of-distribution input data.
Comparing interfold sub-model disagreement against human interobserver values
is an efficient way to approximate a model's epistemic uncertainty - its lack
of knowledge due to insufficient relevant training data - a key functionality
for adopting these applications in clinical practice. | [
"eess.IV",
"cs.CV",
"cs.LG"
] | false |
2305.08291 | 2023-05-15T01:18:23Z | Large Language Model Guided Tree-of-Thought | [
"Jieyi Long"
] | In this paper, we introduce the Tree-of-Thought (ToT) framework, a novel
approach aimed at improving the problem-solving capabilities of auto-regressive
large language models (LLMs). The ToT technique is inspired by the human mind's
approach for solving complex reasoning tasks through trial and error. In this
process, the human mind explores the solution space through a tree-like thought
process, allowing for backtracking when necessary. To implement ToT as a
software system, we augment an LLM with additional modules including a prompter
agent, a checker module, a memory module, and a ToT controller. In order to
solve a given problem, these modules engage in a multi-round conversation with
the LLM. The memory module records the conversation and state history of the
problem solving process, which allows the system to backtrack to the previous
steps of the thought-process and explore other directions from there. To verify
the effectiveness of the proposed technique, we implemented a ToT-based solver
for the Sudoku Puzzle. Experimental results show that the ToT framework can
significantly increase the success rate of Sudoku puzzle solving. Our
implementation of the ToT-based Sudoku solver is available on GitHub:
\url{https://github.com/jieyilong/tree-of-thought-puzzle-solver}. | [
"cs.AI",
"cs.CL",
"cs.CV",
"cs.LG",
"cs.NE"
] | false |
2305.08300 | 2023-05-15T02:01:20Z | "Nothing Abnormal": Disambiguating Medical Reports via Contrastive
Knowledge Infusion | [
"Zexue He",
"An Yan",
"Amilcare Gentili",
"Julian McAuley",
"Chun-Nan Hsu"
] | Sharing medical reports is essential for patient-centered care. A recent line
of work has focused on automatically generating reports with NLP methods.
However, different audiences have different purposes when writing/reading
medical reports -- for example, healthcare professionals care more about
pathology, whereas patients are more concerned with the diagnosis ("Is there
any abnormality?"). The expectation gap results in a common situation where
patients find their medical reports to be ambiguous and therefore unsure about
the next steps. In this work, we explore the audience expectation gap in
healthcare and summarize common ambiguities that lead patients to be confused
about their diagnosis into three categories: medical jargon, contradictory
findings, and misleading grammatical errors. Based on our analysis, we define a
disambiguation rewriting task to regenerate an input to be unambiguous while
preserving information about the original content. We further propose a
rewriting algorithm based on contrastive pretraining and perturbation-based
rewriting. In addition, we create two datasets, OpenI-Annotated based on chest
reports and VA-Annotated based on general medical reports, with available
binary labels for ambiguity and abnormality presence annotated by radiology
specialists. Experimental results on these datasets show that our proposed
algorithm effectively rewrites input sentences in a less ambiguous way with
high content fidelity. Our code and annotated data are released to facilitate
future research. | [
"cs.CL"
] | false |
2305.08428 | 2023-05-15T08:13:00Z | Legal Extractive Summarization of U.S. Court Opinions | [
"Emmanuel Bauer",
"Dominik Stammbach",
"Nianlong Gu",
"Elliott Ash"
] | This paper tackles the task of legal extractive summarization using a dataset
of 430K U.S. court opinions with key passages annotated. According to automated
summary quality metrics, the reinforcement-learning-based MemSum model is best
and even out-performs transformer-based models. In turn, expert human
evaluation shows that MemSum summaries effectively capture the key points of
lengthy court opinions. Motivated by these results, we open-source our models
to the general public. This represents progress towards democratizing law and
making U.S. court opinions more accessible to the general public. | [
"cs.CL"
] | false |
2305.08433 | 2023-05-15T08:21:32Z | EMBRACE: Evaluation and Modifications for Boosting RACE | [
"Mariia Zyrianova",
"Dmytro Kalpakchi",
"Johan Boye"
] | When training and evaluating machine reading comprehension models, it is very
important to work with high-quality datasets that are also representative of
real-world reading comprehension tasks. This requirement includes, for
instance, having questions that are based on texts of different genres and
require generating inferences or reflecting on the reading material.
In this article we turn our attention to RACE, a dataset of English texts and
corresponding multiple-choice questions (MCQs). Each MCQ consists of a question
and four alternatives (of which one is the correct answer). RACE was
constructed by Chinese teachers of English for human reading comprehension and
is widely used as training material for machine reading comprehension models.
By construction, RACE should satisfy the aforementioned quality requirements
and the purpose of this article is to check whether they are indeed satisfied.
We provide a detailed analysis of the test set of RACE for high-school
students (1045 texts and 3498 corresponding MCQs) including (1) an evaluation
of the difficulty of each MCQ and (2) annotations for the relevant pieces of
the texts (called "bases") that are used to justify the plausibility of each
alternative. A considerable number of MCQs appear not to fulfill basic
requirements for this type of reading comprehension tasks, so we additionally
identify the high-quality subset of the evaluated RACE corpus. We also
demonstrate that the distribution of the positions of the bases for the
alternatives is biased towards certain parts of texts, which is not necessarily
desirable when evaluating MCQ answering and generation models. | [
"cs.CL"
] | false |
2305.08487 | 2023-05-15T09:43:32Z | Taxi1500: A Multilingual Dataset for Text Classification in 1500
Languages | [
"Chunlan Ma",
"Ayyoob ImaniGooghari",
"Haotian Ye",
"Ehsaneddin Asgari",
"Hinrich Schütze"
] | While natural language processing tools have been developed extensively for
some of the world's languages, a significant portion of the world's over 7000
languages are still neglected. One reason for this is that evaluation datasets
do not yet cover a wide range of languages, including low-resource and
endangered ones. We aim to address this issue by creating a text classification
dataset encompassing a large number of languages, many of which currently have
little to no annotated data available. We leverage parallel translations of the
Bible to construct such a dataset by first developing applicable topics and
employing a crowdsourcing tool to collect annotated data. By annotating the
English side of the data and projecting the labels onto other languages through
aligned verses, we generate text classification datasets for more than 1500
languages. We extensively benchmark several existing multilingual language
models using our dataset. To facilitate the advancement of research in this
area, we will release our dataset and code. | [
"cs.CL"
] | false |
2305.08572 | 2023-05-15T12:01:09Z | Estimating the Causal Effects of Natural Logic Features in Neural NLI
Models | [
"Julia Rozanova",
"Marco Valentino",
"Andre Freitas"
] | Rigorous evaluation of the causal effects of semantic features on language
model predictions can be hard to achieve for natural language reasoning
problems. However, this is such a desirable form of analysis from both an
interpretability and model evaluation perspective, that it is valuable to zone
in on specific patterns of reasoning with enough structure and regularity to be
able to identify and quantify systematic reasoning failures in widely-used
models. In this vein, we pick a portion of the NLI task for which an explicit
causal diagram can be systematically constructed: in particular, the case where
across two sentences (the premise and hypothesis), two related words/terms
occur in a shared context. In this work, we apply causal effect estimation
strategies to measure the effect of context interventions (whose effect on the
entailment label is mediated by the semantic monotonicity characteristic) and
interventions on the inserted word-pair (whose effect on the entailment label
is mediated by the relation between these words.). Following related work on
causal analysis of NLP models in different settings, we adapt the methodology
for the NLI task to construct comparative model profiles in terms of robustness
to irrelevant changes and sensitivity to impactful changes. | [
"cs.CL"
] | false |
2305.08654 | 2023-05-15T13:58:21Z | Unsupervised Semantic Variation Prediction using the Distribution of
Sibling Embeddings | [
"Taichi Aida",
"Danushka Bollegala"
] | Languages are dynamic entities, where the meanings associated with words
constantly change with time. Detecting the semantic variation of words is an
important task for various NLP applications that must make time-sensitive
predictions. Existing work on semantic variation prediction have predominantly
focused on comparing some form of an averaged contextualised representation of
a target word computed from a given corpus. However, some of the previously
associated meanings of a target word can become obsolete over time (e.g.
meaning of gay as happy), while novel usages of existing words are observed
(e.g. meaning of cell as a mobile phone). We argue that mean representations
alone cannot accurately capture such semantic variations and propose a method
that uses the entire cohort of the contextualised embeddings of the target
word, which we refer to as the sibling distribution. Experimental results on
SemEval-2020 Task 1 benchmark dataset for semantic variation prediction show
that our method outperforms prior work that consider only the mean embeddings,
and is comparable to the current state-of-the-art. Moreover, a qualitative
analysis shows that our method detects important semantic changes in words that
are not captured by the existing methods. Source code is available at
https://github.com/a1da4/svp-gauss . | [
"cs.CL"
] | false |
2305.08655 | 2023-05-15T13:59:23Z | Unsupervised Sentence Representation Learning with Frequency-induced
Adversarial Tuning and Incomplete Sentence Filtering | [
"Bing Wang",
"Ximing Li",
"Zhiyao Yang",
"Yuanyuan Guan",
"Jiayin Li",
"Shengsheng Wang"
] | Pre-trained Language Model (PLM) is nowadays the mainstay of Unsupervised
Sentence Representation Learning (USRL). However, PLMs are sensitive to the
frequency information of words from their pre-training corpora, resulting in
anisotropic embedding space, where the embeddings of high-frequency words are
clustered but those of low-frequency words disperse sparsely. This anisotropic
phenomenon results in two problems of similarity bias and information bias,
lowering the quality of sentence embeddings. To solve the problems, we
fine-tune PLMs by leveraging the frequency information of words and propose a
novel USRL framework, namely Sentence Representation Learning with
Frequency-induced Adversarial tuning and Incomplete sentence filtering
(SLT-FAI). We calculate the word frequencies over the pre-training corpora of
PLMs and assign words thresholding frequency labels. With them, (1) we
incorporate a similarity discriminator used to distinguish the embeddings of
high-frequency and low-frequency words, and adversarially tune the PLM with it,
enabling to achieve uniformly frequency-invariant embedding space; and (2) we
propose a novel incomplete sentence detection task, where we incorporate an
information discriminator to distinguish the embeddings of original sentences
and incomplete sentences by randomly masking several low-frequency words,
enabling to emphasize the more informative low-frequency words. Our SLT-FAI is
a flexible and plug-and-play framework, and it can be integrated with existing
USRL techniques. We evaluate SLT-FAI with various backbones on benchmark
datasets. Empirical results indicate that SLT-FAI can be superior to the
existing USRL baselines. Our code is released in
\url{https://github.com/wangbing1416/SLT-FAI}. | [
"cs.CL"
] | false |
2305.08787 | 2023-05-15T16:46:47Z | Comparing Variation in Tokenizer Outputs Using a Series of Problematic
and Challenging Biomedical Sentences | [
"Christopher Meaney",
"Therese A Stukel",
"Peter C Austin",
"Michael Escobar"
] | Background & Objective: Biomedical text data are increasingly available for
research. Tokenization is an initial step in many biomedical text mining
pipelines. Tokenization is the process of parsing an input biomedical sentence
(represented as a digital character sequence) into a discrete set of word/token
symbols, which convey focused semantic/syntactic meaning. The objective of this
study is to explore variation in tokenizer outputs when applied across a series
of challenging biomedical sentences.
Method: Diaz [2015] introduce 24 challenging example biomedical sentences for
comparing tokenizer performance. In this study, we descriptively explore
variation in outputs of eight tokenizers applied to each example biomedical
sentence. The tokenizers compared in this study are the NLTK white space
tokenizer, the NLTK Penn Tree Bank tokenizer, Spacy and SciSpacy tokenizers,
Stanza/Stanza-Craft tokenizers, the UDPipe tokenizer, and R-tokenizers.
Results: For many examples, tokenizers performed similarly effectively;
however, for certain examples, there were meaningful variation in returned
outputs. The white space tokenizer often performed differently than other
tokenizers. We observed performance similarities for tokenizers implementing
rule-based systems (e.g. pattern matching and regular expressions) and
tokenizers implementing neural architectures for token classification.
Oftentimes, the challenging tokens resulting in the greatest variation in
outputs, are those words which convey substantive and focused
biomedical/clinical meaning (e.g. x-ray, IL-10, TCR/CD3, CD4+ CD8+, and
(Ca2+)-regulated).
Conclusion: When state-of-the-art, open-source tokenizers from Python and R
were applied to a series of challenging biomedical example sentences, we
observed subtle variation in the returned outputs. | [
"cs.CL"
] | false |
2305.08800 | 2023-05-15T17:05:45Z | Measuring Cross-Lingual Transferability of Multilingual Transformers on
Sentence Classification | [
"Zewen Chi",
"Heyan Huang",
"Xian-Ling Mao"
] | Recent studies have exhibited remarkable capabilities of pre-trained
multilingual Transformers, especially cross-lingual transferability. However,
current methods do not measure cross-lingual transferability well, hindering
the understanding of multilingual Transformers. In this paper, we propose IGap,
a cross-lingual transferability metric for multilingual Transformers on
sentence classification tasks. IGap takes training error into consideration,
and can also estimate transferability without end-task data. Experimental
results show that IGap outperforms baseline metrics for transferability
measuring and transfer direction ranking. Besides, we conduct extensive
systematic experiments where we compare transferability among various
multilingual Transformers, fine-tuning algorithms, and transfer directions.
More importantly, our results reveal three findings about cross-lingual
transfer, which helps us to better understand multilingual Transformers. | [
"cs.CL"
] | false |
2305.08804 | 2023-05-15T17:10:19Z | Exploring In-Context Learning Capabilities of Foundation Models for
Generating Knowledge Graphs from Text | [
"Hanieh Khorashadizadeh",
"Nandana Mihindukulasooriya",
"Sanju Tiwari",
"Jinghua Groppe",
"Sven Groppe"
] | Knowledge graphs can represent information about the real-world using
entities and their relations in a structured and semantically rich manner and
they enable a variety of downstream applications such as question-answering,
recommendation systems, semantic search, and advanced analytics. However, at
the moment, building a knowledge graph involves a lot of manual effort and thus
hinders their application in some situations and the automation of this process
might benefit especially for small organizations. Automatically generating
structured knowledge graphs from a large volume of natural language is still a
challenging task and the research on sub-tasks such as named entity extraction,
relation extraction, entity and relation linking, and knowledge graph
construction aims to improve the state of the art of automatic construction and
completion of knowledge graphs from text. The recent advancement of foundation
models with billions of parameters trained in a self-supervised manner with
large volumes of training data that can be adapted to a variety of downstream
tasks has helped to demonstrate high performance on a large range of Natural
Language Processing (NLP) tasks. In this context, one emerging paradigm is
in-context learning where a language model is used as it is with a prompt that
provides instructions and some examples to perform a task without changing the
parameters of the model using traditional approaches such as fine-tuning. This
way, no computing resources are needed for re-training/fine-tuning the models
and the engineering effort is minimal. Thus, it would be beneficial to utilize
such capabilities for generating knowledge graphs from text. | [
"cs.CL"
] | false |
2305.08853 | 2023-05-15T17:59:41Z | CQE: A Comprehensive Quantity Extractor | [
"Satya Almasian",
"Vivian Kazakova",
"Philip Göldner",
"Michael Gertz"
] | Quantities are essential in documents to describe factual information. They
are ubiquitous in application domains such as finance, business, medicine, and
science in general. Compared to other information extraction approaches,
interestingly only a few works exist that describe methods for a proper
extraction and representation of quantities in text. In this paper, we present
such a comprehensive quantity extraction framework from text data. It
efficiently detects combinations of values and units, the behavior of a
quantity (e.g., rising or falling), and the concept a quantity is associated
with. Our framework makes use of dependency parsing and a dictionary of units,
and it provides for a proper normalization and standardization of detected
quantities. Using a novel dataset for evaluation, we show that our open source
framework outperforms other systems and -- to the best of our knowledge -- is
the first to detect concepts associated with identified quantities. The code
and data underlying our framework are available at
https://github.com/vivkaz/CQE. | [
"cs.CL",
"I.7; I.7.1; I.7.5"
] | false |
2305.09022 | 2023-05-15T21:12:07Z | It Takes Two to Tango: Navigating Conceptualizations of NLP Tasks and
Measurements of Performance | [
"Arjun Subramonian",
"Xingdi Yuan",
"Hal Daumé III",
"Su Lin Blodgett"
] | Progress in NLP is increasingly measured through benchmarks; hence,
contextualizing progress requires understanding when and why practitioners may
disagree about the validity of benchmarks. We develop a taxonomy of
disagreement, drawing on tools from measurement modeling, and distinguish
between two types of disagreement: 1) how tasks are conceptualized and 2) how
measurements of model performance are operationalized. To provide evidence for
our taxonomy, we conduct a meta-analysis of relevant literature to understand
how NLP tasks are conceptualized, as well as a survey of practitioners about
their impressions of different factors that affect benchmark validity. Our
meta-analysis and survey across eight tasks, ranging from coreference
resolution to question answering, uncover that tasks are generally not clearly
and consistently conceptualized and benchmarks suffer from operationalization
disagreements. These findings support our proposed taxonomy of disagreement.
Finally, based on our taxonomy, we present a framework for constructing
benchmarks and documenting their limitations. | [
"cs.CL"
] | false |
2305.09067 | 2023-05-15T23:29:56Z | SGP-TOD: Building Task Bots Effortlessly via Schema-Guided LLM Prompting | [
"Xiaoying Zhang",
"Baolin Peng",
"Kun Li",
"Jingyan Zhou",
"Helen Meng"
] | Building end-to-end task bots and maintaining their integration with new
functionalities using minimal human efforts is a long-standing challenge in
dialog research. Recently large language models (LLMs) have demonstrated
exceptional proficiency in conversational engagement and adherence to
instructions across various downstream tasks. In this work, we introduce
SGP-TOD, Schema-Guided Prompting for building Task-Oriented Dialog systems
effortlessly based on LLMs. Utilizing the symbolic knowledge -- task schema, we
instruct fixed LLMs to generate appropriate responses on novel tasks,
circumventing the need for training data. Specifically, SGP-TOD comprises three
components: a LLM for engaging with users, a DST Prompter to aid the LLM with
dialog state tracking, which is then used to retrieve database items, and a
Policy Prompter to elicit proper responses adhering to the provided dialog
policy. Experimental results on Multiwoz, RADDLE and STAR datasets show that
our training-free strategy SGP-TOD, without any task-specific data, yields
state-of-the-art (SOTA) zero-shot performance, greatly surpasses the few-shot
approaches. In a domain-extension setting, SGP-TOD aptly adapts to new
functionalities by merely adding supplementary schema rules. We make our code
and data publicly available. | [
"cs.CL"
] | false |
2305.08347 | 2023-05-15T04:58:37Z | KEPR: Knowledge Enhancement and Plausibility Ranking for Generative
Commonsense Question Answering | [
"Zhifeng Li",
"Bowei Zou",
"Yifan Fan",
"Yu Hong"
] | Generative commonsense question answering (GenCQA) is a task of automatically
generating a list of answers given a question. The answer list is required to
cover all reasonable answers. This presents the considerable challenges of
producing diverse answers and ranking them properly. Incorporating a variety of
closely-related background knowledge into the encoding of questions enables the
generation of different answers. Meanwhile, learning to distinguish positive
answers from negative ones potentially enhances the probabilistic estimation of
plausibility, and accordingly, the plausibility-based ranking. Therefore, we
propose a Knowledge Enhancement and Plausibility Ranking (KEPR) approach
grounded on the Generate-Then-Rank pipeline architecture. Specifically, we
expand questions in terms of Wiktionary commonsense knowledge of keywords, and
reformulate them with normalized patterns. Dense passage retrieval is utilized
for capturing relevant knowledge, and different PLM-based (BART, GPT2 and T5)
networks are used for generating answers. On the other hand, we develop an
ELECTRA-based answer ranking model, where logistic regression is conducted
during training, with the aim of approximating different levels of plausibility
in a polar classification scenario. Extensive experiments on the benchmark
ProtoQA show that KEPR obtains substantial improvements, compared to the strong
baselines. Within the experimental models, the T5-based GenCQA with KEPR
obtains the best performance, which is up to 60.91% at the primary canonical
metric Inc@3. It outperforms the existing GenCQA models on the current
leaderboard of ProtoQA. | [
"cs.CL",
"cs.AI"
] | false |
2305.08414 | 2023-05-15T07:48:31Z | What's the Meaning of Superhuman Performance in Today's NLU? | [
"Simone Tedeschi",
"Johan Bos",
"Thierry Declerck",
"Jan Hajic",
"Daniel Hershcovich",
"Eduard H. Hovy",
"Alexander Koller",
"Simon Krek",
"Steven Schockaert",
"Rico Sennrich",
"Ekaterina Shutova",
"Roberto Navigli"
] | In the last five years, there has been a significant focus in Natural
Language Processing (NLP) on developing larger Pretrained Language Models
(PLMs) and introducing benchmarks such as SuperGLUE and SQuAD to measure their
abilities in language understanding, reasoning, and reading comprehension.
These PLMs have achieved impressive results on these benchmarks, even
surpassing human performance in some cases. This has led to claims of
superhuman capabilities and the provocative idea that certain tasks have been
solved. In this position paper, we take a critical look at these claims and ask
whether PLMs truly have superhuman abilities and what the current benchmarks
are really evaluating. We show that these benchmarks have serious limitations
affecting the comparison between humans and PLMs and provide recommendations
for fairer and more transparent benchmarks. | [
"cs.CL",
"cs.AI"
] | false |
2305.08495 | 2023-05-15T09:52:36Z | Similarity-weighted Construction of Contextualized Commonsense Knowledge
Graphs for Knowledge-intense Argumentation Tasks | [
"Moritz Plenz",
"Juri Opitz",
"Philipp Heinisch",
"Philipp Cimiano",
"Anette Frank"
] | Arguments often do not make explicit how a conclusion follows from its
premises. To compensate for this lack, we enrich arguments with structured
background knowledge to support knowledge-intense argumentation tasks. We
present a new unsupervised method for constructing Contextualized Commonsense
Knowledge Graphs (CCKGs) that selects contextually relevant knowledge from
large knowledge graphs (KGs) efficiently and at high quality. Our work goes
beyond context-insensitive knowledge extraction heuristics by computing
semantic similarity between KG triplets and textual arguments. Using these
triplet similarities as weights, we extract contextualized knowledge paths that
connect a conclusion to its premise, while maximizing similarity to the
argument. We combine multiple paths into a CCKG that we optionally prune to
reduce noise and raise precision. Intrinsic evaluation of the quality of our
graphs shows that our method is effective for (re)constructing human
explanation graphs. Manual evaluations in a large-scale knowledge selection
setup confirm high recall and precision of implicit CSK in the CCKGs. Finally,
we demonstrate the effectiveness of CCKGs in a knowledge-insensitive argument
quality rating task, outperforming strong baselines and rivaling a GPT-3 based
system. | [
"cs.CL",
"cs.DB"
] | false |
2305.08502 | 2023-05-15T10:02:47Z | MeeQA: Natural Questions in Meeting Transcripts | [
"Reut Apel",
"Tom Braude",
"Amir Kantor",
"Eyal Kolman"
] | We present MeeQA, a dataset for natural-language question answering over
meeting transcripts. It includes real questions asked during meetings by its
participants. The dataset contains 48K question-answer pairs, extracted from
422 meeting transcripts, spanning multiple domains. Questions in transcripts
pose a special challenge as they are not always clear, and considerable context
may be required in order to provide an answer. Further, many questions asked
during meetings are left unanswered. To improve baseline model performance on
this type of questions, we also propose a novel loss function, \emph{Flat
Hierarchical Loss}, designed to enhance performance over questions with no
answer in the text. Our experiments demonstrate the advantage of using our
approach over standard QA models. | [
"cs.CL",
"cs.LG"
] | false |
2305.08518 | 2023-05-15T10:28:36Z | Beqi: Revitalize the Senegalese Wolof Language with a Robust Spelling
Corrector | [
"Derguene Mbaye",
"Moussa Diallo"
] | The progress of Natural Language Processing (NLP), although fast in recent
years, is not at the same pace for all languages. African languages in
particular are still behind and lack automatic processing tools. Some of these
tools are very important for the development of these languages but also have
an important role in many NLP applications. This is particularly the case for
automatic spell checkers. Several approaches have been studied to address this
task and the one modeling spelling correction as a translation task from
misspelled (noisy) text to well-spelled (correct) text shows promising results.
However, this approach requires a parallel corpus of noisy data on the one hand
and correct data on the other hand, whereas Wolof is a low-resource language
and does not have such a corpus. In this paper, we present a way to address the
constraint related to the lack of data by generating synthetic data and we
present sequence-to-sequence models using Deep Learning for spelling correction
in Wolof. We evaluated these models in three different scenarios depending on
the subwording method applied to the data and showed that the latter had a
significant impact on the performance of the models, which opens the way for
future research in Wolof spelling correction. | [
"cs.CL",
"cs.AI"
] | false |
2305.08625 | 2023-05-15T13:20:14Z | Adam-Smith at SemEval-2023 Task 4: Discovering Human Values in Arguments
with Ensembles of Transformer-based Models | [
"Daniel Schroter",
"Daryna Dementieva",
"Georg Groh"
] | This paper presents the best-performing approach alias "Adam Smith" for the
SemEval-2023 Task 4: "Identification of Human Values behind Arguments". The
goal of the task was to create systems that automatically identify the values
within textual arguments. We train transformer-based models until they reach
their loss minimum or f1-score maximum. Ensembling the models by selecting one
global decision threshold that maximizes the f1-score leads to the
best-performing system in the competition. Ensembling based on stacking with
logistic regressions shows the best performance on an additional dataset
provided to evaluate the robustness ("Nahj al-Balagha"). Apart from outlining
the submitted system, we demonstrate that the use of the large ensemble model
is not necessary and that the system size can be significantly reduced. | [
"cs.CL",
"cs.AI"
] | false |
2305.08633 | 2023-05-15T13:26:50Z | Text2Gender: A Deep Learning Architecture for Analysis of Blogger's Age
and Gender | [
"Vishesh Thakur",
"Aneesh Tickoo"
] | Deep learning techniques have gained a lot of traction in the field of NLP
research. The aim of this paper is to predict the age and gender of an
individual by inspecting their written text. We propose a supervised BERT-based
classification technique in order to predict the age and gender of bloggers.
The dataset used contains 681284 rows of data, with the information of the
blogger's age, gender, and text of the blog written by them. We compare our
algorithm to previous works in the same domain and achieve a better accuracy
and F1 score. The accuracy reported for the prediction of age group was 84.2%,
while the accuracy for the prediction of gender was 86.32%. This study relies
on the raw capabilities of BERT to predict the classes of textual data
efficiently. This paper shows promising capability in predicting the
demographics of the author with high accuracy and can have wide applicability
across multiple domains. | [
"cs.CL",
"cs.AI"
] | false |
2305.08636 | 2023-05-15T13:28:59Z | AdamR at SemEval-2023 Task 10: Solving the Class Imbalance Problem in
Sexism Detection with Ensemble Learning | [
"Adam Rydelek",
"Daryna Dementieva",
"Georg Groh"
] | The Explainable Detection of Online Sexism task presents the problem of
explainable sexism detection through fine-grained categorisation of sexist
cases with three subtasks. Our team experimented with different ways to combat
class imbalance throughout the tasks using data augmentation and loss
alteration techniques. We tackled the challenge by utilising ensembles of
Transformer models trained on different datasets, which are tested to find the
balance between performance and interpretability. This solution ranked us in
the top 40\% of teams for each of the tracks. | [
"cs.CL",
"cs.AI"
] | false |
2305.08702 | 2023-05-15T15:05:44Z | Recyclable Tuning for Continual Pre-training | [
"Yujia Qin",
"Cheng Qian",
"Xu Han",
"Yankai Lin",
"Huadong Wang",
"Ruobing Xie",
"Zhiyuan Liu",
"Maosong Sun",
"Jie Zhou"
] | Continual pre-training is the paradigm where pre-trained language models
(PLMs) continually acquire fresh knowledge from growing data and gradually get
upgraded. Before an upgraded PLM is released, we may have tuned the original
PLM for various tasks and stored the adapted weights. However, when tuning the
upgraded PLM, these outdated adapted weights will typically be ignored and
discarded, causing a potential waste of resources. We bring this issue to the
forefront and contend that proper algorithms for recycling outdated adapted
weights should be developed. To this end, we formulate the task of recyclable
tuning for continual pre-training. In pilot studies, we find that after
continual pre-training, the upgraded PLM remains compatible with the outdated
adapted weights to some extent. Motivated by this finding, we analyze the
connection between continually pre-trained PLMs from two novel aspects, i.e.,
mode connectivity, and functional similarity. Based on the corresponding
findings, we propose both an initialization-based method and a
distillation-based method for our task. We demonstrate their feasibility in
improving the convergence and performance for tuning the upgraded PLM. We also
show that both methods can be combined to achieve better performance. The
source codes are publicly available at
https://github.com/thunlp/RecyclableTuning. | [
"cs.CL",
"cs.AI"
] | false |
2305.08818 | 2023-05-15T17:28:59Z | Sentence Level Curriculum Learning for Improved Neural Conversational
Models | [
"Sean Paulsen"
] | Designing machine intelligence to converse with a human user necessarily
requires an understanding of how humans participate in conversation, and thus
conversation modeling is an important task in natural language processing. New
breakthroughs in architecture and data gathering continue to push the
performance of such conversational AI models. However, designs neglect the
gradual buildup in sentence structure and complexity experienced by humans as
we learn to communicate. During training, our model accepts one or more
sentences as input and attempts to predict the next sentence in the
conversation one word at a time, so our goal is to separate training into
segments, with each segment's corpus comprised of longer sentence pairs than
the previous one. This will mimic the desired "buildup" component of human
learning. We begin with only "short" length sentence pairs, then only "medium"
length pairs, and so on. A majority of our experiments were toward optimizing
this technique, ensuring a proper representation of the technique's potential,
since many of the details were new questions. Our segment-trained models were
then able to achieve lower validation loss at the end of training than models
trained with standard text preparation. This segmented training is
straightforward to implement and our results provide a general direction for
future research to implement and improve it. | [
"cs.CL",
"cs.LG"
] | false |
2305.08982 | 2023-05-15T19:48:59Z | Helping the Helper: Supporting Peer Counselors via AI-Empowered Practice
and Feedback | [
"Shang-Ling Hsu",
"Raj Sanjay Shah",
"Prathik Senthil",
"Zahra Ashktorab",
"Casey Dugan",
"Werner Geyer",
"Diyi Yang"
] | Millions of users come to online peer counseling platforms to seek support on
diverse topics ranging from relationship stress to anxiety. However, studies
show that online peer support groups are not always as effective as expected
largely due to users' negative experiences with unhelpful counselors. Peer
counselors are key to the success of online peer counseling platforms, but most
of them often do not have systematic ways to receive guidelines or supervision.
In this work, we introduce CARE: an interactive AI-based tool to empower peer
counselors through automatic suggestion generation. During the practical
training stage, CARE helps diagnose which specific counseling strategies are
most suitable in the given context and provides tailored example responses as
suggestions. Counselors can choose to select, modify, or ignore any suggestion
before replying to the support seeker. Building upon the Motivational
Interviewing framework, CARE utilizes large-scale counseling conversation data
together with advanced natural language generation techniques to achieve these
functionalities. We demonstrate the efficacy of CARE by performing both
quantitative evaluations and qualitative user studies through simulated chats
and semi-structured interviews. We also find that CARE especially helps novice
counselors respond better in challenging situations. | [
"cs.HC",
"cs.CL"
] | false |
2305.09025 | 2023-05-15T21:17:17Z | Soft Prompt Decoding for Multilingual Dense Retrieval | [
"Zhiqi Huang",
"Hansi Zeng",
"Hamed Zamani",
"James Allan"
] | In this work, we explore a Multilingual Information Retrieval (MLIR) task,
where the collection includes documents in multiple languages. We demonstrate
that applying state-of-the-art approaches developed for cross-lingual
information retrieval to MLIR tasks leads to sub-optimal performance. This is
due to the heterogeneous and imbalanced nature of multilingual collections --
some languages are better represented in the collection and some benefit from
large-scale training data. To address this issue, we present KD-SPD, a novel
soft prompt decoding approach for MLIR that implicitly "translates" the
representation of documents in different languages into the same embedding
space. To address the challenges of data scarcity and imbalance, we introduce a
knowledge distillation strategy. The teacher model is trained on rich English
retrieval data, and by leveraging bi-text data, our distillation framework
transfers its retrieval knowledge to the multilingual document encoder.
Therefore, our approach does not require any multilingual retrieval training
data. Extensive experiments on three MLIR datasets with a total of 15 languages
demonstrate that KD-SPD significantly outperforms competitive baselines in all
cases. We conduct extensive analyses to show that our method has less language
bias and better zero-shot transfer ability towards new languages. | [
"cs.IR",
"cs.CL"
] | false |