text
stringlengths 29
3.31k
| label
sequencelengths 1
11
|
---|---|
Visual object tracking, which is representing a major interest in image
processing field, has facilitated numerous real world applications. Among them,
equipping unmanned aerial vehicle (UAV) with real time robust visual trackers
for all day aerial maneuver, is currently attracting incremental attention and
has remarkably broadened the scope of applications of object tracking. However,
prior tracking methods have merely focused on robust tracking in the
well-illuminated scenes, while ignoring trackers' capabilities to be deployed
in the dark. In darkness, the conditions can be more complex and harsh, easily
posing inferior robust tracking or even tracking failure. To this end, this
work proposed a novel discriminative correlation filter based tracker with
illumination adaptive and anti dark capability, namely ADTrack. ADTrack firstly
exploits image illuminance information to enable adaptability of the model to
the given light condition. Then, by virtue of an efficient and effective image
enhancer, ADTrack carries out image pretreatment, where a target aware mask is
generated. Benefiting from the mask, ADTrack aims to solve a dual regression
problem where dual filters, i.e., the context filter and target focused filter,
are trained with mutual constraint. Thus ADTrack is able to maintain
continuously favorable performance in all-day conditions. Besides, this work
also constructed one UAV nighttime tracking benchmark UAVDark135, comprising of
more than 125k manually annotated frames, which is also very first UAV
nighttime tracking benchmark. Exhaustive experiments are extended on
authoritative daytime benchmarks, i.e., UAV123 10fps, DTB70, and the newly
built dark benchmark UAVDark135, which have validated the superiority of
ADTrack in both bright and dark conditions on a single CPU. | [
"cs.CV"
] |
Symmetries and equivariance are fundamental to the generalization of neural
networks on domains such as images, graphs, and point clouds. Existing work has
primarily focused on a small number of groups, such as the translation,
rotation, and permutation groups. In this work we provide a completely general
algorithm for solving for the equivariant layers of matrix groups. In addition
to recovering solutions from other works as special cases, we construct
multilayer perceptrons equivariant to multiple groups that have never been
tackled before, including $\mathrm{O}(1,3)$, $\mathrm{O}(5)$, $\mathrm{Sp}(n)$,
and the Rubik's cube group. Our approach outperforms non-equivariant baselines,
with applications to particle physics and dynamical systems. We release our
software library to enable researchers to construct equivariant layers for
arbitrary matrix groups. | [
"cs.LG",
"math.DS",
"stat.ML"
] |
Designing RNA molecules has garnered recent interest in medicine, synthetic
biology, biotechnology and bioinformatics since many functional RNA molecules
were shown to be involved in regulatory processes for transcription,
epigenetics and translation. Since an RNA's function depends on its structural
properties, the RNA Design problem is to find an RNA sequence which satisfies
given structural constraints. Here, we propose a new algorithm for the RNA
Design problem, dubbed LEARNA. LEARNA uses deep reinforcement learning to train
a policy network to sequentially design an entire RNA sequence given a
specified target structure. By meta-learning across 65000 different RNA Design
tasks for one hour on 20 CPU cores, our extension Meta-LEARNA constructs an RNA
Design policy that can be applied out of the box to solve novel RNA Design
tasks. Methodologically, for what we believe to be the first time, we jointly
optimize over a rich space of architectures for the policy network, the
hyperparameters of the training procedure and the formulation of the decision
process. Comprehensive empirical results on two widely-used RNA Design
benchmarks, as well as a third one that we introduce, show that our approach
achieves new state-of-the-art performance on the former while also being orders
of magnitudes faster in reaching the previous state-of-the-art performance. In
an ablation study, we analyze the importance of our method's different
components. | [
"cs.LG",
"q-bio.QM",
"stat.ML"
] |
We introduce a two-stream model for dynamic texture synthesis. Our model is
based on pre-trained convolutional networks (ConvNets) that target two
independent tasks: (i) object recognition, and (ii) optical flow prediction.
Given an input dynamic texture, statistics of filter responses from the object
recognition ConvNet encapsulate the per-frame appearance of the input texture,
while statistics of filter responses from the optical flow ConvNet model its
dynamics. To generate a novel texture, a randomly initialized input sequence is
optimized to match the feature statistics from each stream of an example
texture. Inspired by recent work on image style transfer and enabled by the
two-stream model, we also apply the synthesis approach to combine the texture
appearance from one texture with the dynamics of another to generate entirely
novel dynamic textures. We show that our approach generates novel, high quality
samples that match both the framewise appearance and temporal evolution of
input texture. Finally, we quantitatively evaluate our texture synthesis
approach with a thorough user study. | [
"cs.CV"
] |
Is it possible to use convolutional neural networks pre-trained without any
natural images to assist natural image understanding? The paper proposes a
novel concept, Formula-driven Supervised Learning. We automatically generate
image patterns and their category labels by assigning fractals, which are based
on a natural law existing in the background knowledge of the real world.
Theoretically, the use of automatically generated images instead of natural
images in the pre-training phase allows us to generate an infinite scale
dataset of labeled images. Although the models pre-trained with the proposed
Fractal DataBase (FractalDB), a database without natural images, does not
necessarily outperform models pre-trained with human annotated datasets at all
settings, we are able to partially surpass the accuracy of ImageNet/Places
pre-trained models. The image representation with the proposed FractalDB
captures a unique feature in the visualization of convolutional layers and
attentions. | [
"cs.CV",
"cs.LG"
] |
In this paper, we present a regression-based pose recognition method using
cascade Transformers. One way to categorize the existing approaches in this
domain is to separate them into 1). heatmap-based and 2). regression-based. In
general, heatmap-based methods achieve higher accuracy but are subject to
various heuristic designs (not end-to-end mostly), whereas regression-based
approaches attain relatively lower accuracy but they have less intermediate
non-differentiable steps. Here we utilize the encoder-decoder structure in
Transformers to perform regression-based person and keypoint detection that is
general-purpose and requires less heuristic design compared with the existing
approaches. We demonstrate the keypoint hypothesis (query) refinement process
across different self-attention layers to reveal the recursive self-attention
mechanism in Transformers. In the experiments, we report competitive results
for pose recognition when compared with the competing regression-based methods. | [
"cs.CV"
] |
To achieve peak predictive performance, hyperparameter optimization (HPO) is
a crucial component of machine learning and its applications. Over the last
years,the number of efficient algorithms and tools for HPO grew substantially.
At the same time, the community is still lacking realistic, diverse,
computationally cheap,and standardized benchmarks. This is especially the case
for multi-fidelity HPO methods. To close this gap, we propose HPOBench, which
includes 7 existing and 5 new benchmark families, with in total more than 100
multi-fidelity benchmark problems. HPOBench allows to run this extendable set
of multi-fidelity HPO benchmarks in a reproducible way by isolating and
packaging the individual benchmarks in containers. It also provides surrogate
and tabular benchmarks for computationally affordable yet statistically sound
evaluations. To demonstrate the broad compatibility of HPOBench and its
usefulness, we conduct an exemplary large-scale study evaluating 6 well known
multi-fidelity HPO tools. | [
"cs.LG"
] |
In this work, we focus on a challenging task: synthesizing multiple imaginary
videos given a single image. Major problems come from high dimensionality of
pixel space and the ambiguity of potential motions. To overcome those problems,
we propose a new framework that produce imaginary videos by transformation
generation. The generated transformations are applied to the original image in
a novel volumetric merge network to reconstruct frames in imaginary video.
Through sampling different latent variables, our method can output different
imaginary video samples. The framework is trained in an adversarial way with
unsupervised learning. For evaluation, we propose a new assessment metric
$RIQA$. In experiments, we test on 3 datasets varying from synthetic data to
natural scene. Our framework achieves promising performance in image quality
assessment. The visual inspection indicates that it can successfully generate
diverse five-frame videos in acceptable perceptual quality. | [
"cs.CV"
] |
Self-supervised vision-and-language pretraining (VLP) aims to learn
transferable multi-modal representations from large-scale image-text data and
to achieve strong performances on a broad scope of vision-language tasks after
finetuning. Previous mainstream VLP approaches typically adopt a two-step
strategy relying on external object detectors to encode images in a multi-modal
Transformer framework, which suffer from restrictive object concept space,
limited image context and inefficient computation. In this paper, we propose an
object-aware end-to-end VLP framework, which directly feeds image grid features
from CNNs into the Transformer and learns the multi-modal representations
jointly. More importantly, we propose to perform object knowledge distillation
to facilitate learning cross-modal alignment at different semantic levels. To
achieve that, we design two novel pretext tasks by taking object features and
their semantic labels from external detectors as supervision: 1.) Object-guided
masked vision modeling task focuses on enforcing object-aware representation
learning in the multi-modal Transformer; 2.) Phrase-region alignment task aims
to improve cross-modal alignment by utilizing the similarities between noun
phrases and object labels in the linguistic space. Extensive experiments on a
wide range of vision-language tasks demonstrate the efficacy of our proposed
framework, and we achieve competitive or superior performances over the
existing pretraining strategies. The code is available in supplementary
materials. | [
"cs.CV"
] |
Sample efficiency is a critical property when optimizing policy parameters
for the controller of a robot. In this paper, we evaluate two state-of-the-art
policy optimization algorithms. One is a recent deep reinforcement learning
method based on an actor-critic algorithm, Deep Deterministic Policy Gradient
(DDPG), that has been shown to perform well on various control benchmarks. The
other one is a direct policy search method, Covariance Matrix Adaptation
Evolution Strategy (CMA-ES), a black-box optimization method that is widely
used for robot learning. The algorithms are evaluated on a continuous version
of the mountain car benchmark problem, so as to compare their sample
complexity. From a preliminary analysis, we expect DDPG to be more sample
efficient than CMA-ES, which is confirmed by our experimental results. | [
"cs.LG"
] |
Video anomaly detection is of critical practical importance to a variety of
real applications because it allows human attention to be focused on events
that are likely to be of interest, in spite of an otherwise overwhelming volume
of video. We show that applying self-trained deep ordinal regression to video
anomaly detection overcomes two key limitations of existing methods, namely, 1)
being highly dependent on manually labeled normal training data; and 2)
sub-optimal feature learning. By formulating a surrogate two-class ordinal
regression task we devise an end-to-end trainable video anomaly detection
approach that enables joint representation learning and anomaly scoring without
manually labeled normal/abnormal data. Experiments on eight real-world video
scenes show that our proposed method outperforms state-of-the-art methods that
require no labeled training data by a substantial margin, and enables easy and
accurate localization of the identified anomalies. Furthermore, we demonstrate
that our method offers effective human-in-the-loop anomaly detection which can
be critical in applications where anomalies are rare and the false-negative
cost is high. | [
"cs.CV"
] |
Reconstructing 3D models from 2D images is one of the fundamental problems in
computer vision. In this work, we propose a deep learning technique for 3D
object reconstruction from a single image. Contrary to recent works that either
use 3D supervision or multi-view supervision, we use only single view images
with no pose information during training as well. This makes our approach more
practical requiring only an image collection of an object category and the
corresponding silhouettes. We learn both 3D point cloud reconstruction and pose
estimation networks in a self-supervised manner, making use of differentiable
point cloud renderer to train with 2D supervision. A key novelty of the
proposed technique is to impose 3D geometric reasoning into predicted 3D point
clouds by rotating them with randomly sampled poses and then enforcing cycle
consistency on both 3D reconstructions and poses. In addition, using
single-view supervision allows us to do test-time optimization on a given test
image. Experiments on the synthetic ShapeNet and real-world Pix3D datasets
demonstrate that our approach, despite using less supervision, can achieve
competitive performance compared to pose-supervised and multi-view supervised
approaches. | [
"cs.CV"
] |
Computational color constancy has the important task of reducing the
influence of the scene illumination on the object colors. As such, it is an
essential part of the image processing pipelines of most digital cameras. One
of the important parts of the computational color constancy is illumination
estimation, i.e. estimating the illumination color. When an illumination
estimation method is proposed, its accuracy is usually reported by providing
the values of error metrics obtained on the images of publicly available
datasets. However, over time it has been shown that many of these datasets have
problems such as too few images, inappropriate image quality, lack of scene
diversity, absence of version tracking, violation of various assumptions, GDPR
regulation violation, lack of additional shooting procedure info, etc. In this
paper, a new illumination estimation dataset is proposed that aims to alleviate
many of the mentioned problems and to help the illumination estimation
research. It consists of 4890 images with known illumination colors as well as
with additional semantic data that can further make the learning process more
accurate. Due to the usage of the SpyderCube color target, for every image
there are two ground-truth illumination records covering different directions.
Because of that, the dataset can be used for training and testing of methods
that perform single or two-illuminant estimation. This makes it superior to
many similar existing datasets. The datasets, it's smaller version
SimpleCube++, and the accompanying code are available at
https://github.com/Visillect/CubePlusPlus/. | [
"cs.CV"
] |
A generalist robot must be able to complete a variety of tasks in its
environment. One appealing way to specify each task is in terms of a goal
observation. However, learning goal-reaching policies with reinforcement
learning remains a challenging problem, particularly when hand-engineered
reward functions are not available. Learned dynamics models are a promising
approach for learning about the environment without rewards or task-directed
data, but planning to reach goals with such a model requires a notion of
functional similarity between observations and goal states. We present a
self-supervised method for model-based visual goal reaching, which uses both a
visual dynamics model as well as a dynamical distance function learned using
model-free reinforcement learning. Our approach learns entirely using offline,
unlabeled data, making it practical to scale to large and diverse datasets. In
our experiments, we find that our method can successfully learn models that
perform a variety of tasks at test-time, moving objects amid distractors with a
simulated robotic arm and even learning to open and close a drawer using a
real-world robot. In comparisons, we find that this approach substantially
outperforms both model-free and model-based prior methods. Videos and
visualizations are available here: http://sites.google.com/berkeley.edu/mbold. | [
"cs.LG",
"cs.AI",
"cs.CV",
"cs.RO"
] |
With the broad use of face recognition, its weakness gradually emerges that
it is able to be attacked. So, it is important to study how face recognition
networks are subject to attacks. In this paper, we focus on a novel way to do
attacks against face recognition network that misleads the network to identify
someone as the target person not misclassify inconspicuously. Simultaneously,
for this purpose, we introduce a specific attentional adversarial attack
generative network to generate fake face images. For capturing the semantic
information of the target person, this work adds a conditional variational
autoencoder and attention modules to learn the instance-level correspondences
between faces. Unlike traditional two-player GAN, this work introduces face
recognition networks as the third player to participate in the competition
between generator and discriminator which allows the attacker to impersonate
the target person better. The generated faces which are hard to arouse the
notice of onlookers can evade recognition by state-of-the-art networks and most
of them are recognized as the target person. | [
"cs.CV"
] |
Variants of Graph Neural Networks (GNNs) for representation learning have
been proposed recently and achieved fruitful results in various fields. Among
them, Graph Attention Network (GAT) first employs a self-attention strategy to
learn attention weights for each edge in the spatial domain. However, learning
the attentions over edges can only focus on the local information of graphs and
greatly increases the computational costs. In this paper, we first introduce
the attention mechanism in the spectral domain of graphs and present Spectral
Graph Attention Network (SpGAT) that learns representations for different
frequency components regarding weighted filters and graph wavelets bases. In
this way, SpGAT can better capture global patterns of graphs in an efficient
manner with much fewer learned parameters than that of GAT. Further, to reduce
the computational cost of SpGAT brought by the eigen-decomposition, we propose
a fast approximation variant SpGAT-Cheby. We thoroughly evaluate the
performance of SpGAT and SpGAT-Cheby in semi-supervised node classification
tasks and verify the effectiveness of the learned attentions in the spectral
domain. | [
"cs.LG",
"cs.SI",
"stat.ML"
] |
Sparse labels have been attracting much attention in recent years. However,
the performance gap between weakly supervised and fully supervised salient
object detection methods is huge, and most previous weakly supervised works
adopt complex training methods with many bells and whistles. In this work, we
propose a one-round end-to-end training approach for weakly supervised salient
object detection via scribble annotations without pre/post-processing
operations or extra supervision data. Since scribble labels fail to offer
detailed salient regions, we propose a local coherence loss to propagate the
labels to unlabeled regions based on image features and pixel distance, so as
to predict integral salient regions with complete object structures. We design
a saliency structure consistency loss as self-consistent mechanism to ensure
consistent saliency maps are predicted with different scales of the same image
as input, which could be viewed as a regularization technique to enhance the
model generalization ability. Additionally, we design an aggregation module
(AGGM) to better integrate high-level features, low-level features and global
context information for the decoder to aggregate various information. Extensive
experiments show that our method achieves a new state-of-the-art performance on
six benchmarks (e.g. for the ECSSD dataset: F_\beta = 0.8995, E_\xi = 0.9079
and MAE = 0.0489$), with an average gain of 4.60\% for F-measure, 2.05\% for
E-measure and 1.88\% for MAE over the previous best method on this task. Source
code is available at http://github.com/siyueyu/SCWSSOD. | [
"cs.CV"
] |
Detection of objects in cluttered indoor environments is one of the key
enabling functionalities for service robots. The best performing object
detection approaches in computer vision exploit deep Convolutional Neural
Networks (CNN) to simultaneously detect and categorize the objects of interest
in cluttered scenes. Training of such models typically requires large amounts
of annotated training data which is time consuming and costly to obtain. In
this work we explore the ability of using synthetically generated composite
images for training state-of-the-art object detectors, especially for object
instance detection. We superimpose 2D images of textured object models into
images of real environments at variety of locations and scales. Our experiments
evaluate different superimposition strategies ranging from purely image-based
blending all the way to depth and semantics informed positioning of the object
models into real scenes. We demonstrate the effectiveness of these object
detector training strategies on two publicly available datasets, the
GMU-Kitchens and the Washington RGB-D Scenes v2. As one observation, augmenting
some hand-labeled training data with synthetic examples carefully composed onto
scenes yields object detectors with comparable performance to using much more
hand-labeled data. Broadly, this work charts new opportunities for training
detectors for new objects by exploiting existing object model repositories in
either a purely automatic fashion or with only a very small number of
human-annotated examples. | [
"cs.CV",
"cs.RO"
] |
We study the problem of evaluating super resolution methods. Traditional
evaluation methods usually judge the quality of super resolved images based on
a single measure of their difference with the original high resolution images.
In this paper, we proposed to use both fidelity (the difference with original
images) and naturalness (human visual perception of super resolved images) for
evaluation. For fidelity evaluation, a new metric is proposed to solve the bias
problem of traditional evaluation. For naturalness evaluation, we let humans
label preference of super resolution results using pair-wise comparison, and
test the correlation between human labeling results and image quality
assessment metrics' outputs. Experimental results show that our
fidelity-naturalness method is better than the traditional evaluation method
for super resolution methods, which could help future research on single-image
super resolution. | [
"cs.CV"
] |
Many recent algorithms for reinforcement learning are model-free and founded
on the Bellman equation. Here we present a method founded on the costate
equation and models of the state dynamics. We use the costate -- the gradient
of cost with respect to state -- to improve the policy and also to "focus" the
model, training it to detect and mimic those features of the environment that
are most relevant to its task. We show that this method can handle difficult
time-optimal control problems, driving deterministic or stochastic mechanical
systems quickly to a target. On these tasks it works well compared to deep
deterministic policy gradient, a recent Bellman method. And because it creates
a model, the costate method can also learn from mental practice. | [
"cs.LG",
"math.OC"
] |
Most SLAM algorithms are based on the assumption that the scene is static.
However, in practice, most scenes are dynamic which usually contains moving
objects, these methods are not suitable. In this paper, we introduce DymSLAM, a
dynamic stereo visual SLAM system being capable of reconstructing a 4D (3D +
time) dynamic scene with rigid moving objects. The only input of DymSLAM is
stereo video, and its output includes a dense map of the static environment, 3D
model of the moving objects and the trajectories of the camera and the moving
objects. We at first detect and match the interesting points between successive
frames by using traditional SLAM methods. Then the interesting points belonging
to different motion models (including ego-motion and motion models of rigid
moving objects) are segmented by a multi-model fitting approach. Based on the
interesting points belonging to the ego-motion, we are able to estimate the
trajectory of the camera and reconstruct the static background. The interesting
points belonging to the motion models of rigid moving objects are then used to
estimate their relative motion models to the camera and reconstruct the 3D
models of the objects. We then transform the relative motion to the
trajectories of the moving objects in the global reference frame. Finally, we
then fuse the 3D models of the moving objects into the 3D map of the
environment by considering their motion trajectories to obtain a 4D (3D+time)
sequence. DymSLAM obtains information about the dynamic objects instead of
ignoring them and is suitable for unknown rigid objects. Hence, the proposed
system allows the robot to be employed for high-level tasks, such as obstacle
avoidance for dynamic objects. We conducted experiments in a real-world
environment where both the camera and the objects were moving in a wide range. | [
"cs.CV",
"cs.RO"
] |
Due to the sparsity and irregularity of the point cloud data, methods that
directly consume points have become popular. Among all point-based models,
graph convolutional networks (GCN) lead to notable performance by fully
preserving the data granularity and exploiting point interrelation. However,
point-based networks spend a significant amount of time on data structuring
(e.g., Farthest Point Sampling (FPS) and neighbor points querying), which limit
the speed and scalability. In this paper, we present a method, named Grid-GCN,
for fast and scalable point cloud learning. Grid-GCN uses a novel data
structuring strategy, Coverage-Aware Grid Query (CAGQ). By leveraging the
efficiency of grid space, CAGQ improves spatial coverage while reducing the
theoretical time complexity. Compared with popular sampling methods such as
Farthest Point Sampling (FPS) and Ball Query, CAGQ achieves up to 50X speed-up.
With a Grid Context Aggregation (GCA) module, Grid-GCN achieves
state-of-the-art performance on major point cloud classification and
segmentation benchmarks with significantly faster runtime than previous
studies. Remarkably, Grid-GCN achieves the inference speed of 50fps on ScanNet
using 81920 points per scene as input. | [
"cs.CV",
"cs.LG"
] |
Knowledge Distillation (KD) methods are capable of transferring the knowledge
encoded in a large and complex teacher into a smaller and faster student. Early
methods were usually limited to transferring the knowledge only between the
last layers of the networks, while latter approaches were capable of performing
multi-layer KD, further increasing the accuracy of the student. However,
despite their improved performance, these methods still suffer from several
limitations that restrict both their efficiency and flexibility. First,
existing KD methods typically ignore that neural networks undergo through
different learning phases during the training process, which often requires
different types of supervision for each one. Furthermore, existing multi-layer
KD methods are usually unable to effectively handle networks with significantly
different architectures (heterogeneous KD). In this paper we propose a novel KD
method that works by modeling the information flow through the various layers
of the teacher model and then train a student model to mimic this information
flow. The proposed method is capable of overcoming the aforementioned
limitations by using an appropriate supervision scheme during the different
phases of the training process, as well as by designing and training an
appropriate auxiliary teacher model that acts as a proxy model capable of
"explaining" the way the teacher works to the student. The effectiveness of the
proposed method is demonstrated using four image datasets and several different
evaluation setups. | [
"cs.CV"
] |
Semantic Segmentation is a crucial component in the perception systems of
many applications, such as robotics and autonomous driving that rely on
accurate environmental perception and understanding. In literature, several
approaches are introduced to attempt LiDAR semantic segmentation task, such as
projection-based (range-view or birds-eye-view), and voxel-based approaches.
However, they either abandon the valuable 3D topology and geometric relations
and suffer from information loss introduced in the projection process or are
inefficient. Therefore, there is a need for accurate models capable of
processing the 3D driving-scene point cloud in 3D space. In this paper, we
propose S3Net, a novel convolutional neural network for LiDAR point cloud
semantic segmentation. It adopts an encoder-decoder backbone that consists of
Sparse Intra-channel Attention Module (SIntraAM), and Sparse Inter-channel
Attention Module (SInterAM) to emphasize the fine details of both within each
feature map and among nearby feature maps. To extract the global contexts in
deeper layers, we introduce Sparse Residual Tower based upon sparse convolution
that suits varying sparsity of LiDAR point cloud. In addition, geo-aware
anisotrophic loss is leveraged to emphasize the semantic boundaries and
penalize the noise within each predicted regions, leading to a robust
prediction. Our experimental results show that the proposed method leads to a
large improvement (12\%) compared to its baseline counterpart (MinkNet42
\cite{choy20194d}) on SemanticKITTI \cite{DBLP:conf/iccv/BehleyGMQBSG19} test
set and achieves state-of-the-art mIoU accuracy of semantic segmentation
approaches. | [
"cs.CV"
] |
Scene graph generation refers to the task of automatically mapping an image
into a semantic structural graph, which requires correctly labeling each
extracted object and their interaction relationships. Despite the recent
success in object detection using deep learning techniques, inferring complex
contextual relationships and structured graph representations from visual data
remains a challenging topic. In this study, we propose a novel Attentive
Relational Network that consists of two key modules with an object detection
backbone to approach this problem. The first module is a semantic
transformation module utilized to capture semantic embedded relation features,
by translating visual features and linguistic features into a common semantic
space. The other module is a graph self-attention module introduced to embed a
joint graph representation through assigning various importance weights to
neighboring nodes. Finally, accurate scene graphs are produced by the relation
inference module to recognize all entities and the corresponding relations. We
evaluate our proposed method on the widely-adopted Visual Genome Dataset, and
the results demonstrate the effectiveness and superiority of our model. | [
"cs.CV"
] |
Heterogeneous presentation of a neurological disorder suggests potential
differences in the underlying pathophysiological changes that occur in the
brain. We propose to model heterogeneous patterns of functional network
differences using a demographic-guided attention (DGA) mechanism for recurrent
neural network models for prediction from functional magnetic resonance imaging
(fMRI) time-series data. The context computed from the DGA head is used to help
focus on the appropriate functional networks based on individual demographic
information. We demonstrate improved classification on 3 subsets of the ABIDE I
dataset used in published studies that have previously produced
state-of-the-art results, evaluating performance under a leave-one-site-out
cross-validation framework for better generalizeability to new data. Finally,
we provide examples of interpreting functional network differences based on
individual demographic variables. | [
"cs.LG",
"cs.CV",
"eess.IV",
"q-bio.QM",
"stat.AP"
] |
Monocular 3D object detection is an important task in autonomous driving. It
can be easily intractable where there exists ego-car pose change w.r.t. ground
plane. This is common due to the slight fluctuation of road smoothness and
slope. Due to the lack of insight in industrial application, existing methods
on open datasets neglect the camera pose information, which inevitably results
in the detector being susceptible to camera extrinsic parameters. The
perturbation of objects is very popular in most autonomous driving cases for
industrial products. To this end, we propose a novel method to capture camera
pose to formulate the detector free from extrinsic perturbation. Specifically,
the proposed framework predicts camera extrinsic parameters by detecting
vanishing point and horizon change. A converter is designed to rectify
perturbative features in the latent space. By doing so, our 3D detector works
independent of the extrinsic parameter variations and produces accurate results
in realistic cases, e.g., potholed and uneven roads, where almost all existing
monocular detectors fail to handle. Experiments demonstrate our method yields
the best performance compared with the other state-of-the-arts by a large
margin on both KITTI 3D and nuScenes datasets. | [
"cs.CV"
] |
Object-oriented maps are important for scene understanding since they jointly
capture geometry and semantics, allow individual instantiation and meaningful
reasoning about objects. We introduce FroDO, a method for accurate 3D
reconstruction of object instances from RGB video that infers object location,
pose and shape in a coarse-to-fine manner. Key to FroDO is to embed object
shapes in a novel learnt space that allows seamless switching between sparse
point cloud and dense DeepSDF decoding. Given an input sequence of localized
RGB frames, FroDO first aggregates 2D detections to instantiate a
category-aware 3D bounding box per object. A shape code is regressed using an
encoder network before optimizing shape and pose further under the learnt shape
priors using sparse and dense shape representations. The optimization uses
multi-view geometric, photometric and silhouette losses. We evaluate on
real-world datasets, including Pix3D, Redwood-OS, and ScanNet, for single-view,
multi-view, and multi-object reconstruction. | [
"cs.CV"
] |
We address the challenging task of cross-modal moment retrieval, which aims
to localize a temporal segment from an untrimmed video described by a natural
language query. It poses great challenges over the proper semantic alignment
between vision and linguistic domains. Existing methods independently extract
the features of videos and sentences and purely utilize the sentence embedding
in the multi-modal fusion stage, which do not make full use of the potential of
language. In this paper, we present Language Guided Networks (LGN), a new
framework that leverages the sentence embedding to guide the whole process of
moment retrieval. In the first feature extraction stage, we propose to jointly
learn visual and language features to capture the powerful visual information
which can cover the complex semantics in the sentence query. Specifically, the
early modulation unit is designed to modulate the visual feature extractor's
feature maps by a linguistic embedding. Then we adopt a multi-modal fusion
module in the second fusion stage. Finally, to get a precise localizer, the
sentence information is utilized to guide the process of predicting temporal
positions. Specifically, the late guidance module is developed to linearly
transform the output of localization networks via the channel attention
mechanism. The experimental results on two popular datasets demonstrate the
superior performance of our proposed method on moment retrieval (improving by
5.8\% in terms of [email protected] on Charades-STA and 5.2\% on TACoS). The source
code for the complete system will be publicly available. | [
"cs.CV"
] |
We present a novel architecture for 3D object detection, M3DeTR, which
combines different point cloud representations (raw, voxels, bird-eye view)
with different feature scales based on multi-scale feature pyramids. M3DeTR is
the first approach that unifies multiple point cloud representations, feature
scales, as well as models mutual relationships between point clouds
simultaneously using transformers. We perform extensive ablation experiments
that highlight the benefits of fusing representation and scale, and modeling
the relationships. Our method achieves state-of-the-art performance on the
KITTI 3D object detection dataset and Waymo Open Dataset. Results show that
M3DeTR improves the baseline significantly by 1.48% mAP for all classes on
Waymo Open Dataset. In particular, our approach ranks 1st on the well-known
KITTI 3D Detection Benchmark for both car and cyclist classes, and ranks 1st on
Waymo Open Dataset with single frame point cloud input. | [
"cs.CV"
] |
It is a long-standing question to discover causal relations among a set of
variables in many empirical sciences. Recently, Reinforcement Learning (RL) has
achieved promising results in causal discovery from observational data.
However, searching the space of directed graphs and enforcing acyclicity by
implicit penalties tend to be inefficient and restrict the existing RL-based
method to small scale problems. In this work, we propose a novel RL-based
approach for causal discovery, by incorporating RL into the ordering-based
paradigm. Specifically, we formulate the ordering search problem as a
multi-step Markov decision process, implement the ordering generating process
with an encoder-decoder architecture, and finally use RL to optimize the
proposed model based on the reward mechanisms designed for~each ordering. A
generated ordering would then be processed using variable selection to obtain
the final causal graph. We analyze the consistency and computational complexity
of the proposed method, and empirically show that a pretrained model can be
exploited to accelerate training. Experimental results on both synthetic and
real data sets shows that the proposed method achieves a much improved
performance over existing RL-based method. | [
"cs.LG"
] |
Audio-visual event localization aims to localize an event that is both
audible and visible in the wild, which is a widespread audio-visual scene
analysis task for unconstrained videos. To address this task, we propose a
Multimodal Parallel Network (MPN), which can perceive global semantics and
unmixed local information parallelly. Specifically, our MPN framework consists
of a classification subnetwork to predict event categories and a localization
subnetwork to predict event boundaries. The classification subnetwork is
constructed by the Multimodal Co-attention Module (MCM) and obtains global
contexts. The localization subnetwork consists of Multimodal Bottleneck
Attention Module (MBAM), which is designed to extract fine-grained
segment-level contents. Extensive experiments demonstrate that our framework
achieves the state-of-the-art performance both in fully supervised and weakly
supervised settings on the Audio-Visual Event (AVE) dataset. | [
"cs.CV",
"cs.AI",
"cs.MM"
] |
Object detection involves two sub-tasks, i.e. localizing objects in an image
and classifying them into various categories. For existing CNN-based detectors,
we notice the widespread divergence between localization and classification,
which leads to degradation in performance. In this work, we propose a mutual
learning framework to modulate the two tasks. In particular, the two tasks are
forced to learn from each other with a novel mutual labeling strategy. Besides,
we introduce a simple yet effective IoU rescoring scheme, which further reduces
the divergence. Moreover, we define a Spearman rank correlation-based metric to
quantify the divergence, which correlates well with the detection performance.
The proposed approach is general-purpose and can be easily injected into
existing detectors such as FCOS and RetinaNet. We achieve a significant
performance gain over the baseline detectors on the COCO dataset. | [
"cs.CV"
] |
In this paper, we tackle the problem of online road network extraction from
sparse 3D point clouds. Our method is inspired by how an annotator builds a
lane graph, by first identifying how many lanes there are and then drawing each
one in turn. We develop a hierarchical recurrent network that attends to
initial regions of a lane boundary and traces them out completely by outputting
a structured polyline. We also propose a novel differentiable loss function
that measures the deviation of the edges of the ground truth polylines and
their predictions. This is more suitable than distances on vertices, as there
exists many ways to draw equivalent polylines. We demonstrate the effectiveness
of our method on a 90 km stretch of highway, and show that we can recover the
right topology 92\% of the time. | [
"cs.CV"
] |
Point cloud is an important type of geometric data structure. Due to its
irregular format, most researchers transform such data to regular 3D voxel
grids or collections of images. This, however, renders data unnecessarily
voluminous and causes issues. In this paper, we design a novel type of neural
network that directly consumes point clouds and well respects the permutation
invariance of points in the input. Our network, named PointNet, provides a
unified architecture for applications ranging from object classification, part
segmentation, to scene semantic parsing. Though simple, PointNet is highly
efficient and effective. Empirically, it shows strong performance on par or
even better than state of the art. Theoretically, we provide analysis towards
understanding of what the network has learnt and why the network is robust with
respect to input perturbation and corruption. | [
"cs.CV"
] |
This paper presents an unobtrusive solution that can automatically identify
deep breath when a person is walking past the global depth camera. Existing
non-contact breath assessments achieve satisfactory results under restricted
conditions when human body stays relatively still. When someone moves forward,
the breath signals detected by depth camera are hidden within signals of trunk
displacement and deformation, and the signal length is short due to the short
stay time, posing great challenges for us to establish models. To overcome
these challenges, multiple region of interests (ROIs) based signal extraction
and selection method is proposed to automatically obtain the signal informative
to breath from depth video. Subsequently, graph signal analysis (GSA) is
adopted as a spatial-temporal filter to wipe the components unrelated to
breath. Finally, a classifier for identifying deep breath is established based
on the selected breath-informative signal. In validation experiments, the
proposed approach outperforms the comparative methods with the accuracy,
precision, recall and F1 of 75.5%, 76.2%, 75.0% and 75.2%, respectively. This
system can be extended to public places to provide timely and ubiquitous help
for those who may have or are going through physical or mental trouble. | [
"cs.CV",
"cs.MM",
"cs.SY",
"eess.SY"
] |
Deep Reinforcement Learning (DRL) has shown impressive performance on domains
with visual inputs, in particular various games. However, the agent is usually
trained on a fixed environment, e.g. a fixed number of levels. A growing mass
of evidence suggests that these trained models fail to generalize to even
slight variations of the environments they were trained on. This paper advances
the hypothesis that the lack of generalization is partly due to the input
representation, and explores how rotation, cropping and translation could
increase generality. We show that a cropped, translated and rotated observation
can get better generalization on unseen levels of two-dimensional arcade games
from the GVGAI framework. The generality of the agents is evaluated on both
human-designed and procedurally generated levels. | [
"cs.LG",
"cs.CV",
"stat.ML"
] |
Building an interactive artificial intelligence that can ask questions about
the real world is one of the biggest challenges for vision and language
problems. In particular, goal-oriented visual dialogue, where the aim of the
agent is to seek information by asking questions during a turn-taking dialogue,
has been gaining scholarly attention recently. While several existing models
based on the GuessWhat?! dataset have been proposed, the Questioner typically
asks simple category-based questions or absolute spatial questions. This might
be problematic for complex scenes where the objects share attributes or in
cases where descriptive questions are required to distinguish objects. In this
paper, we propose a novel Questioner architecture, called Unified Questioner
Transformer (UniQer), for descriptive question generation with referring
expressions. In addition, we build a goal-oriented visual dialogue task called
CLEVR Ask. It synthesizes complex scenes that require the Questioner to
generate descriptive questions. We train our model with two variants of CLEVR
Ask datasets. The results of the quantitative and qualitative evaluations show
that UniQer outperforms the baseline. | [
"cs.CV"
] |
In this work, we take a representation learning perspective on hierarchical
reinforcement learning, where the problem of learning lower layers in a
hierarchy is transformed into the problem of learning trajectory-level
generative models. We show that we can learn continuous latent representations
of trajectories, which are effective in solving temporally extended and
multi-stage problems. Our proposed model, SeCTAR, draws inspiration from
variational autoencoders, and learns latent representations of trajectories. A
key component of this method is to learn both a latent-conditioned policy and a
latent-conditioned model which are consistent with each other. Given the same
latent, the policy generates a trajectory which should match the trajectory
predicted by the model. This model provides a built-in prediction mechanism, by
predicting the outcome of closed loop policy behavior. We propose a novel
algorithm for performing hierarchical RL with this model, combining model-based
planning in the learned latent space with an unsupervised exploration
objective. We show that our model is effective at reasoning over long horizons
with sparse rewards for several simulated tasks, outperforming standard
reinforcement learning methods and prior methods for hierarchical reasoning,
model-based planning, and exploration. | [
"cs.LG",
"cs.AI",
"stat.ML"
] |
Sub-pixel registration is a crucial step for applications such as
super-resolution in remote sensing, motion compensation in magnetic resonance
imaging, and non-destructive testing in manufacturing, to name a few. Recently,
these technologies have been trending towards wavelet encoded imaging and
sparse/compressive sensing. The former plays a crucial role in reducing imaging
artifacts, while the latter significantly increases the acquisition speed. In
view of these new emerging needs for applications of wavelet encoded imaging,
we propose a sub-pixel registration method that can achieve direct wavelet
domain registration from a sparse set of coefficients. We make the following
contributions: (i) We devise a method of decoupling scale, rotation, and
translation parameters in the Haar wavelet domain, (ii) We derive explicit
mathematical expressions that define in-band sub-pixel registration in terms of
wavelet coefficients, (iii) Using the derived expressions, we propose an
approach to achieve in-band subpixel registration, avoiding back and forth
transformations. (iv) Our solution remains highly accurate even when a sparse
set of coefficients are used, which is due to localization of signals in a
sparse set of wavelet coefficients. We demonstrate the accuracy of our method,
and show that it outperforms the state-of-the-art on simulated and real data,
even when the data is sparse. | [
"cs.CV"
] |
In this paper, we propose TauRieL and target Traveling Salesman Problem (TSP)
since it has broad applicability in theoretical and applied sciences. TauRieL
utilizes an actor-critic inspired architecture that adopts ordinary feedforward
nets to obtain a policy update vector $v$. Then, we use $v$ to improve the
state transition matrix from which we generate the policy. Also, the state
transition matrix allows the solver to initialize from precomputed solutions
such as nearest neighbors. In an online learning setting, TauRieL unifies the
training and the search where it can generate near-optimal results in seconds.
The input to the neural nets in the actor-critic architecture are raw 2-D
inputs, and the design idea behind this decision is to keep neural nets
relatively smaller than the architectures with wide embeddings with the
tradeoff of omitting any distributed representations of the embeddings.
Consequently, TauRieL generates TSP solutions two orders of magnitude faster
per TSP instance as compared to state-of-the-art offline techniques with a
performance impact of 6.1\% in the worst case. | [
"cs.LG",
"cs.AI",
"cs.NE"
] |
We study in this paper how to initialize the parameters of multinomial
logistic regression (a fully connected layer followed with softmax and cross
entropy loss), which is widely used in deep neural network (DNN) models for
classification problems. As logistic regression is widely known not having a
closed-form solution, it is usually randomly initialized, leading to several
deficiencies especially in transfer learning where all the layers except for
the last task-specific layer are initialized using a pre-trained model. The
deficiencies include slow convergence speed, possibility of stuck in local
minimum, and the risk of over-fitting. To address those deficiencies, we first
study the properties of logistic regression and propose a closed-form
approximate solution named regularized Gaussian classifier (RGC). Then we adopt
this approximate solution to initialize the task-specific linear layer and
demonstrate superior performance over random initialization in terms of both
accuracy and convergence speed on various tasks and datasets. For example, for
image classification, our approach can reduce the training time by 10 times and
achieve 3.2% gain in accuracy for Flickr-style classification. For object
detection, our approach can also be 10 times faster in training for the same
accuracy, or 5% better in terms of mAP for VOC 2007 with slightly longer
training. | [
"cs.CV",
"cs.LG"
] |
In this paper we introduce a novel method for general semantic segmentation
that can benefit from general semantics of Convolutional Neural Network (CNN).
Our segmentation proposes visually and semantically coherent image segments. We
use binary encoding of CNN features to overcome the difficulty of the
clustering on the high-dimensional CNN feature space. These binary codes are
very robust against noise and non-semantic changes in the image. These binary
encoding can be embedded into the CNN as an extra layer at the end of the
network. This results in real-time segmentation. To the best of our knowledge
our method is the first attempt on general semantic image segmentation using
CNN. All the previous papers were limited to few number of category of the
images (e.g. PASCAL VOC). Experiments show that our segmentation algorithm
outperform the state-of-the-art non-semantic segmentation methods by large
margin. | [
"cs.CV"
] |
In machine learning, one must acquire labels to help supervise a model that
will be able to generalize to unseen data. However, the labeling process can be
tedious, long, costly, and error-prone. It is often the case that most of our
data is unlabeled. Semi-supervised learning (SSL) alleviates that by making
strong assumptions about the relation between the labels and the input data
distribution. This paradigm has been successful in practice, but most SSL
algorithms end up fully trusting the few available labels. In real life, both
humans and automated systems are prone to mistakes; it is essential that our
algorithms are able to work with labels that are both few and also unreliable.
Our work aims to perform an extensive empirical evaluation of existing
graph-based semi-supervised algorithms, like Gaussian Fields and Harmonic
Functions, Local and Global Consistency, Laplacian Eigenmaps, Graph
Transduction Through Alternating Minimization. To do that, we compare the
accuracy of classifiers while varying the amount of labeled data and label
noise for many different samples. Our results show that, if the dataset is
consistent with SSL assumptions, we are able to detect the noisiest instances,
although this gets harder when the number of available labels decreases. Also,
the Laplacian Eigenmaps algorithm performed better than label propagation when
the data came from high-dimensional clusters. | [
"cs.LG",
"stat.ML"
] |
In this paper, the concept of representation learning based on deep neural
networks is applied as an alternative to the use of handcrafted features in a
method for automatic visual inspection of corroded thermoelectric metallic
pipes. A texture convolutional neural network (TCNN) replaces handcrafted
features based on Local Phase Quantization (LPQ) and Haralick descriptors (HD)
with the advantage of learning an appropriate textural representation and the
decision boundaries into a single optimization process. Experimental results
have shown that it is possible to reach the accuracy of 99.20% in the task of
identifying different levels of corrosion in the internal surface of
thermoelectric pipe walls, while using a compact network that requires much
less effort in tuning parameters when compared to the handcrafted approach
since the TCNN architecture is compact regarding the number of layers and
connections. The observed results open up the possibility of using deep neural
networks in real-time applications such as the automatic inspection of
thermoelectric metal pipes. | [
"cs.CV"
] |
Inspired by the recently proposed successive subspace learning (SSL)
principles, we develop a successive subspace graph transform (SSGT) to address
point cloud attribute compression in this work. The octree geometry structure
is utilized to partition the point cloud, where every node of the octree
represents a point cloud subspace with a certain spatial size. We design a
weighted graph with self-loop to describe the subspace and define a graph
Fourier transform based on the normalized graph Laplacian. The transforms are
applied to large point clouds from the leaf nodes to the root node of the
octree recursively, while the represented subspace is expanded from the
smallest one to the whole point cloud successively. It is shown by experimental
results that the proposed SSGT method offers better R-D performances than the
previous Region Adaptive Haar Transform (RAHT) method. | [
"cs.CV",
"eess.IV",
"eess.SP"
] |
In this paper, we study 1-bit convolutional neural networks (CNNs), of which
both the weights and activations are binary. While efficient, the lacking of
representational capability and the training difficulty impede 1-bit CNNs from
performing as well as real-valued networks. We propose Bi-Real net with a novel
training algorithm to tackle these two challenges. To enhance the
representational capability, we propagate the real-valued activations generated
by each 1-bit convolution via a parameter-free shortcut. To address the
training difficulty, we propose a training algorithm using a tighter
approximation to the derivative of the sign function, a magnitude-aware
gradient for weight updating, a better initialization method, and a two-step
scheme for training a deep network. Experiments on ImageNet show that an
18-layer Bi-Real net with the proposed training algorithm achieves 56.4% top-1
classification accuracy, which is 10% higher than the state-of-the-arts (e.g.,
XNOR-Net) with greater memory saving and lower computational cost. Bi-Real net
is also the first to scale up 1-bit CNNs to an ultra-deep network with 152
layers, and achieves 64.5% top-1 accuracy on ImageNet. A 50-layer Bi-Real net
shows comparable performance to a real-valued network on the depth estimation
task with only a 0.3% accuracy gap. | [
"cs.CV"
] |
We present a novel adaptation of active learning to graph-based
semi-supervised learning (SSL) under non-Gaussian Bayesian models. We present
an approximation of non-Gaussian distributions to adapt previously
Gaussian-based acquisition functions to these more general cases. We develop an
efficient rank-one update for applying "look-ahead" based methods as well as
model retraining. We also introduce a novel "model change" acquisition function
based on these approximations that further expands the available collection of
active learning acquisition functions for such methods. | [
"stat.ML",
"cs.LG"
] |
Visual Question Answering (VQA) has emerged as a Visual Turing Test to
validate the reasoning ability of AI agents. The pivot to existing VQA models
is the joint embedding that is learned by combining the visual features from an
image and the semantic features from a given question. Consequently, a large
body of literature has focused on developing complex joint embedding strategies
coupled with visual attention mechanisms to effectively capture the interplay
between these two modalities. However, modelling the visual and semantic
features in a high dimensional (joint embedding) space is computationally
expensive, and more complex models often result in trivial improvements in the
VQA accuracy. In this work, we systematically study the trade-off between the
model complexity and the performance on the VQA task. VQA models have a diverse
architecture comprising of pre-processing, feature extraction, multimodal
fusion, attention and final classification stages. We specifically focus on the
effect of "multi-modal fusion" in VQA models that is typically the most
expensive step in a VQA pipeline. Our thorough experimental evaluation leads us
to two proposals, one optimized for minimal complexity and the other one
optimized for state-of-the-art VQA performance. | [
"cs.CV",
"cs.CC"
] |
Real-time semantic video segmentation is a challenging task due to the strict
requirements of inference speed. Recent approaches mainly devote great efforts
to reducing the model size for high efficiency. In this paper, we rethink this
problem from a different viewpoint: using knowledge contained in compressed
videos. We propose a simple and effective framework, dubbed TapLab, to tap into
resources from the compressed domain. Specifically, we design a fast feature
warping module using motion vectors for acceleration. To reduce the noise
introduced by motion vectors, we design a residual-guided correction module and
a residual-guided frame selection module using residuals. TapLab significantly
reduces redundant computations of the state-of-the-art fast semantic image
segmentation models, running 3 to 10 times faster with controllable accuracy
degradation. The experimental results show that TapLab achieves 70.6% mIoU on
the Cityscapes dataset at 99.8 FPS with a single GPU card for the 1024x2048
videos. A high-speed version even reaches the speed of 160+ FPS. Codes will be
available soon at https://github.com/Sixkplus/TapLab. | [
"cs.CV"
] |
Recently, Vision Transformer and its variants have shown great promise on
various computer vision tasks. The ability of capturing short- and long-range
visual dependencies through self-attention is arguably the main source for the
success. But it also brings challenges due to quadratic computational overhead,
especially for the high-resolution vision tasks (e.g., object detection). In
this paper, we present focal self-attention, a new mechanism that incorporates
both fine-grained local and coarse-grained global interactions. Using this new
mechanism, each token attends the closest surrounding tokens at fine
granularity but the tokens far away at coarse granularity, and thus can capture
both short- and long-range visual dependencies efficiently and effectively.
With focal self-attention, we propose a new variant of Vision Transformer
models, called Focal Transformer, which achieves superior performance over the
state-of-the-art vision Transformers on a range of public image classification
and object detection benchmarks. In particular, our Focal Transformer models
with a moderate size of 51.1M and a larger size of 89.8M achieve 83.5 and 83.8
Top-1 accuracy, respectively, on ImageNet classification at 224x224 resolution.
Using Focal Transformers as the backbones, we obtain consistent and substantial
improvements over the current state-of-the-art Swin Transformers for 6
different object detection methods trained with standard 1x and 3x schedules.
Our largest Focal Transformer yields 58.7/58.9 box mAPs and 50.9/51.3 mask mAPs
on COCO mini-val/test-dev, and 55.4 mIoU on ADE20K for semantic segmentation,
creating new SoTA on three of the most challenging computer vision tasks. | [
"cs.CV",
"cs.AI",
"cs.LG"
] |
Designing a multi-layer optical system with designated optical
characteristics is an inverse design problem in which the resulting design is
determined by several discrete and continuous parameters. In particular, we
consider three design parameters to describe a multi-layer stack: Each layer's
dielectric material and thickness as well as the total number of layers. Such a
combination of both, discrete and continuous parameters is a challenging
optimization problem that often requires a computationally expensive search for
an optimal system design. Hence, most methods merely determine the optimal
thicknesses of the system's layers. To incorporate layer material and the total
number of layers as well, we propose a method that considers the stacking of
consecutive layers as parameterized actions in a Markov decision process. We
propose an exponentially transformed reward signal that eases policy
optimization and adapt a recent variant of Q-learning for inverse design
optimization. We demonstrate that our method outperforms human experts and a
naive reinforcement learning algorithm concerning the achieved optical
characteristics. Moreover, the learned Q-values contain information about the
optical properties of multi-layer optical systems, thereby allowing physical
interpretation or what-if analysis. | [
"cs.LG",
"cs.AI",
"physics.optics"
] |
Summarizing video content is an important task in many applications. This
task can be defined as the computation of the ordered list of actions present
in a video. Such a list could be extracted using action detection algorithms.
However, it is not necessary to determine the temporal boundaries of actions to
know their existence. Moreover, localizing precise boundaries usually requires
dense video analysis to be effective. In this work, we propose to directly
compute this ordered list by sparsely browsing the video and selecting one
frame per action instance, task known as action spotting in literature. To do
this, we propose ActionSpotter, a spotting algorithm that takes advantage of
Deep Reinforcement Learning to efficiently spot actions while adapting its
video browsing speed, without additional supervision. Experiments performed on
datasets THUMOS14 and ActivityNet show that our framework outperforms state of
the art detection methods. In particular, the spotting mean Average Precision
on THUMOS14 is significantly improved from 59.7% to 65.6% while skipping 23% of
video. | [
"cs.LG",
"cs.AI",
"cs.CV"
] |
In 2D image processing, some attempts decompose images into high and low
frequency components for describing edge and smooth parts respectively.
Similarly, the contour and flat area of 3D objects, such as the boundary and
seat area of a chair, describe different but also complementary geometries.
However, such investigation is lost in previous deep networks that understand
point clouds by directly treating all points or local patches equally. To solve
this problem, we propose Geometry-Disentangled Attention Network (GDANet).
GDANet introduces Geometry-Disentangle Module to dynamically disentangle point
clouds into the contour and flat part of 3D objects, respectively denoted by
sharp and gentle variation components. Then GDANet exploits Sharp-Gentle
Complementary Attention Module that regards the features from sharp and gentle
variation components as two holistic representations, and pays different
attentions to them while fusing them respectively with original point cloud
features. In this way, our method captures and refines the holistic and
complementary 3D geometric semantics from two distinct disentangled components
to supplement the local information. Extensive experiments on 3D object
classification and segmentation benchmarks demonstrate that GDANet achieves the
state-of-the-arts with fewer parameters. Code is released on
https://github.com/mutianxu/GDANet. | [
"cs.CV"
] |
Reading of mathematical expression or equation in the document images is very
challenging due to the large variability of mathematical symbols and
expressions. In this paper, we pose reading of mathematical equation as a task
of generation of the textual description which interprets the internal meaning
of this equation. Inspired by the natural image captioning problem in computer
vision, we present a mathematical equation description (MED) model, a novel
end-to-end trainable deep neural network based approach that learns to generate
a textual description for reading mathematical equation images. Our MED model
consists of a convolution neural network as an encoder that extracts features
of input mathematical equation images and a recurrent neural network with
attention mechanism which generates description related to the input
mathematical equation images. Due to the unavailability of mathematical
equation image data sets with their textual descriptions, we generate two data
sets for experimental purpose. To validate the effectiveness of our MED model,
we conduct a real-world experiment to see whether the students are able to
write equations by only reading or listening their textual descriptions or not.
Experiments conclude that the students are able to write most of the equations
correctly by reading their textual descriptions only. | [
"cs.CV"
] |
Although wireless capsule endoscopy is the preferred modality for diagnosis
and assessment of small bowel diseases, the poor camera resolution is a
substantial limitation for both subjective and automated diagnostics.
Enhanced-resolution endoscopy has shown to improve adenoma detection rate for
conventional endoscopy and is likely to do the same for capsule endoscopy. In
this work, we propose and quantitatively validate a novel framework to learn a
mapping from low-to-high resolution endoscopic images. We combine conditional
adversarial networks with a spatial attention block to improve the resolution
by up to factors of 8x, 10x, 12x, respectively. Quantitative and qualitative
studies performed demonstrate the superiority of EndoL2H over state-of-the-art
deep super-resolution methods DBPN, RCAN and SRGAN. MOS tests performed by 30
gastroenterologists qualitatively assess and confirm the clinical relevance of
the approach. EndoL2H is generally applicable to any endoscopic capsule system
and has the potential to improve diagnosis and better harness computational
approaches for polyp detection and characterization. Our code and trained
models are available at https://github.com/CapsuleEndoscope/EndoL2H. | [
"cs.CV",
"cs.LG",
"eess.IV"
] |
Combinatorial optimization problem (COP) over graphs is a fundamental
challenge in optimization. Reinforcement learning (RL) has recently emerged as
a new framework to tackle these problems and has demonstrated promising
results. However, most RL solutions employ a greedy manner to construct the
solution incrementally, thus inevitably pose unnecessary dependency on action
sequences and need a lot of problem-specific designs. We propose a general RL
framework that not only exhibits state-of-the-art empirical performance but
also generalizes to a variety class of COPs. Specifically, we define state as a
solution to a problem instance and action as a perturbation to this solution.
We utilize graph neural networks (GNN) to extract latent representations for
given problem instances for state-action encoding, and then apply deep
Q-learning to obtain a policy that gradually refines the solution by flipping
or swapping vertex labels. Experiments are conducted on Maximum $k$-Cut and
Traveling Salesman Problem and performance improvement is achieved against a
set of learning-based and heuristic baselines. | [
"cs.LG"
] |
This paper addresses the problem of decentralized spectrum sharing in
vehicle-to-everything (V2X) communication networks. The aim is to provide
resource-efficient coexistence of vehicle-to-infrastructure(V2I) and
vehicle-to-vehicle(V2V) links. A recent work on the topic proposes a
multi-agent reinforcement learning (MARL) approach based on deep Q-learning,
which leverages a fingerprint-based deep Q-network (DQN) architecture. This
work considers an extension of this framework by combining Double Q-learning
(via Double DQN) and transfer learning. The motivation behind is that Double
Q-learning can alleviate the problem of overestimation of the action values
present in conventional Q-learning, while transfer learning can leverage
knowledge acquired by an expert model to accelerate learning in the MARL
setting. The proposed algorithm is evaluated in a realistic V2X setting, with
synthetic data generated based on a geometry-based propagation model that
incorporates location-specific geographical descriptors of the simulated
environment(outlines of buildings, foliage, and vehicles). The advantages of
the proposed approach are demonstrated via numerical simulations. | [
"cs.LG",
"cs.IT",
"eess.SP",
"math.IT"
] |
Background: The human mind is multimodal. Yet most behavioral studies rely on
century-old measures such as task accuracy and latency. To create a better
understanding of human behavior and brain functionality, we should introduce
other measures and analyze behavior from various aspects. However, it is
technically complex and costly to design and implement the experiments that
record multiple measures. To address this issue, a platform that allows
synchronizing multiple measures from human behavior is needed. Method: This
paper introduces an opensource platform named OpenSync, which can be used to
synchronize multiple measures in neuroscience experiments. This platform helps
to automatically integrate, synchronize and record physiological measures
(e.g., electroencephalogram (EEG), galvanic skin response (GSR), eye-tracking,
body motion, etc.), user input response (e.g., from mouse, keyboard, joystick,
etc.), and task-related information (stimulus markers). In this paper, we
explain the structure and details of OpenSync, provide two case studies in
PsychoPy and Unity. Comparison with existing tools: Unlike proprietary systems
(e.g., iMotions), OpenSync is free and it can be used inside any opensource
experiment design software (e.g., PsychoPy, OpenSesame, Unity, etc.,
https://pypi.org/project/OpenSync/ and
https://github.com/moeinrazavi/OpenSync_Unity). Results: Our experimental
results show that the OpenSync platform is able to synchronize multiple
measures with microsecond resolution. | [
"cs.LG",
"eess.SP",
"q-bio.NC"
] |
Learning-based visual odometry and SLAM methods demonstrate a steady
improvement over past years. However, collecting ground truth poses to train
these methods is difficult and expensive. This could be resolved by training in
an unsupervised mode, but there is still a large gap between performance of
unsupervised and supervised methods. In this work, we focus on generating
synthetic data for deep learning-based visual odometry and SLAM methods that
take optical flow as an input. We produce training data in a form of optical
flow that corresponds to arbitrary camera movement between a real frame and a
virtual frame. For synthesizing data we use depth maps either produced by a
depth sensor or estimated from stereo pair. We train visual odometry model on
synthetic data and do not use ground truth poses hence this model can be
considered unsupervised. Also it can be classified as monocular as we do not
use depth maps on inference. We also propose a simple way to convert any visual
odometry model into a SLAM method based on frame matching and graph
optimization. We demonstrate that both the synthetically-trained visual
odometry model and the proposed SLAM method build upon this model yields
state-of-the-art results among unsupervised methods on KITTI dataset and shows
promising results on a challenging EuRoC dataset. | [
"cs.CV"
] |
Generative adversarial networks (GANs) can be trained to generate 3D image
data, which is useful for design optimisation. However, this conventionally
requires 3D training data, which is challenging to obtain. 2D imaging
techniques tend to be faster, higher resolution, better at phase identification
and more widely available. Here, we introduce a generative adversarial network
architecture, SliceGAN, which is able to synthesise high fidelity 3D datasets
using a single representative 2D image. This is especially relevant for the
task of material microstructure generation, as a cross-sectional micrograph can
contain sufficient information to statistically reconstruct 3D samples. Our
architecture implements the concept of uniform information density, which both
ensures that generated volumes are equally high quality at all points in space,
and that arbitrarily large volumes can be generated. SliceGAN has been
successfully trained on a diverse set of materials, demonstrating the
widespread applicability of this tool. The quality of generated micrographs is
shown through a statistical comparison of synthetic and real datasets of a
battery electrode in terms of key microstructural metrics. Finally, we find
that the generation time for a $10^8$ voxel volume is on the order of a few
seconds, yielding a path for future studies into high-throughput
microstructural optimisation. | [
"cs.CV",
"cs.LG"
] |
In this paper, we focus on designing effective method for fast and accurate
scene parsing. A common practice to improve the performance is to attain high
resolution feature maps with strong semantic representation. Two strategies are
widely used -- atrous convolutions and feature pyramid fusion, are either
computation intensive or ineffective. Inspired by the Optical Flow for motion
alignment between adjacent video frames, we propose a Flow Alignment Module
(FAM) to learn Semantic Flow between feature maps of adjacent levels, and
broadcast high-level features to high resolution features effectively and
efficiently. Furthermore, integrating our module to a common feature pyramid
structure exhibits superior performance over other real-time methods even on
light-weight backbone networks, such as ResNet-18. Extensive experiments are
conducted on several challenging datasets, including Cityscapes, PASCAL
Context, ADE20K and CamVid. Especially, our network is the first to achieve
80.4\% mIoU on Cityscapes with a frame rate of 26 FPS. The code is available at
\url{https://github.com/lxtGH/SFSegNets}. | [
"cs.CV",
"cs.RO"
] |
Low-rank approximation is an effective model compression technique to not
only reduce parameter storage requirements, but to also reduce computations.
For convolutional neural networks (CNNs), however, well-known low-rank
approximation methods, such as Tucker or CP decomposition, result in degraded
model accuracy because decomposed layers hinder training convergence. In this
paper, we propose a new training technique that finds a flat minimum in the
view of low-rank approximation without a decomposed structure during training.
By preserving the original model structure, 2-dimensional low-rank
approximation demanding lowering (such as im2col) is available in our proposed
scheme. We show that CNN models can be compressed by low-rank approximation
with much higher compression ratio than conventional training methods while
maintaining or even enhancing model accuracy. We also discuss various
2-dimensional low-rank approximation techniques for CNNs. | [
"cs.LG",
"stat.ML"
] |
Due to the sparsity of features, noise has proven to be a great inhibitor in
the classification of handwritten characters. To combat this, most techniques
perform denoising of the data before classification. In this paper, we
consolidate the approach by training an all-in-one model that is able to
classify even noisy characters. For classification, we progressively train a
classifier generative adversarial network on the characters from low to high
resolution. We show that by learning the features at each resolution
independently a trained model is able to accurately classify characters even in
the presence of noise. We experimentally demonstrate the effectiveness of our
approach by classifying noisy versions of MNIST, handwritten Bangla Numeral,
and Basic Character datasets. | [
"cs.CV",
"cs.IR"
] |
Denoising Score Matching with Annealed Langevin Sampling (DSM-ALS) has
recently found success in generative modeling. The approach works by first
training a neural network to estimate the score of a distribution, and then
using Langevin dynamics to sample from the data distribution assumed by the
score network. Despite the convincing visual quality of samples, this method
appears to perform worse than Generative Adversarial Networks (GANs) under the
Fr\'echet Inception Distance, a standard metric for generative models.
We show that this apparent gap vanishes when denoising the final Langevin
samples using the score network. In addition, we propose two improvements to
DSM-ALS: 1) Consistent Annealed Sampling as a more stable alternative to
Annealed Langevin Sampling, and 2) a hybrid training formulation, composed of
both Denoising Score Matching and adversarial objectives. By combining these
two techniques and exploring different network architectures, we elevate score
matching methods and obtain results competitive with state-of-the-art image
generation on CIFAR-10. | [
"cs.LG",
"cs.CV",
"stat.ML"
] |
Referring video object segmentation (RVOS) aims to segment video objects with
the guidance of natural language reference. Previous methods typically tackle
RVOS through directly grounding linguistic reference over the image lattice.
Such bottom-up strategy fails to explore object-level cues, easily leading to
inferior results. In this work, we instead put forward a two-stage, top-down
RVOS solution. First, an exhaustive set of object tracklets is constructed by
propagating object masks detected from several sampled frames to the entire
video. Second, a Transformer-based tracklet-language grounding module is
proposed, which models instance-level visual relations and cross-modal
interactions simultaneously and efficiently. Our model ranks first place on
CVPR2021 Referring Youtube-VOS challenge. | [
"cs.CV"
] |
Generating natural language descriptions for in-the-wild videos is a
challenging task. Most state-of-the-art methods for solving this problem borrow
existing deep convolutional neural network (CNN) architectures (AlexNet,
GoogLeNet) to extract a visual representation of the input video. However,
these deep CNN architectures are designed for single-label centered-positioned
object classification. While they generate strong semantic features, they have
no inherent structure allowing them to detect multiple objects of different
sizes and locations in the frame. Our paper tries to solve this problem by
integrating the base CNN into several fully convolutional neural networks
(FCNs) to form a multi-scale network that handles multiple receptive field
sizes in the original image. FCNs, previously applied to image segmentation,
can generate class heat-maps efficiently compared to sliding window mechanisms,
and can easily handle multiple scales. To further handle the ambiguity over
multiple objects and locations, we incorporate the Multiple Instance Learning
mechanism (MIL) to consider objects in different positions and at different
scales simultaneously. We integrate our multi-scale multi-instance architecture
with a sequence-to-sequence recurrent neural network to generate sentence
descriptions based on the visual representation. Ours is the first end-to-end
trainable architecture that is capable of multi-scale region processing.
Evaluation on a Youtube video dataset shows the advantage of our approach
compared to the original single-scale whole frame CNN model. Our flexible and
efficient architecture can potentially be extended to support other video
processing tasks. | [
"cs.CV"
] |
Commonly used human motion capture systems require intrusive attachment of
markers that are visually tracked with multiple cameras. In this work we
present an efficient and inexpensive solution to markerless motion capture
using only a few Kinect sensors. Unlike the previous work on 3d pose estimation
using a single depth camera, we relax constraints on the camera location and do
not assume a co-operative user. We apply recent image segmentation techniques
to depth images and use curriculum learning to train our system on purely
synthetic data. Our method accurately localizes body parts without requiring an
explicit shape model. The body joint locations are then recovered by combining
evidence from multiple views in real-time. We also introduce a dataset of ~6
million synthetic depth frames for pose estimation from multiple cameras and
exceed state-of-the-art results on the Berkeley MHAD dataset. | [
"cs.CV"
] |
The 2019 novel coronavirus (SARS-CoV-2) pandemic has resulted in more than a
million deaths, high morbidities, and economic distress worldwide. There is an
urgent need to identify medications that would treat and prevent novel diseases
like the 2019 coronavirus disease (COVID-19). Drug repurposing is a promising
strategy to discover new medical indications of the existing approved drugs due
to several advantages in terms of the costs, safety factors, and quick results
compared to new drug design and discovery. In this work, we explore
computational data-driven methods for drug repurposing and propose a dedicated
graph neural network (GNN) based drug repurposing model, called Dr-COVID.
Although we analyze the predicted drugs in detail for COVID-19, the model is
generic and can be used for any novel diseases. We construct a four-layered
heterogeneous graph to model the complex interactions between drugs, diseases,
genes, and anatomies. We pose drug repurposing as a link prediction problem.
Specifically, we design an encoder based on the scalable inceptive graph neural
network (SIGN) to generate embeddings for all the nodes in the four-layered
graph and propose a quadratic norm scorer as a decoder to predict treatment for
a disease. We provide a detailed analysis of the 150 potential drugs (such as
Dexamethasone, Ivermectin) predicted by Dr-COVID for COVID-19 from different
pharmacological classes (e.g., corticosteroids, antivirals, antiparasitic). Out
of these 150 drugs, 46 drugs are currently in clinical trials. Dr-COVID is
evaluated in terms of its prediction performance and its ability to rank the
known treatment drugs for diseases as high as possible. For a majority of the
diseases, Dr-COVID ranks the actual treatment drug in the top 15. | [
"cs.LG",
"q-bio.MN",
"q-bio.QM"
] |
Ranked data sets, where m judges/voters specify a preference ranking of n
objects/candidates, are increasingly prevalent in contexts such as political
elections, computer vision, recommender systems, and bioinformatics. The vote
counts for each ranking can be viewed as an n! data vector lying on the
permutahedron, which is a Cayley graph of the symmetric group with vertices
labeled by permutations and an edge when two permutations differ by an adjacent
transposition. Leveraging combinatorial representation theory and recent
progress in signal processing on graphs, we investigate a novel, scalable
transform method to interpret and exploit structure in ranked data. We
represent data on the permutahedron using an overcomplete dictionary of atoms,
each of which captures both smoothness information about the data (typically
the focus of spectral graph decomposition methods in graph signal processing)
and structural information about the data (typically the focus of symmetry
decomposition methods from representation theory). These atoms have a more
naturally interpretable structure than any known basis for signals on the
permutahedron, and they form a Parseval frame, ensuring beneficial numerical
properties such as energy preservation. We develop specialized algorithms and
open software that take advantage of the symmetry and structure of the
permutahedron to improve the scalability of the proposed method, making it more
applicable to the high-dimensional ranked data found in applications. | [
"stat.ML",
"cs.LG",
"eess.SP",
"math.RT"
] |
This paper raises an implicit manifold learning perspective in Generative
Adversarial Networks (GANs), by studying how the support of the learned
distribution, modelled as a submanifold $\mathcal{M}_{\theta}$, perfectly match
with $\mathcal{M}_{r}$, the support of the real data distribution. We show that
optimizing Jensen-Shannon divergence forces $\mathcal{M}_{\theta}$ to perfectly
match with $\mathcal{M}_{r}$, while optimizing Wasserstein distance does not.
On the other hand, by comparing the gradients of the Jensen-Shannon divergence
and the Wasserstein distances ($W_1$ and $W_2^2$) in their primal forms, we
conjecture that Wasserstein $W_2^2$ may enjoy desirable properties such as
reduced mode collapse. It is therefore interesting to design new distances that
inherit the best from both distances. | [
"stat.ML"
] |
Camera and 3D LiDAR sensors have become indispensable devices in modern
autonomous driving vehicles, where the camera provides the fine-grained
texture, color information in 2D space and LiDAR captures more precise and
farther-away distance measurements of the surrounding environments. The
complementary information from these two sensors makes the two-modality fusion
be a desired option. However, two major issues of the fusion between camera and
LiDAR hinder its performance, \ie, how to effectively fuse these two modalities
and how to precisely align them (suffering from the weak spatiotemporal
synchronization problem). In this paper, we propose a coarse-to-fine LiDAR and
camera fusion-based network (termed as LIF-Seg) for LiDAR segmentation. For the
first issue, unlike these previous works fusing the point cloud and image
information in a one-to-one manner, the proposed method fully utilizes the
contextual information of images and introduces a simple but effective
early-fusion strategy. Second, due to the weak spatiotemporal synchronization
problem, an offset rectification approach is designed to align these
two-modality features. The cooperation of these two components leads to the
success of the effective camera-LiDAR fusion. Experimental results on the
nuScenes dataset show the superiority of the proposed LIF-Seg over existing
methods with a large margin. Ablation studies and analyses demonstrate that our
proposed LIF-Seg can effectively tackle the weak spatiotemporal synchronization
problem. | [
"cs.CV"
] |
A number of psychological and physiological evidences suggest that early
visual attention works in a coarse-to-fine way, which lays a basis for the
reverse hierarchy theory (RHT). This theory states that attention propagates
from the top level of the visual hierarchy that processes gist and abstract
information of input, to the bottom level that processes local details.
Inspired by the theory, we develop a computational model for saliency detection
in images. First, the original image is downsampled to different scales to
constitute a pyramid. Then, saliency on each layer is obtained by image
super-resolution reconstruction from the layer above, which is defined as
unpredictability from this coarse-to-fine reconstruction. Finally, saliency on
each layer of the pyramid is fused into stochastic fixations through a
probabilistic model, where attention initiates from the top layer and
propagates downward through the pyramid. Extensive experiments on two standard
eye-tracking datasets show that the proposed method can achieve competitive
results with state-of-the-art models. | [
"cs.CV"
] |
Graphs as a type of data structure have recently attracted significant
attention. Representation learning of geometric graphs has achieved great
success in many fields including molecular, social, and financial networks. It
is natural to present proteins as graphs in which nodes represent the residues
and edges represent the pairwise interactions between residues. However, 3D
protein structures have rarely been studied as graphs directly. The challenges
include: 1) Proteins are complex macromolecules composed of thousands of atoms
making them much harder to model than micro-molecules. 2) Capturing the
long-range pairwise relations for protein structure modeling remains
under-explored. 3) Few studies have focused on learning the different
attributes of proteins together. To address the above challenges, we propose a
new graph neural network architecture to represent the proteins as 3D graphs
and predict both distance geometric graph representation and dihedral geometric
graph representation together. This gives a significant advantage because this
network opens a new path from the sequence to structure. We conducted extensive
experiments on four different datasets and demonstrated the effectiveness of
the proposed method. | [
"cs.LG",
"q-bio.BM",
"q-bio.QM"
] |
The recent explosion in applications of machine learning to satellite imagery
often rely on visible images and therefore suffer from a lack of data during
the night. The gap can be filled by employing available infra-red observations
to generate visible images. This work presents how deep learning can be applied
successfully to create those images by using U-Net based architectures. The
proposed methods show promising results, achieving a structural similarity
index (SSIM) up to 86\% on an independent test set and providing visually
convincing output images, generated from infra-red observations. | [
"cs.CV",
"cs.LG",
"eess.IV"
] |
Technologies to predict human actions are extremely important for
applications such as human robot cooperation and autonomous driving. However, a
majority of the existing algorithms focus on exploiting visual features of the
videos and do not consider the mining of relationships, which include spatial
relationships between human and scene elements as well as causal relationships
in temporal action sequences. In fact, human beings are good at using spatial
and causal relational reasoning mechanism to predict the actions of others.
Inspired by this idea, we proposed a Spatial and Causal Relationship based
Graph Reasoning Network (SCR-Graph), which can be used to predict human actions
by modeling the action-scene relationship, and causal relationship between
actions, in spatial and temporal dimensions respectively. Here, in spatial
dimension, a hierarchical graph attention module is designed by iteratively
aggregating the features of different kinds of scene elements in different
level. In temporal dimension, we designed a knowledge graph based causal
reasoning module and map the past actions to temporal causal features through
Diffusion RNN. Finally, we integrated the causality features into the
heterogeneous graph in the form of shadow node, and introduced a self-attention
module to determine the time when the knowledge graph information should be
activated. Extensive experimental results on the VIRAT datasets demonstrate the
favorable performance of the proposed framework. | [
"cs.CV",
"eess.IV"
] |
U-Net has been providing state-of-the-art performance in many medical image
segmentation problems. Many modifications have been proposed for U-Net, such as
attention U-Net, recurrent residual convolutional U-Net (R2-UNet), and U-Net
with residual blocks or blocks with dense connections. However, all these
modifications have an encoder-decoder structure with skip connections, and the
number of paths for information flow is limited. We propose LadderNet in this
paper, which can be viewed as a chain of multiple U-Nets. Instead of only one
pair of encoder branch and decoder branch in U-Net, a LadderNet has multiple
pairs of encoder-decoder branches, and has skip connections between every pair
of adjacent decoder and decoder branches in each level. Inspired by the success
of ResNet and R2-UNet, we use modified residual blocks where two convolutional
layers in one block share the same weights. A LadderNet has more paths for
information flow because of skip connections and residual blocks, and can be
viewed as an ensemble of Fully Convolutional Networks (FCN). The equivalence to
an ensemble of FCNs improves segmentation accuracy, while the shared weights
within each residual block reduce parameter number. Semantic segmentation is
essential for retinal disease detection. We tested LadderNet on two benchmark
datasets for blood vessel segmentation in retinal images, and achieved superior
performance over methods in the literature. The implementation is provided
\url{https://github.com/juntang-zhuang/LadderNet} | [
"cs.CV",
"eess.IV"
] |
Segmentation-based tracking has been actively studied in computer vision and
multimedia. Superpixel based object segmentation and tracking methods are
usually developed for this task. However, they independently perform feature
representation and learning of superpixels which may lead to sub-optimal
results. In this paper, we propose to utilize graph convolutional network (GCN)
model for superpixel based object tracking. The proposed model provides a
general end-to-end framework which integrates i) label linear prediction, and
ii) structure-aware feature information of each superpixel together to obtain
object segmentation and further improves the performance of tracking. The main
benefits of the proposed GCN method have two main aspects. First, it provides
an effective end-to-end way to exploit both spatial and temporal consistency
constraint for target object segmentation. Second, it utilizes a mixed graph
convolution module to learn a context-aware and discriminative feature for
superpixel representation and labeling. An effective algorithm has been
developed to optimize the proposed model. Extensive experiments on five
datasets demonstrate that our method obtains better performance against
existing alternative methods. | [
"cs.CV"
] |
Finding a suitable data representation for a specific task has been shown to
be crucial in many applications. The success of subspace clustering depends on
the assumption that the data can be separated into different subspaces.
However, this simple assumption does not always hold since the raw data might
not be separable into subspaces. To recover the ``clustering-friendly''
representation and facilitate the subsequent clustering, we propose a graph
filtering approach by which a smooth representation is achieved. Specifically,
it injects graph similarity into data features by applying a low-pass filter to
extract useful data representations for clustering. Extensive experiments on
image and document clustering datasets demonstrate that our method improves
upon state-of-the-art subspace clustering techniques. Especially, its
comparable performance with deep learning methods emphasizes the effectiveness
of the simple graph filtering scheme for many real-world applications. An
ablation study shows that graph filtering can remove noise, preserve structure
in the image, and increase the separability of classes. | [
"cs.CV",
"cs.AI",
"cs.LG"
] |
Supervised learning of time series data has been extensively studied for the
case of a categorical target variable. In some application domains, e.g.,
energy, environment and health monitoring, it occurs that the target variable
is numerical and the problem is known as time series extrinsic regression
(TSER). In the literature, some well-known time series classifiers have been
extended for TSER problems. As first benchmarking studies have focused on
predictive performance, very little attention has been given to
interpretability. To fill this gap, in this paper, we suggest an extension of a
Bayesian method for robust and interpretable feature construction and selection
in the context of TSER. Our approach exploits a relational way to tackle with
TSER: (i), we build various and simple representations of the time series which
are stored in a relational data scheme, then, (ii), a propositionalisation
technique (based on classical aggregation / selection functions from the
relational data field) is applied to build interpretable features from
secondary tables to "flatten" the data; and (iii), the constructed features are
filtered out through a Bayesian Maximum A Posteriori approach. The resulting
transformed data can be processed with various existing regressors.
Experimental validation on various benchmark data sets demonstrates the
benefits of the suggested approach. | [
"cs.LG"
] |
A semi-parametric, non-linear regression model in the presence of latent
variables is introduced. These latent variables can correspond to unmodeled
phenomena or unmeasured agents in a complex networked system. This new
formulation allows joint estimation of certain non-linearities in the system,
the direct interactions between measured variables, and the effects of
unmodeled elements on the observed system. The particular form of the model
adopted is justified, and learning is posed as a regularized empirical risk
minimization. This leads to classes of structured convex optimization problems
with a "sparse plus low-rank" flavor. Relations between the proposed model and
several common model paradigms, such as those of Robust Principal Component
Analysis (PCA) and Vector Autoregression (VAR), are established. Particularly
in the VAR setting, the low-rank contributions can come from broad trends
exhibited in the time series. Details of the algorithm for learning the model
are presented. Experiments demonstrate the performance of the model and the
estimation algorithm on simulated and real data. | [
"stat.ML"
] |
Bottom-up and top-down visual cues are two types of information that helps
the visual saliency models. These salient cues can be from spatial
distributions of the features (space-based saliency) or contextual /
task-dependent features (object based saliency). Saliency models generally
incorporate salient cues either in bottom-up or top-down norm separately. In
this work, we combine bottom-up and top-down cues from both space and object
based salient features on RGB-D data. In addition, we also investigated the
ability of various pre-trained convolutional neural networks for extracting
top-down saliency on color images based on the object dependent feature
activation. We demonstrate that combining salient features from color and dept
through bottom-up and top-down methods gives significant improvement on the
salient object detection with space based and object based salient cues. RGB-D
saliency integration framework yields promising results compared with the
several state-of-the-art-models. | [
"cs.CV"
] |
Controlling the internal representation space of a neural network is a
desirable feature because it allows to generate new data in a supervised
manner. In this paper we will show how this can be achieved while building a
low-dimensional mapping of the input stream, by deriving a generalized
algorithm starting from Self Organizing Maps (SOMs). SOMs are a kind of neural
network which can be trained with unsupervised learning to produce a
low-dimensional discretized mapping of the input space. They can be used for
the generation of new data through backward propagation of interpolations made
from the mapping grid. Unfortunately the final topology of the mapping space of
a SOM is not known before learning, so interpolating new data in a supervised
way is not an easy task. Here we will show a variation from the SOM algorithm
consisting in constraining the update of prototypes so that it is also a
function of the distance of its prototypes from extrinsically given targets in
the mapping space. We will demonstrate how such variants, that we will call
Supervised Topological Maps (STMs), allow for a supervised mapping where the
position of internal representations in the mapping space is determined by the
experimenter. Controlling the internal representation space in STMs reveals to
be an easier task than what is currently done using other algorithms such as
variational or adversarial autoencoders. | [
"cs.LG",
"cs.NE",
"stat.ML",
"68T07",
"I.5.1; I.5.3"
] |
Controllable Image Captioning (CIC) -- generating image descriptions
following designated control signals -- has received unprecedented attention
over the last few years. To emulate the human ability in controlling caption
generation, current CIC studies focus exclusively on control signals concerning
objective properties, such as contents of interest or descriptive patterns.
However, we argue that almost all existing objective control signals have
overlooked two indispensable characteristics of an ideal control signal: 1)
Event-compatible: all visual contents referred to in a single sentence should
be compatible with the described activity. 2) Sample-suitable: the control
signals should be suitable for a specific image sample. To this end, we propose
a new control signal for CIC: Verb-specific Semantic Roles (VSR). VSR consists
of a verb and some semantic roles, which represents a targeted activity and the
roles of entities involved in this activity. Given a designated VSR, we first
train a grounded semantic role labeling (GSRL) model to identify and ground all
entities for each role. Then, we propose a semantic structure planner (SSP) to
learn human-like descriptive semantic structures. Lastly, we use a role-shift
captioning model to generate the captions. Extensive experiments and ablations
demonstrate that our framework can achieve better controllability than several
strong baselines on two challenging CIC benchmarks. Besides, we can generate
multi-level diverse captions easily. The code is available at:
https://github.com/mad-red/VSR-guided-CIC. | [
"cs.CV",
"cs.AI",
"cs.CL",
"cs.MM"
] |
Contour information plays a vital role in salient object detection. However,
excessive false positives remain in predictions from existing contour-based
models due to insufficient contour-saliency fusion. In this work, we designed a
network for better edge quality in salient object detection. We proposed a
contour-saliency blending module to exchange information between contour and
saliency. We adopted recursive CNN to increase contour-saliency fusion while
keeping the total trainable parameters the same. Furthermore, we designed a
stage-wise feature extraction module to help the model pick up the most helpful
features from previous intermediate saliency predictions. Besides, we proposed
two new loss functions, namely Dual Confinement Loss and Confidence Loss, for
our model to generate better boundary predictions. Evaluation results on five
common benchmark datasets reveal that our model achieves competitive
state-of-the-art performance. | [
"cs.CV"
] |
We present a method for recognition of isolated Swedish Sign Language signs.
The method will be used in a game intended to help children training signing at
home, as a complement to training with a teacher. The target group is not
primarily deaf children, but children with language disorders. Using sign
language as a support in conversation has been shown to greatly stimulate the
speech development of such children. The signer is captured with an RGB-D
(Kinect) sensor, which has three advantages over a regular RGB camera. Firstly,
it allows complex backgrounds to be removed easily. We segment the hands and
face based on skin color and depth information. Secondly, it helps with the
resolution of hand over face occlusion. Thirdly, signs take place in 3D; some
aspects of the signs are defined by hand motion vertically to the image plane.
This motion can be estimated if the depth is observable. The 3D motion of the
hands relative to the torso are used as a cue together with the hand shape, and
HMMs trained with this input are used for classification. To obtain higher
robustness towards differences across signers, Fisher Linear Discriminant
Analysis is used to find the combinations of features that are most descriptive
for each sign, regardless of signer. Experiments show that the system can
distinguish signs from a challenging 94 word vocabulary with a precision of up
to 94% in the signer dependent case and up to 47% in the signer independent
case. | [
"cs.CV"
] |
Estimating the amount of electricity that can be produced by rooftop
photovoltaic systems is a time-consuming process that requires on-site
measurements, a difficult task to achieve on a large scale. In this paper, we
present an approach to estimate the solar potential of rooftops based on their
location and architectural characteristics, as well as the amount of solar
radiation they receive annually. Our technique uses computer vision to achieve
semantic segmentation of roof sections and roof objects on the one hand, and a
machine learning model based on structured building features to predict roof
pitch on the other hand. We then compute the azimuth and maximum number of
solar panels that can be installed on a rooftop with geometric approaches.
Finally, we compute precise shading masks and combine them with solar
irradiation data that enables us to estimate the yearly solar potential of a
rooftop. | [
"cs.CV",
"cs.LG"
] |
Deep image prior (DIP), which utilizes a deep convolutional network (ConvNet)
structure itself as an image prior, has attracted attentions in computer vision
and machine learning communities. It empirically shows the effectiveness of
ConvNet structure for various image restoration applications. However, why the
DIP works so well is still unknown, and why convolution operation is useful for
image reconstruction or enhancement is not very clear. In this study, we tackle
these questions. The proposed approach is dividing the convolution into
``delay-embedding'' and ``transformation (\ie encoder-decoder)'', and proposing
a simple, but essential, image/tensor modeling method which is closely related
to dynamical systems and self-similarity. The proposed method named as manifold
modeling in embedded space (MMES) is implemented by using a novel
denoising-auto-encoder in combination with multi-way delay-embedding transform.
In spite of its simplicity, the image/tensor completion, super-resolution,
deconvolution, and denoising results of MMES are quite similar even competitive
to DIP in our extensive experiments, and these results would help us for
reinterpreting/characterizing the DIP from a perspective of ``low-dimensional
patch-manifold prior''. | [
"cs.CV",
"cs.LG"
] |
High spectral dimensionality and the shortage of annotations make
hyperspectral image (HSI) classification a challenging problem. Recent studies
suggest that convolutional neural networks can learn discriminative spatial
features, which play a paramount role in HSI interpretation. However, most of
these methods ignore the distinctive spectral-spatial characteristic of
hyperspectral data. In addition, a large amount of unlabeled data remains an
unexploited gold mine for efficient data use. Therefore, we proposed an
integration of generative adversarial networks (GANs) and probabilistic
graphical models for HSI classification. Specifically, we used a
spectral-spatial generator and a discriminator to identify land cover
categories of hyperspectral cubes. Moreover, to take advantage of a large
amount of unlabeled data, we adopted a conditional random field to refine the
preliminary classification results generated by GANs. Experimental results
obtained using two commonly studied datasets demonstrate that the proposed
framework achieved encouraging classification accuracy using a small number of
data for training. | [
"cs.CV",
"cs.AI"
] |
We study the adversarial multi-armed bandit problem and create a completely
online algorithmic framework that is invariant under arbitrary translations and
scales of the arm losses. We study the expected performance of our algorithm
against a generic competition class, which makes it applicable for a wide
variety of problem scenarios. Our algorithm works from a universal prediction
perspective and the performance measure used is the expected regret against
arbitrary arm selection sequences, which is the difference between our losses
and a competing loss sequence. The competition class can be designed to include
fixed arm selections, switching bandits, contextual bandits, or any other
competition of interest. The sequences in the competition class are generally
determined by the specific application at hand and should be designed
accordingly. Our algorithm neither uses nor needs any preliminary information
about the loss sequences and is completely online. Its performance bounds are
the second order bounds in terms of sum of the squared losses, where any affine
transform of the losses has no effect on the normalized regret. | [
"cs.LG",
"stat.ML"
] |
We propose a method for generating video-realistic animations of real humans
under user control. In contrast to conventional human character rendering, we
do not require the availability of a production-quality photo-realistic 3D
model of the human, but instead rely on a video sequence in conjunction with a
(medium-quality) controllable 3D template model of the person. With that, our
approach significantly reduces production cost compared to conventional
rendering approaches based on production-quality 3D models, and can also be
used to realistically edit existing videos. Technically, this is achieved by
training a neural network that translates simple synthetic images of a human
character into realistic imagery. For training our networks, we first track the
3D motion of the person in the video using the template model, and subsequently
generate a synthetically rendered version of the video. These images are then
used to train a conditional generative adversarial network that translates
synthetic images of the 3D model into realistic imagery of the human. We
evaluate our method for the reenactment of another person that is tracked in
order to obtain the motion data, and show video results generated from
artist-designed skeleton motion. Our results outperform the state-of-the-art in
learning-based human image synthesis. Project page:
http://gvv.mpi-inf.mpg.de/projects/wxu/HumanReenactment/ | [
"cs.CV"
] |
We study compositional generalization, viz., the problem of zero-shot
generalization to novel compositions of concepts in a domain. Standard neural
networks fail to a large extent on compositional learning. We propose Tree
Stack Memory Units (Tree-SMU) to enable strong compositional generalization.
Tree-SMU is a recursive neural network with Stack Memory Units (\SMU s), a
novel memory augmented neural network whose memory has a differentiable stack
structure. Each SMU in the tree architecture learns to read from its stack and
to write to it by combining the stacks and states of its children through
gating. The stack helps capture long-range dependencies in the problem domain,
thereby enabling compositional generalization. Additionally, the stack also
preserves the ordering of each node's descendants, thereby retaining locality
on the tree. We demonstrate strong empirical results on two mathematical
reasoning benchmarks. We use four compositionality tests to assess the
generalization performance of Tree-SMU and show that it enables accurate
compositional generalization compared to strong baselines such as Transformers
and Tree-LSTMs. | [
"cs.LG",
"cs.NE",
"stat.ML"
] |
Accurate estimation of remaining useful life (RUL) of industrial equipment
can enable advanced maintenance schedules, increase equipment availability and
reduce operational costs. However, existing deep learning methods for RUL
prediction are not completely successful due to the following two reasons.
First, relying on a single objective function to estimate the RUL will limit
the learned representations and thus affect the prediction accuracy. Second,
while longer sequences are more informative for modelling the sensor dynamics
of equipment, existing methods are less effective to deal with very long
sequences, as they mainly focus on the latest information. To address these two
problems, we develop a novel attention-based sequence to sequence with
auxiliary task (ATS2S) model. In particular, our model jointly optimizes both
reconstruction loss to empower our model with predictive capabilities (by
predicting next input sequence given current input sequence) and RUL prediction
loss to minimize the difference between the predicted RUL and actual RUL.
Furthermore, to better handle longer sequence, we employ the attention
mechanism to focus on all the important input information during training
process. Finally, we propose a new dual-latent feature representation to
integrate the encoder features and decoder hidden states, to capture rich
semantic information in data. We conduct extensive experiments on four real
datasets to evaluate the efficacy of the proposed method. Experimental results
show that our proposed method can achieve superior performance over 13
state-of-the-art methods consistently. | [
"cs.LG",
"stat.ML"
] |
A main issue preventing the use of Convolutional Neural Networks (CNN) in end
user applications is the low level of transparency in the decision process.
Previous work on CNN interpretability has mostly focused either on localizing
the regions of the image that contribute to the result or on building an
external model that generates plausible explanations. However, the former does
not provide any semantic information and the latter does not guarantee the
faithfulness of the explanation. We propose an intermediate representation
composed of multiple Semantically Interpretable Activation Maps (SIAM)
indicating the presence of predefined attributes at different locations of the
image. These attribute maps are then linearly combined to produce the final
output. This gives the user insight into what the model has seen, where, and a
final output directly linked to this information in a comprehensive and
interpretable way. We test the method on the task of landscape scenicness
(aesthetic value) estimation, using an intermediate representation of 33
attributes from the SUN Attributes database. The results confirm that SIAM
makes it possible to understand what attributes in the image are contributing
to the final score and where they are located. Since it is based on learning
from multiple tasks and datasets, SIAM improve the explanability of the
prediction without additional annotation efforts or computational overhead at
inference time, while keeping good performances on both the final and
intermediate tasks. | [
"cs.CV"
] |
Zero-Shot Learning (ZSL) is an emerging research that aims to solve the
classification problems with very few training data. The present works on ZSL
mainly focus on the mapping of learning semantic space to visual space. It
encounters many challenges that obstruct the progress of ZSL research. First,
the representation of the semantic feature is inadequate to represent all
features of the categories. Second, the domain drift problem still exists
during the transfer from semantic space to visual space. In this paper, we
introduce knowledge sharing (KS) to enrich the representation of semantic
features. Based on KS, we apply a generative adversarial network to generate
pseudo visual features from semantic features that are very close to the real
visual features. Abundant experimental results from two benchmark datasets of
ZSL show that the proposed approach has a consistent improvement. | [
"cs.CV",
"cs.AI"
] |
Lidar based 3D object detection is inevitable for autonomous driving, because
it directly links to environmental understanding and therefore builds the base
for prediction and motion planning. The capacity of inferencing highly sparse
3D data in real-time is an ill-posed problem for lots of other application
areas besides automated vehicles, e.g. augmented reality, personal robotics or
industrial automation. We introduce Complex-YOLO, a state of the art real-time
3D object detection network on point clouds only. In this work, we describe a
network that expands YOLOv2, a fast 2D standard object detector for RGB images,
by a specific complex regression strategy to estimate multi-class 3D boxes in
Cartesian space. Thus, we propose a specific Euler-Region-Proposal Network
(E-RPN) to estimate the pose of the object by adding an imaginary and a real
fraction to the regression network. This ends up in a closed complex space and
avoids singularities, which occur by single angle estimations. The E-RPN
supports to generalize well during training. Our experiments on the KITTI
benchmark suite show that we outperform current leading methods for 3D object
detection specifically in terms of efficiency. We achieve state of the art
results for cars, pedestrians and cyclists by being more than five times faster
than the fastest competitor. Further, our model is capable of estimating all
eight KITTI-classes, including Vans, Trucks or sitting pedestrians
simultaneously with high accuracy. | [
"cs.CV"
] |
Dynamic graph modeling has recently attracted much attention due to its
extensive applications in many real-world scenarios, such as recommendation
systems, financial transactions, and social networks. Although many works have
been proposed for dynamic graph modeling in recent years, effective and
scalable models are yet to be developed. In this paper, we propose a novel
graph neural network approach, called TCL, which deals with the
dynamically-evolving graph in a continuous-time fashion and enables effective
dynamic node representation learning that captures both the temporal and
topology information. Technically, our model contains three novel aspects.
First, we generalize the vanilla Transformer to temporal graph learning
scenarios and design a graph-topology-aware transformer. Secondly, on top of
the proposed graph transformer, we introduce a two-stream encoder that
separately extracts representations from temporal neighborhoods associated with
the two interaction nodes and then utilizes a co-attentional transformer to
model inter-dependencies at a semantic level. Lastly, we are inspired by the
recently developed contrastive learning and propose to optimize our model by
maximizing mutual information (MI) between the predictive representations of
two future interaction nodes. Benefiting from this, our dynamic representations
can preserve high-level (or global) semantics about interactions and thus is
robust to noisy interactions. To the best of our knowledge, this is the first
attempt to apply contrastive learning to representation learning on dynamic
graphs. We evaluate our model on four benchmark datasets for interaction
prediction and experiment results demonstrate the superiority of our model. | [
"cs.LG",
"cs.AI"
] |
We present a first attempt to elucidate a theoretical and empirical approach
to design the reward provided by a natural language environment to some
structure learning agent. To this end, we revisit the Information Theory of
unsupervised induction of phrase-structure grammars to characterize the
behavior of simulated actions modeled as set-valued random variables (random
sets of linguistic samples) constituting semantic structures. Our results
showed empirical evidence of that simulated semantic structures (Open
Information Extraction triplets) can be distinguished from randomly constructed
ones by observing the Mutual Information among their constituents. This
suggests the possibility of rewarding structure learning agents without using
pretrained structural analyzers (oracle actors/experts). | [
"cs.LG",
"cs.AI",
"cs.CL",
"cs.IT",
"math.IT",
"stat.ML"
] |
A bipartite network is a graph structure where nodes are from two distinct
domains and only inter-domain interactions exist as edges. A large number of
network embedding methods exist to learn vectorial node representations from
general graphs with both homogeneous and heterogeneous node and edge types,
including some that can specifically model the distinct properties of bipartite
networks. However, these methods are inadequate to model multiplex bipartite
networks (e.g., in e-commerce), that have multiple types of interactions (e.g.,
click, inquiry, and buy) and node attributes. Most real-world multiplex
bipartite networks are also sparse and have imbalanced node distributions that
are challenging to model. In this paper, we develop an unsupervised Dual
HyperGraph Convolutional Network (DualHGCN) model that scalably transforms the
multiplex bipartite network into two sets of homogeneous hypergraphs and uses
spectral hypergraph convolutional operators, along with intra- and
inter-message passing strategies to promote information exchange within and
across domains, to learn effective node embedding. We benchmark DualHGCN using
four real-world datasets on link prediction and node classification tasks. Our
extensive experiments demonstrate that DualHGCN significantly outperforms
state-of-the-art methods, and is robust to varying sparsity levels and
imbalanced node distributions. | [
"cs.LG",
"cs.SI"
] |
Recent years have witnessed the popularity and success of graph neural
networks (GNN) in various scenarios. To obtain data-specific GNN architectures,
researchers turn to neural architecture search (NAS), which has made impressive
success in discovering effective architectures in convolutional neural
networks. However, it is non-trivial to apply NAS approaches to GNN due to
challenges in search space design and the expensive searching cost of existing
NAS methods. In this work, to obtain the data-specific GNN architectures and
address the computational challenges facing by NAS approaches, we propose a
framework, which tries to Search to Aggregate NEighborhood (SANE), to
automatically design data-specific GNN architectures. By designing a novel and
expressive search space, we propose a differentiable search algorithm, which is
more efficient than previous reinforcement learning based methods. Experimental
results on four tasks and seven real-world datasets demonstrate the superiority
of SANE compared to existing GNN models and NAS approaches in terms of
effectiveness and efficiency. (Code is available at:
https://github.com/AutoML-4Paradigm/SANE). | [
"cs.LG"
] |