Ashish Bora , Sugato Basu , Joydeep Ghosh Subjects : Neural and Evolutionary Computing (cs.NE) ; Learning (cs.LG)
Many time series are generated by a set of entities that interact with one
another over time. This paper introduces a broad, flexible framework to learn
from multiple inter-dependent time series generated by such entities. Our
framework explicitly models the entities and their interactions through time.
It achieves this by building on the capabilities of Recurrent Neural Networks,
while also offering several ways to incorporate domain knowledge/constraints
into the model architecture. The capabilities of our approach are showcased
through an application to weather prediction, which shows gains over strong
baselines.
Comments: Pre-print. The final publication is available at Springer via this http URL
Subjects:
Neural and Evolutionary Computing (cs.NE)
; Learning (cs.LG)
In an attempt to solve the lengthy training times of neural networks, we
proposed Parallel Circuits (PCs), a biologically inspired architecture.
Previous work has shown that this approach fails to maintain generalization
performance in spite of achieving sharp speed gains. To address this issue, and
motivated by the way Dropout prevents node co-adaption, in this paper, we
suggest an improvement by extending Dropout to the PC architecture. The paper
provides multiple insights into this combination, including a variety of fusion
approaches. Experiments show promising results in which improved error rates
are achieved in most cases, whilst maintaining the speed advantage of the PC
approach.
Comments: This paper was accepted and presented at Computing with Spikes NIPS 2016 Workshop, Barcelona, Spain, December 2016
Subjects:
Neural and Evolutionary Computing (cs.NE)
; Machine Learning (stat.ML)
A dynamic Boltzmann machine (DyBM) has been proposed as a model of a spiking
neural network, and its learning rule of maximizing the log-likelihood of given
time-series has been shown to exhibit key properties of spike-timing dependent
plasticity (STDP), which had been postulated and experimentally confirmed in
the field of neuroscience as a learning rule that refines the Hebbian rule.
Here, we relax some of the constraints in the DyBM in a way that it becomes
more suitable for computation and learning. We show that learning the DyBM can
be considered as logistic regression for binary-valued time-series. We also
show how the DyBM can learn real-valued data in the form of a Gaussian DyBM and
discuss its relation to the vector autoregressive (VAR) model. The Gaussian
DyBM extends the VAR by using additional explanatory variables, which
correspond to the eligibility traces of the DyBM and capture long term
dependency of the time-series. Numerical experiments show that the Gaussian
DyBM significantly improves the predictive accuracy over VAR.
Franck Dernoncourt , Ji Young Lee , Peter Szolovits Subjects : Computation and Language (cs.CL) ; Artificial Intelligence (cs.AI); Neural and Evolutionary Computing (cs.NE); Machine Learning (stat.ML)
Existing models based on artificial neural networks (ANNs) for sentence
classification often do not incorporate the context in which sentences appear,
and classify sentences individually. However, traditional sentence
classification approaches have been shown to greatly benefit from jointly
classifying subsequent sentences, such as with conditional random fields. In
this work, we present an ANN architecture that combines the effectiveness of
typical ANN models to classify sentences in isolation, with the strength of
structured prediction. Our model achieves state-of-the-art results on two
different datasets for sequential sentence classification in medical abstracts.
Comments: 9 pages, 4 figures
Subjects:
Learning (cs.LG)
; Neural and Evolutionary Computing (cs.NE); Machine Learning (stat.ML)
We present a method for implementing an Efficient Unitary Neural Network
(EUNN) whose computational complexity is merely (mathcal{O}(1)) per parameter
and has full tunability, from spanning part of unitary space to all of it. We
apply the EUNN in Recurrent Neural Networks, and test its performance on the
standard copying task and the MNIST digit recognition benchmark, finding that
it significantly outperforms a non-unitary RNN, an LSTM network, an exclusively
partial space URNN and a projective URNN with comparable parameter numbers.
Comments: submitted to CVPR 2017
Subjects:
Computer Vision and Pattern Recognition (cs.CV)
We introduce the concept of a Visual Compiler that generates a scene specific
pedestrian detector and pose estimator without any pedestrian observations.
Given a single image and auxiliary scene information in the form of camera
parameters and geometric layout of the scene, the Visual Compiler first infers
geometrically and photometrically accurate images of humans in that scene
through the use of computer graphics rendering. Using these renders we learn a
scene-and-region specific spatially-varying fully convolutional neural network,
for simultaneous detection, pose estimation and segmentation of pedestrians. We
demonstrate that when real human annotated data is scarce or non-existent, our
data generation strategy can provide an excellent solution for bootstrapping
human detection and pose estimation. Experimental results show that our
approach outperforms off-the-shelf state-of-the-art pedestrian detectors and
pose estimators that are trained on real data.
CSVideoNet: A Recurrent Convolutional Neural Network for Compressive Sensing Video Reconstruction
Comments: 10 pages, 6 pages, 2 tables
Subjects:
Computer Vision and Pattern Recognition (cs.CV)
; Learning (cs.LG)
In this paper, we develop a deep neural network architecture called
“CSVideoNet” that can learn visual representations from random measurements for
compressive sensing (CS) video reconstruction. CSVideoNet is an end-to-end
trainable and non-iterative model that combines convolutional neural networks
(CNNs) with a recurrent neural networks (RNN) to facilitate video
reconstruction by leveraging temporal-spatial features. The proposed network
can accept random measurements with a multi-level compression ratio (CR). The
lightly and aggressively compressed measurements offer background information
and object details, respectively. This is similar to the variable bit rate
techniques widely used in conventional video coding approaches. The RNN
employed by CSVideoNet can leverage temporal coherence that exists in adjacent
video frames to extrapolate motion features and merge them with spatial visual
features extracted by the CNNs to further enhance reconstruction quality,
especially at high CRs. We test our CSVideoNet on the UCF-101 dataset.
Experimental results show that CSVideoNet outperforms the existing video CS
reconstruction approaches. The results demonstrate that our method can preserve
relatively excellent visual details from original videos even at a 100x CR,
which is difficult to realize with the reference approaches. Also, the
non-iterative nature of CSVideoNet results in an decrease in runtime by three
orders of magnitude over iterative reconstruction algorithms. Furthermore,
CSVideoNet can enhance the CR of CS cameras beyond the limitation of
conventional approaches, ensuring a reduction in bandwidth for data
transmission. These benefits are especially favorable to high-frame-rate video
applications.
SceneNet RGB-D: 5M Photorealistic Images of Synthetic Indoor Trajectories with Ground Truth
John McCormac , Ankur Handa , Stefan Leutenegger , Andrew J. Davison Subjects : Computer Vision and Pattern Recognition (cs.CV)
We introduce SceneNet RGB-D, expanding the previous work of SceneNet to
enable large scale photorealistic rendering of indoor scene trajectories. It
provides pixel-perfect ground truth for scene understanding problems such as
semantic segmentation, instance segmentation, and object detection, and also
for geometric computer vision problems such as optical flow, depth estimation,
camera pose estimation, and 3D reconstruction. Random sampling permits
virtually unlimited scene configurations, and here we provide a set of 5M
rendered RGB-D images from over 15K trajectories in synthetic layouts with
random but physically simulated object poses. Each layout also has random
lighting, camera trajectories, and textures. The scale of this dataset is well
suited for pre-training data-driven computer vision techniques from scratch
with RGB-D inputs, which previously has been limited by relatively small
labelled datasets in NYUv2 and SUN RGB-D. It also provides a basis for
investigating 3D scene labelling tasks by providing perfect camera poses and
depth data as proxy for a SLAM system. We host the dataset at
this http URLThomas Nestmeyer , Peter V. Gehler Subjects : Computer Vision and Pattern Recognition (cs.CV)
Separation of an input image into its reflectance and shading layers poses a
challenge for learning approaches because no large corpus of precise and
realistic ground truth decompositions exists. The Intrinsic Images in the Wild
dataset (IIW) provides a sparse set of relative human reflectance judgments,
which serves as a standard benchmark for intrinsic images. This dataset led to
an increase in methods that learn statistical dependencies between the images
and their reflectance layer. Although learning plays a role in pushing
state-of-the-art performance, we show that a standard signal processing
technique achieves performance on par with recent developments. We propose a
loss function that enables learning dense reflectance predictions with a CNN.
Our results show a simple pixel-wise decision, without any context or prior
knowledge, is sufficient to provide a strong baseline on IIW. This sets a
competitive bar and we find that only two approaches surpass this result. We
then develop a joint bilateral filtering method that implements strong prior
knowledge about reflectance constancy. This filtering operation can be applied
to any intrinsic image algorithm and we improve several previous results
achieving a new state-of-the-art on IIW. Our findings suggest that the effect
of learning-based approaches may be over-estimated and that it is still the use
of explicit prior knowledge that drives performance on intrinsic image
decompositions.
Adrian K. Davison , Cliff Lansley , Choon Ching Ng , Kevin Tan , Moi Hoon Yap Subjects : Computer Vision and Pattern Recognition (cs.CV)
Micro-facial expressions are regarded as an important human behavioural event
that can highlight emotional deception. Spotting these movements is difficult
for humans and machines, however research into using computer vision to detect
subtle facial expressions is growing in popularity. This paper proposes an
individualised baseline micro-movement detection method using 3D Histogram of
Oriented Gradients (3D HOG) temporal difference method. We define a face
template consisting of 26 regions based on the Facial Action Coding System
(FACS). We extract the temporal features of each region using 3D HOG. Then, we
use Chi-square distance to find subtle facial motion in the local regions.
Finally, an automatic peak detector is used to detect micro-movements above the
newly proposed adaptive baseline threshold. The performance is validated on two
FACS coded datasets: SAMM and CASME II. This objective method focuses on the
movement of the 26 face regions. When comparing with the ground truth, the best
result was an AUC of 0.7512 and 0.7261 on SAMM and CASME II, respectively. The
results show that 3D HOG outperformed for micro-movement detection, compared to
state-of-the-art feature representations: Local Binary Patterns in Three
Orthogonal Planes and Histograms of Oriented Optical Flow.
Alexander Hewer , Stefanie Wuhrer , Ingmar Steiner , Korin Richmond Subjects : Computer Vision and Pattern Recognition (cs.CV)
We present a multilinear statistical model of the human tongue that captures
anatomical and tongue pose related shape variations separately. The model was
derived from 3D magnetic resonance imaging data of 11 speakers sustaining
speech related vocal tract configurations. The extraction was performed by
using a minimally supervised method that uses as basis an image segmentation
approach and a template fitting technique. Furthermore, it uses image denoising
to deal with possibly corrupt data, palate surface information reconstruction
to handle palatal tongue contacts, and a bootstrap strategy to refine the
obtained shapes. Our experiments concluded that limiting the degrees of freedom
for the anatomical and speech related variations to 5 and 4 respectively
produces a model that can reliably register unknown data while avoiding
overfitting effects.
Development of a Real-time Colorectal Tumor Classification System for Narrow-band Imaging zoom-videoendoscopy
Comments: 9 pages, 8 figures
Subjects:
Computer Vision and Pattern Recognition (cs.CV)
Colorectal endoscopy is important for the early detection and treatment of
colorectal cancer and is used worldwide. A computer-aided diagnosis (CAD)
system that provides an objective measure to endoscopists during colorectal
endoscopic examinations would be of great value. In this study, we describe a
newly developed CAD system that provides real-time objective measures. Our
system captures the video stream from an endoscopic system and transfers it to
a desktop computer. The captured video stream is then classified by a
pretrained classifier and the results are displayed on a monitor. The
experimental results show that our developed system works efficiently in actual
endoscopic examinations and is medically significant.
Naushad Ansari , Anubha Gupta , Rahul Duggal Subjects : Computer Vision and Pattern Recognition (cs.CV)
Image-matched nonseparable wavelets can find potential use in many
applications including image classification, segmen- tation, compressive
sensing, etc. This paper proposes a novel design methodology that utilizes
convolutional neural net- work (CNN) to design two-channel non-separable
wavelet matched to a given image. The design is proposed on quin- cunx lattice.
The loss function of the convolutional neural network is setup with total
squared error between the given input image to CNN and the reconstructed image
at the output of CNN, leading to perfect reconstruction at the end of train-
ing. Simulation results have been shown on some standard images.
Or Litany , Tal Remez , Alex Bronstein Subjects : Computer Vision and Pattern Recognition (cs.CV) ; Graphics (cs.GR)
With the development of range sensors such as LIDAR and time-of-flight
cameras, 3D point cloud scans have become ubiquitous in computer vision
applications, the most prominent ones being gesture recognition and autonomous
driving. Parsimony-based algorithms have shown great success on images and
videos where data points are sampled on a regular Cartesian grid. We propose an
adaptation of these techniques to irregularly sampled signals by using
continuous dictionaries. We present an example application in the form of point
cloud denoising.
Recurrent Image Captioner: Describing Images with Spatial-Invariant Transformation and Attention Filtering
Hao Liu , Yang Yang , Fumin Shen , Lixin Duan , Heng Tao Shen Subjects : Computer Vision and Pattern Recognition (cs.CV) ; Computation and Language (cs.CL)
Along with the prosperity of recurrent neural network in modelling sequential
data and the power of attention mechanism in automatically identify salient
information, image captioning, a.k.a., image description, has been remarkably
advanced in recent years. Nonetheless, most existing paradigms may suffer from
the deficiency of invariance to images with different scaling, rotation, etc.;
and effective integration of standalone attention to form a holistic end-to-end
system. In this paper, we propose a novel image captioning architecture, termed
Recurrent Image Captioner ( extbf{RIC}), which allows visual encoder and
language decoder to coherently cooperate in a recurrent manner. Specifically,
we first equip CNN-based visual encoder with a differentiable layer to enable
spatially invariant transformation of visual signals. Moreover, we deploy an
attention filter module (differentiable) between encoder and decoder to
dynamically determine salient visual parts. We also employ bidirectional LSTM
to preprocess sentences for generating better textual representations. Besides,
we propose to exploit variational inference to optimize the whole architecture.
Extensive experimental results on three benchmark datasets (i.e., Flickr8k,
Flickr30k and MS COCO) demonstrate the superiority of our proposed architecture
as compared to most of the state-of-the-art methods.
Anh Tuan Tran , Tal Hassner , Iacopo Masi , Gerard Medioni Subjects : Computer Vision and Pattern Recognition (cs.CV)
The 3D shapes of faces are well known to be discriminative. Yet despite this,
they are rarely used for face recognition and always under controlled viewing
conditions. We claim that this is a symptom of a serious but often overlooked
problem with existing methods for single view 3D face reconstruction: when
applied “in the wild”, their 3D estimates are either unstable and change for
different photos of the same subject or they are over-regularized and generic.
In response, we describe a robust method for regressing discriminative 3D
morphable face models (3DMM). We use a convolutional neural network (CNN) to
regress 3DMM shape and texture parameters directly from an input photo. We
overcome the shortage of training data required for this purpose by offering a
method for generating huge numbers of labeled examples. The 3D estimates
produced by our CNN surpass state of the art accuracy on the MICC data set.
Coupled with a 3D-3D face matching pipeline, we show the first competitive face
recognition results on the LFW, YTF and IJB-A benchmarks using 3D face shapes
as representations, rather than the opaque deep feature vectors used by other
modern systems.
Vivek Krishnan , Deva Ramanan Subjects : Computer Vision and Pattern Recognition (cs.CV)
We consider the task of visual net surgery, in which a CNN can be
reconfigured without extra data to recognize novel concepts that may be omitted
from the training set. While most prior work make use of linguistic cues for
such “zero-shot” learning, we do so by using a pictorial language
representation of the training set, implicitly learned by a CNN, to generalize
to new classes. To this end, we introduce a set of visualization techniques
that better reveal the activation patterns and relations between groups of CNN
filters. We next demonstrate that knowledge of pictorial languages can be used
to rewire certain CNN neurons into a part model, which we call a pictorial
language classifier. We demonstrate the robustness of simple PLCs by applying
them in a weakly supervised manner: labeling unlabeled concepts for visual
classes present in the training data. Specifically we show that a PLC built on
top of a CNN trained for ImageNet classification can localize humans in Graz-02
and determine the pose of birds in PASCAL-VOC without extra labeled data or
additional training. We then apply PLCs in an interactive zero-shot manner,
demonstrating that pictorial languages are expressive enough to detect a set of
visual classes in MS-COCO that never appear in the ImageNet training set.
Fahad Shahbaz Khan , Joost van de Weijer , Rao Muhammad Anwer , Andrew D. Bagdanov , Michael Felsberg , Jorma Laaksonen Subjects : Computer Vision and Pattern Recognition (cs.CV)
Most approaches to human attribute and action recognition in still images are
based on image representation in which multi-scale local features are pooled
across scale into a single, scale-invariant encoding. Both in bag-of-words and
the recently popular representations based on convolutional neural networks,
local features are computed at multiple scales. However, these multi-scale
convolutional features are pooled into a single scale-invariant representation.
We argue that entirely scale-invariant image representations are sub-optimal
and investigate approaches to scale coding within a Bag of Deep Features
framework.
Our approach encodes multi-scale information explicitly during the image
encoding stage. We propose two strategies to encode multi-scale information
explicitly in the final image representation. We validate our two scale coding
techniques on five datasets: Willow, PASCAL VOC 2010, PASCAL VOC 2012,
Stanford-40 and Human Attributes (HAT-27). On all datasets, the proposed scale
coding approaches outperform both the scale-invariant method and the standard
deep features of the same network. Further, combining our scale coding
approaches with standard deep features leads to consistent improvement over the
state-of-the-art.
Comments: 9 pages, 9 figures, supplementary material added as ancillary file
Subjects:
Computer Vision and Pattern Recognition (cs.CV)
In this paper, we present a novel non-parametric clustering technique, which
is based on an iterative algorithm that peels off layers of points around the
clusters. Our technique is based on the notion that each latent cluster is
comprised of layers that surround its core, where the external layers, or
border points, implicitly separate the clusters. Analyzing the K-nearest
neighbors of the points makes it possible to identify the border points and
associate them with points of inner layers. Our clustering algorithm
iteratively identifies border points, peels them, and separates the latent
clusters. We show that the peeling process adapts to the local density and
successfully separates adjacent clusters. A notable quality of the
Border-Peeling algorithm is that it does not require any parameter tuning in
order to outperform state-of-the-art finely-tuned non-parametric clustering
methods, including Mean-Shift and DBSCAN. We further assess our technique on
high-dimensional datasets that vary in size and characteristics. In particular,
we analyze the space of deep features that were trained by a convolutional
neural network.
Giuseppe Airò Farulla , Nadir Murru , Rosaria Rossini Subjects : Computer Vision and Pattern Recognition (cs.CV)
The problem of correctly segmenting touching characters is an hard task to
solve and it is of major relevance in pattern recognition. In the recent years,
many methods and algorithms have been proposed; still, a definitive solution is
far from being found. In this paper, we propose a novel method based on fuzzy
logic. The proposed method combines in a novel way three features for
segmenting touching characters that have been already proposed in other studies
but have been exploited only singularly so far. The proposed strategy is based
on a 3–input/1–output fuzzy inference system with fuzzy rules specifically
optimized for segmenting touching characters in the case of Latin printed and
handwritten characters. The system performances are illustrated and supported
by numerical examples showing that our approach can achieve a reasonable good
overall accuracy in segmenting characters even on tricky conditions of touching
characters. Moreover, numerical results suggest that the method can be applied
to many different datasets of characters by means of a convenient tuning of the
fuzzy sets and rules.
Michal Yarom , Michal Irani Subjects : Computer Vision and Pattern Recognition (cs.CV)
The ability to detect similar actions across videos can be very useful for
real-world applications in many fields. However, this task is still challenging
for existing systems, since videos that present the same action, can be taken
from significantly different viewing directions, performed by different actors
and backgrounds and under various video qualities. Video descriptors play a
significant role in these systems. In this work we propose the
“temporal-needle” descriptor which captures the dynamic behavior, while being
invariant to viewpoint and appearance. The descriptor is computed using multi
temporal scales of the video and by computing self-similarity for every patch
through time in every temporal scale. The descriptor is computed for every
pixel in the video. However, to find similar actions across videos, we consider
only a small subset of the descriptors – the statistical significant
descriptors. This allow us to find good correspondences across videos more
efficiently. Using the descriptor, we were able to detect the same behavior
across videos in a variety of scenarios. We demonstrate the use of the
descriptor in tasks such as temporal and spatial alignment, action detection
and even show its potential in unsupervised video clustering into categories.
In this work we handled only videos taken with stationary cameras, but the
descriptor can be extended to handle moving camera as well.
Kenneth Marino , Ruslan Salakhutdinov , Abhinav Gupta Subjects : Computer Vision and Pattern Recognition (cs.CV)
Humans have the remarkable capability to learn a large variety of visual
concepts, often with very few examples, whereas current state-of-the-art vision
algorithms require hundreds or thousands of examples per category and struggle
with ambiguity. One characteristic that sets humans apart is our ability to
acquire knowledge about the world and reason using this knowledge. This paper
investigates the use of structured prior knowledge in the form of knowledge
graphs and shows that using this knowledge improves performance on image
classification. Specifically, we introduce the Graph Search Neural Network as a
way of efficiently incorporating large knowledge graphs into a fully end-to-end
learning system. We show in a number of experiments that our method outperforms
baselines for multi-label classification, even under low data and few-shot
settings.
Lukas Balles , Javier Romero , Philipp Hennig Subjects : Learning (cs.LG) ; Computer Vision and Pattern Recognition (cs.CV); Machine Learning (stat.ML)
Mini-batch stochastic gradient descent and variants thereof have become
standard for large-scale empirical risk minimization like the training of
neural networks. These methods are usually used with a constant batch size
chosen by simple empirical inspection. The batch size significantly influences
the behavior of the stochastic optimization algorithm, though, since it
determines the variance of the gradient estimates. This variance also changes
over the optimization process; when using a constant batch size, stability and
convergence is thus often enforced by means of a (manually tuned) decreasing
learning rate schedule. We propose a practical method for dynamic batch size
adaptation. It estimates the variance of the stochastic gradients and adapts
the batch size to decrease the variance proportionally to the value of the
objective function, removing the need for the aforementioned learning rate
decrease. In contrast to recent related work, our algorithm couples the batch
size to the learning rate, directly reflecting the known relationship between
the two. On three image classification benchmarks, our batch size adaptation
yields faster optimization convergence, while simultaneously simplifying
learning rate tuning. A TensorFlow implementation is available.
Comments: Published In Proceedings of the 17th International Society for Music Information Retrieval Conference (2016)
Subjects:
Learning (cs.LG)
; Computer Vision and Pattern Recognition (cs.CV)
This paper addresses the matching of short music audio snippets to the
corresponding pixel location in images of sheet music. A system is presented
that simultaneously learns to read notes, listens to music and matches the
currently played music to its corresponding notes in the sheet. It consists of
an end-to-end multi-modal convolutional neural network that takes as input
images of sheet music and spectrograms of the respective audio snippets. It
learns to predict, for a given unseen audio snippet (covering approximately one
bar of music), the corresponding position in the respective score line. Our
results suggest that with the use of (deep) neural networks — which have
proven to be powerful image processing models — working with sheet music
becomes feasible and a promising future research direction.
Deep learning is effective for the classification of OCT images of normal versus Age-related Macular Degeneration
Comments: 4 Figures, 1 Table
Subjects:
Machine Learning (stat.ML)
; Computer Vision and Pattern Recognition (cs.CV); Learning (cs.LG)
Objective: The advent of Electronic Medical Records (EMR) with large
electronic imaging databases along with advances in deep neural networks with
machine learning has provided a unique opportunity to achieve milestones in
automated image analysis. Optical coherence tomography (OCT) is the most
commonly obtained imaging modality in ophthalmology and represents a dense and
rich dataset when combined with labels derived from the EMR. We sought to
determine if deep learning could be utilized to distinguish normal OCT images
from images from patients with Age-related Macular Degeneration (AMD). Methods:
Automated extraction of an OCT imaging database was performed and linked to
clinical endpoints from the EMR. OCT macula scans were obtained by Heidelberg
Spectralis, and each OCT scan was linked to EMR clinical endpoints extracted
from EPIC. The central 11 images were selected from each OCT scan of two
cohorts of patients: normal and AMD. Cross-validation was performed using a
random subset of patients. Area under receiver operator curves (auROC) were
constructed at an independent image level, macular OCT level, and patient
level. Results: Of an extraction of 2.6 million OCT images linked to clinical
datapoints from the EMR, 52,690 normal and 48,312 AMD macular OCT images were
selected. A deep neural network was trained to categorize images as either
normal or AMD. At the image level, we achieved an auROC of 92.78% with an
accuracy of 87.63%. At the macula level, we achieved an auROC of 93.83% with an
accuracy of 88.98%. At a patient level, we achieved an auROC of 97.45% with an
accuracy of 93.45%. Peak sensitivity and specificity with optimal cutoffs were
92.64% and 93.69% respectively. Conclusions: Deep learning techniques are
effective for classifying OCT images. These findings have important
implications in utilizing OCT in automated screening and computer aided
diagnosis tools.
Comments: Preprint, journal special issue
Subjects:
Artificial Intelligence (cs.AI)
Ontohub is a repository engine for managing distributed heterogeneous
ontologies. The distributed nature enables communities to share and exchange
their contributions easily. The heterogeneous nature makes it possible to
integrate ontologies written in various ontology languages. Ontohub supports a
wide range of formal logical and ontology languages, as well as various
structuring and modularity constructs and inter-theory (concept) mappings,
building on the OMG-standardized DOL language. Ontohub repositories are
organised as Git repositories, thus inheriting all features of this popular
version control system. Moreover, Ontohub is the first repository engine
meeting a substantial amount of the requirements formulated in the context of
the Open Ontology Repository (OOR) initiative, including an API for federation
as well as support for logical inference and axiom selection.
Rupert Freeman , Sebastien Lahaie , David M. Pennock Subjects : Artificial Intelligence (cs.AI) ; Computer Science and Game Theory (cs.GT)
A prediction market is a useful means of aggregating information about a
future event. To function, the market needs a trusted entity who will verify
the true outcome in the end. Motivated by the recent introduction of
decentralized prediction markets, we introduce a mechanism that allows for the
outcome to be determined by the votes of a group of arbiters who may themselves
hold stakes in the market. Despite the potential conflict of interest, we
derive conditions under which we can incentivize arbiters to vote truthfully by
using funds raised from market fees to implement a peer prediction mechanism.
Finally, we investigate what parameter values could be used in a real-world
implementation of our mechanism.
Comments: Presented at the Constructive Machine Learning workshop at NIPS 2016 as a poster and spotlight talk. 8 pages including 2 page references, 2 page appendix, 3 figures. Blog post (including videos) at this https URL
Subjects:
Artificial Intelligence (cs.AI)
We investigate a human-machine collaborative drawing environment in which an
autonomous agent sketches images while optionally allowing a user to directly
influence the agent’s trajectory. We combine Monte Carlo Tree Search with image
classifiers and test both shallow models (e.g. multinomial logistic regression)
and deep Convolutional Neural Networks (e.g. LeNet, Inception v3). We found
that using the shallow model, the agent produces a limited variety of images,
which are noticably recogonisable by humans. However, using the deeper models,
the agent produces a more diverse range of images, and while the agent remains
very confident (99.99%) in having achieved its objective, to humans they mostly
resemble unrecognisable ‘random’ noise. We relate this to recent research which
also discovered that ‘deep neural networks are easily fooled’ cite{Nguyen2015}
and we discuss possible solutions and future directions for the research.
Franck Dernoncourt , Ji Young Lee , Peter Szolovits Subjects : Computation and Language (cs.CL) ; Artificial Intelligence (cs.AI); Neural and Evolutionary Computing (cs.NE); Machine Learning (stat.ML)
Existing models based on artificial neural networks (ANNs) for sentence
classification often do not incorporate the context in which sentences appear,
and classify sentences individually. However, traditional sentence
classification approaches have been shown to greatly benefit from jointly
classifying subsequent sentences, such as with conditional random fields. In
this work, we present an ANN architecture that combines the effectiveness of
typical ANN models to classify sentences in isolation, with the strength of
structured prediction. Our model achieves state-of-the-art results on two
different datasets for sequential sentence classification in medical abstracts.
Harm van Seijen , Mehdi Fatemi , Joshua Romoff Subjects : Learning (cs.LG) ; Artificial Intelligence (cs.AI)
In this paper, we propose a framework for solving a single-agent task by
using multiple agents, each focusing on different aspects of the task. This
approach has two main advantages: 1) it allows for specialized agents for
different parts of the task, and 2) it provides a new way to transfer
knowledge, by transferring trained agents. Our framework generalizes the
traditional hierarchical decomposition, in which, at any moment in time, a
single agent has control until it has solved its particular subtask. We
illustrate our framework using a number of examples.
Comments: (12 pages, 2 figures) Presented at NIPS Advances In Approximate Inference 2016 (AABI 2016)
Subjects:
Machine Learning (stat.ML)
; Artificial Intelligence (cs.AI)
Bayesian inference on structured models typically relies on the ability to
infer posterior distributions of underlying hidden variables. However,
inference in implicit models or complex posterior distributions is hard. A
popular tool for learning implicit models are generative adversarial networks
(GANs) which learn parameters of generators by fooling discriminators.
Typically, GANs are considered to be models themselves and are not understood
in the context of inference. Current techniques rely on inefficient global
discrimination of joint distributions to perform learning, or only consider
discriminating a single output variable. We overcome these limitations by
treating GANs as a basis for likelihood-free inference in generative models and
generalize them to Bayesian posterior inference over factor graphs. We propose
local learning rules based on message passing minimizing a global divergence
criterion involving cooperating local adversaries used to sidestep explicit
likelihood evaluations. This allows us to compose models and yields a unified
inference and learning framework for adversarial learning. Our framework treats
model specification and inference separately and facilitates richly structured
models within the family of Directed Acyclic Graphs, including components such
as intractable likelihoods, non-differentiable models, simulators and generally
cumbersome models. A key result of our treatment is the insight that Bayesian
inference on structured models can be performed only with sampling and
discrimination when using nonparametric variational families, without access to
explicit distributions. As a side-result, we discuss the link to likelihood
maximization. These approaches hold promise to be useful in the toolbox of
probabilistic modelers and enrich the gamut of current probabilistic
programming applications.
Prajna Upadhyay , Tanuma Patra , Ashwini Purkar , Maya Ramanath Subjects : Computation and Language (cs.CL) ; Artificial Intelligence (cs.AI)
In this paper, we describe the construction of TeKnowbase, a knowledge-base
of technical concepts in computer science. Our main information sources are
technical websites such as Webopedia and Techtarget as well as Wikipedia and
online textbooks. We divide the knowledge-base construction problem into two
parts — the acquisition of entities and the extraction of relationships among
these entities. Our knowledge-base consists of approximately 100,000 triples.
We conducted an evaluation on a sample of triples and report an accuracy of a
little over 90/%. We additionally conducted classification experiments on
StackOverflow data with features from TeKnowbase and achieved improved
classification accuracy.
Jiwei Li , Alexander H. Miller , Sumit Chopra , Marc'Aurelio Ranzato , Jason Weston Subjects : Computation and Language (cs.CL) ; Artificial Intelligence (cs.AI)
A good dialogue agent should have the ability to interact with users. In this
work, we explore this direction by designing a simulator and a set of synthetic
tasks in the movie domain that allow the learner to interact with a teacher by
both asking and answering questions. We investigate how a learner can benefit
from asking questions in both an offline and online reinforcement learning
setting. We demonstrate that the learner improves when asking questions. Our
work represents a first step in developing end-to-end learned interactive
dialogue agents.
Comments: Accepted for the proceedings of the Causation: Foundation to Application Workshop, UAI 2016
Subjects:
Machine Learning (stat.ML)
; Artificial Intelligence (cs.AI); Learning (cs.LG); Systems and Control (cs.SY)
We demonstrate the possibility of classifying causal systems into kinds that
share a common structure without first constructing an explicit dynamical model
or using prior knowledge of the system dynamics. The algorithmic ability to
determine whether arbitrary systems are governed by causal relations of the
same form offers significant practical applications in the development and
validation of dynamical models. It is also of theoretical interest as an
essential stage in the scientific inference of laws from empirical data. The
algorithm presented is based on the dynamical symmetry approach to dynamical
kinds. A dynamical symmetry with respect to time is an intervention on one or
more variables of a system that commutes with the time evolution of the system.
A dynamical kind is a class of systems sharing a set of dynamical symmetries.
The algorithm presented classifies deterministic, time-dependent causal systems
by directly comparing their exhibited symmetries. Using simulated, noisy data
from a variety of nonlinear systems, we show that this algorithm correctly
sorts systems into dynamical kinds. It is robust under significant sampling
error, is immune to violations of normality in sampling error, and fails
gracefully with increasing dynamical similarity. The algorithm we demonstrate
is the first to address this aspect of automated scientific discovery.
Interpretable Semantic Textual Similarity: Finding and explaining differences between sentences
Comments: Preprint version, Knowledge-Based Systems (ISSN: 0950-7051). (2016)
Subjects:
Computation and Language (cs.CL)
; Artificial Intelligence (cs.AI); Learning (cs.LG)
User acceptance of artificial intelligence agents might depend on their
ability to explain their reasoning, which requires adding an interpretability
layer that fa- cilitates users to understand their behavior. This paper focuses
on adding an in- terpretable layer on top of Semantic Textual Similarity (STS),
which measures the degree of semantic equivalence between two sentences. The
interpretability layer is formalized as the alignment between pairs of segments
across the two sentences, where the relation between the segments is labeled
with a relation type and a similarity score. We present a publicly available
dataset of sentence pairs annotated following the formalization. We then
develop a system trained on this dataset which, given a sentence pair, explains
what is similar and different, in the form of graded and typed segment
alignments. When evaluated on the dataset, the system performs better than an
informed baseline, showing that the dataset and task are well-defined and
feasible. Most importantly, two user studies show how the system output can be
used to automatically produce explanations in natural language. Users performed
better when having access to the explanations, pro- viding preliminary evidence
that our dataset and method to automatically produce explanations is useful in
real applications.
Comments: In Proceedings MEMICS 2016, arXiv:1612.04037
Journal-ref: EPTCS 233, 2016, pp. 1-12
Subjects:
Information Retrieval (cs.IR)
; Human-Computer Interaction (cs.HC)
Our work is generally focused on recommending for small or medium-sized
e-commerce portals, where explicit feedback is absent and thus the usage of
implicit feedback is necessary. Nonetheless, for some implicit feedback
features, the presentation context may be of high importance. In this paper, we
present a model of relevant contextual features affecting user feedback,
propose methods leveraging those features, publish a dataset of real e-commerce
users containing multiple user feedback indicators as well as its context and
finally present results of purchase prediction and recommendation experiments.
Off-line experiments with real users of a Czech travel agency website
corroborated the importance of leveraging presentation context in both purchase
prediction and recommendation tasks.
Yike Liu , Abhilash Dighe , Tara Safavi , Danai Koutra Subjects : Information Retrieval (cs.IR)
While advances in computing resources have made processing enormous amounts
of data possible, human ability to identify patterns in such data has not
scaled accordingly. Thus, efficient computational methods for condensing and
simplifying data are becoming vital for extracting actionable insights. In
particular, while data summarization techniques have been studied extensively,
only recently has summarizing interconnected data, or graphs, become popular.
This survey is a structured, comprehensive overview of the state-of-the-art
methods for summarizing graph data. We first broach the motivation behind and
the challenges of graph summarization. We then categorize summarization
approaches by the type of graphs taken as input and further organize each
category by core methodology. Finally, we discuss applications of summarization
on real-world graphs and conclude by describing some open problems in the
field.
Comments: In NIPS 2016 End-to-end Learning for Speech and Audio Processing Workshop, Barcelona, Spain
Subjects:
Sound (cs.SD)
; Information Retrieval (cs.IR); Learning (cs.LG)
This paper demonstrates the feasibility of learning to retrieve short
snippets of sheet music (images) when given a short query excerpt of music
(audio) — and vice versa –, without any symbolic representation of music or
scores. This would be highly useful in many content-based musical retrieval
scenarios. Our approach is based on Deep Canonical Correlation Analysis (DCCA)
and learns correlated latent spaces allowing for cross-modality retrieval in
both directions. Initial experiments with relatively simple monophonic music
show promising results.
Franck Dernoncourt , Ji Young Lee , Peter Szolovits Subjects : Computation and Language (cs.CL) ; Artificial Intelligence (cs.AI); Neural and Evolutionary Computing (cs.NE); Machine Learning (stat.ML)
Existing models based on artificial neural networks (ANNs) for sentence
classification often do not incorporate the context in which sentences appear,
and classify sentences individually. However, traditional sentence
classification approaches have been shown to greatly benefit from jointly
classifying subsequent sentences, such as with conditional random fields. In
this work, we present an ANN architecture that combines the effectiveness of
typical ANN models to classify sentences in isolation, with the strength of
structured prediction. Our model achieves state-of-the-art results on two
different datasets for sequential sentence classification in medical abstracts.
Mickael Rouvier , Benoit Favre Subjects : Computation and Language (cs.CL)
Creating sentiment polarity lexicons is labor intensive. Automatically
translating them from resourceful languages requires in-domain machine
translation systems, which rely on large quantities of bi-texts. In this paper,
we propose to replace machine translation by transferring words from the
lexicon through word embeddings aligned across languages with a simple linear
transform. The approach leads to no degradation, compared to machine
translation, when tested on sentiment polarity classification on tweets from
four languages.
Fugen Zhou , Fuxiang Wu , Zhengchen Zhang , Minghui Dong Subjects : Computation and Language (cs.CL)
This paper presents a novel reranking model, future reward reranking, to
re-score the actions in a transition-based parser by using a global scorer.
Different to conventional reranking parsing, the model searches for the best
dependency tree in all feasible trees constraining by a sequence of actions to
get the future reward of the sequence. The scorer is based on a first-order
graph-based parser with bidirectional LSTM, which catches different parsing
view compared with the transition-based parser. Besides, since context
enhancement has shown substantial improvement in the arc-stand transition-based
parsing over the parsing accuracy, we implement context enhancement on an
arc-eager transition-base parser with stack LSTMs, the dynamic oracle and
dropout supporting and achieve further improvement. With the global scorer and
context enhancement, the results show that UAS of the parser increases as much
as 1.20% for English and 1.66% for Chinese, and LAS increases as much as 1.32%
for English and 1.63% for Chinese. Moreover, we get state-of-the-art LASs,
achieving 87.58% for Chinese and 93.37% for English.
Prajna Upadhyay , Tanuma Patra , Ashwini Purkar , Maya Ramanath Subjects : Computation and Language (cs.CL) ; Artificial Intelligence (cs.AI)
In this paper, we describe the construction of TeKnowbase, a knowledge-base
of technical concepts in computer science. Our main information sources are
technical websites such as Webopedia and Techtarget as well as Wikipedia and
online textbooks. We divide the knowledge-base construction problem into two
parts — the acquisition of entities and the extraction of relationships among
these entities. Our knowledge-base consists of approximately 100,000 triples.
We conducted an evaluation on a sample of triples and report an accuracy of a
little over 90/%. We additionally conducted classification experiments on
StackOverflow data with features from TeKnowbase and achieved improved
classification accuracy.
Jiwei Li , Alexander H. Miller , Sumit Chopra , Marc'Aurelio Ranzato , Jason Weston Subjects : Computation and Language (cs.CL) ; Artificial Intelligence (cs.AI)
A good dialogue agent should have the ability to interact with users. In this
work, we explore this direction by designing a simulator and a set of synthetic
tasks in the movie domain that allow the learner to interact with a teacher by
both asking and answering questions. We investigate how a learner can benefit
from asking questions in both an offline and online reinforcement learning
setting. We demonstrate that the learner improves when asking questions. Our
work represents a first step in developing end-to-end learned interactive
dialogue agents.
Interpretable Semantic Textual Similarity: Finding and explaining differences between sentences
Comments: Preprint version, Knowledge-Based Systems (ISSN: 0950-7051). (2016)
Subjects:
Computation and Language (cs.CL)
; Artificial Intelligence (cs.AI); Learning (cs.LG)
User acceptance of artificial intelligence agents might depend on their
ability to explain their reasoning, which requires adding an interpretability
layer that fa- cilitates users to understand their behavior. This paper focuses
on adding an in- terpretable layer on top of Semantic Textual Similarity (STS),
which measures the degree of semantic equivalence between two sentences. The
interpretability layer is formalized as the alignment between pairs of segments
across the two sentences, where the relation between the segments is labeled
with a relation type and a similarity score. We present a publicly available
dataset of sentence pairs annotated following the formalization. We then
develop a system trained on this dataset which, given a sentence pair, explains
what is similar and different, in the form of graded and typed segment
alignments. When evaluated on the dataset, the system performs better than an
informed baseline, showing that the dataset and task are well-defined and
feasible. Most importantly, two user studies show how the system output can be
used to automatically produce explanations in natural language. Users performed
better when having access to the explanations, pro- viding preliminary evidence
that our dataset and method to automatically produce explanations is useful in
real applications.
Recurrent Image Captioner: Describing Images with Spatial-Invariant Transformation and Attention Filtering
Hao Liu , Yang Yang , Fumin Shen , Lixin Duan , Heng Tao Shen Subjects : Computer Vision and Pattern Recognition (cs.CV) ; Computation and Language (cs.CL)
Along with the prosperity of recurrent neural network in modelling sequential
data and the power of attention mechanism in automatically identify salient
information, image captioning, a.k.a., image description, has been remarkably
advanced in recent years. Nonetheless, most existing paradigms may suffer from
the deficiency of invariance to images with different scaling, rotation, etc.;
and effective integration of standalone attention to form a holistic end-to-end
system. In this paper, we propose a novel image captioning architecture, termed
Recurrent Image Captioner ( extbf{RIC}), which allows visual encoder and
language decoder to coherently cooperate in a recurrent manner. Specifically,
we first equip CNN-based visual encoder with a differentiable layer to enable
spatially invariant transformation of visual signals. Moreover, we deploy an
attention filter module (differentiable) between encoder and decoder to
dynamically determine salient visual parts. We also employ bidirectional LSTM
to preprocess sentences for generating better textual representations. Besides,
we propose to exploit variational inference to optimize the whole architecture.
Extensive experimental results on three benchmark datasets (i.e., Flickr8k,
Flickr30k and MS COCO) demonstrate the superiority of our proposed architecture
as compared to most of the state-of-the-art methods.
Shripad Gade , Nitin H. Vaidya Subjects : Distributed, Parallel, and Cluster Computing (cs.DC) ; Learning (cs.LG); Optimization and Control (math.OC)
Continual data collection and widespread deployment of machine learning
algorithms, particularly the distributed variants, have raised new privacy
challenges. In a distributed machine learning scenario, the dataset is stored
among several machines and they solve a distributed optimization problem to
collectively learn the underlying model. We present a secure multi-party
computation inspired privacy preserving distributed algorithm for optimizing a
convex function consisting of several possibly non-convex functions. Each
individual objective function is privately stored with an agent while the
agents communicate model parameters with neighbor machines connected in a
network. We show that our algorithm can correctly optimize the overall
objective function and learn the underlying model accurately. We further prove
that under a vertex connectivity condition on the topology, our algorithm
preserves privacy of individual objective functions. We establish limits on the
what a coalition of adversaries can learn by observing the messages and states
shared over a network.
Mohammad Roohitavaf , Sandeep Kulkarni Subjects : Distributed, Parallel, and Cluster Computing (cs.DC)
Causal consistency is in an intermediate consistency model that can be
achieved together with high availability and high performance requirements even
in presence of network partitions. There are several proposals in the
literature for causally consistent data stores. Thanks to the use of single
scalar physical clocks, GentleRain has a throughput higher than other proposals
such as COPS or Orbe. However, both of its correctness and performance relay on
monotonic synchronized physical clocks. Specifically, if physical clocks go
backward its correctness is violated. In addition, GentleRain is sensitive on
the clock synchronization, and clock skew may slow write operations in
GenlteRain. In this paper, we want to solve this issue in GenlteRain by using
Hybrid Logical Clock (HLC) instead of physical clocks. Using HLC, GentleRain
protocl is not sensitive on the clock skew anymore. In addition, even if clocks
go backward, the correctness of the system is not violated. Furthermore, by
HLC, we timestamp versions with a clock very close to the physical clocks.
Thus, we can take causally consistency snapshot of the system at any give
physical time. We call GentleRain protocol with HLCs GentleRain+. We have
implemented GentleRain+ protocol, and have evaluated it experimentally.
GentleRain+ provides faster write operations compare to GentleRain that rely
solely on physical clocks to achieve causal consistency. We have also shown
that using HLC instead of physical clock does not have any overhead. Thus, it
makes GentleRain more robust on clock anomalies at no cost.
Comments: 11 pages, 10 figures
Subjects:
Cryptography and Security (cs.CR)
; Distributed, Parallel, and Cluster Computing (cs.DC)
The surging interest in blockchain technology has revitalized the search for
effective Byzantine consensus schemes. In particular, the blockchain community
has been looking for ways to effectively integrate traditional Byzantine
fault-tolerant (BFT) protocols into a blockchain consensus layer allowing
various financial institutions to securely agree on the order of transactions.
However, existing BFT protocols can only scale to tens of nodes due to their
(O(n^2)) message complexity.
In this paper, we propose FastBFT, the fastest and most scalable BFT protocol
to-date. At the heart of FastBFT is a novel message aggregation technique that
combines hardware-based trusted execution environments (TEEs) with lightweight
secret sharing primitives. Combining this technique with several other
optimizations (i.e., optimistic execution, tree topology and failure
detection), FastBFT achieves low latency and high throughput even for large
scale networks. Via systematic analysis and experiments, we demonstrate that
FastBFT has better scalability and performance than previous BFT protocols.
Efficient Distributed Semi-Supervised Learning using Stochastic Regularization over Affinity Graphs
Comments: NIPS 2016 Workshop on Machine Learning Systems
Subjects:
Machine Learning (stat.ML)
; Distributed, Parallel, and Cluster Computing (cs.DC); Learning (cs.LG)
We describe a computationally efficient, stochastic graph-regularization
technique that can be utilized for the semi-supervised training of deep neural
networks in a parallel or distributed setting. We utilize a technique, first
described in [13] for the construction of mini-batches for stochastic gradient
descent (SGD) based on synthesized partitions of an affinity graph that are
consistent with the graph structure, but also preserve enough stochasticity for
convergence of SGD to good local minima. We show how our technique allows a
graph-based semi-supervised loss function to be decomposed into a sum over
objectives, facilitating data parallelism for scalable training of machine
learning models. Empirical results indicate that our method significantly
improves classification accuracy compared to the fully-supervised case when the
fraction of labeled data is low, and in the parallel case, achieves significant
speed-up in terms of wall-clock time to convergence. We show the results for
both sequential and distributed-memory semi-supervised DNN training on a speech
corpus.
Comments: 9 pages, 4 figures
Subjects:
Learning (cs.LG)
; Neural and Evolutionary Computing (cs.NE); Machine Learning (stat.ML)
We present a method for implementing an Efficient Unitary Neural Network
(EUNN) whose computational complexity is merely (mathcal{O}(1)) per parameter
and has full tunability, from spanning part of unitary space to all of it. We
apply the EUNN in Recurrent Neural Networks, and test its performance on the
standard copying task and the MNIST digit recognition benchmark, finding that
it significantly outperforms a non-unitary RNN, an LSTM network, an exclusively
partial space URNN and a projective URNN with comparable parameter numbers.
Harm van Seijen , Mehdi Fatemi , Joshua Romoff Subjects : Learning (cs.LG) ; Artificial Intelligence (cs.AI)
In this paper, we propose a framework for solving a single-agent task by
using multiple agents, each focusing on different aspects of the task. This
approach has two main advantages: 1) it allows for specialized agents for
different parts of the task, and 2) it provides a new way to transfer
knowledge, by transferring trained agents. Our framework generalizes the
traditional hierarchical decomposition, in which, at any moment in time, a
single agent has control until it has solved its particular subtask. We
illustrate our framework using a number of examples.
Lukas Balles , Javier Romero , Philipp Hennig Subjects : Learning (cs.LG) ; Computer Vision and Pattern Recognition (cs.CV); Machine Learning (stat.ML)
Mini-batch stochastic gradient descent and variants thereof have become
standard for large-scale empirical risk minimization like the training of
neural networks. These methods are usually used with a constant batch size
chosen by simple empirical inspection. The batch size significantly influences
the behavior of the stochastic optimization algorithm, though, since it
determines the variance of the gradient estimates. This variance also changes
over the optimization process; when using a constant batch size, stability and
convergence is thus often enforced by means of a (manually tuned) decreasing
learning rate schedule. We propose a practical method for dynamic batch size
adaptation. It estimates the variance of the stochastic gradients and adapts
the batch size to decrease the variance proportionally to the value of the
objective function, removing the need for the aforementioned learning rate
decrease. In contrast to recent related work, our algorithm couples the batch
size to the learning rate, directly reflecting the known relationship between
the two. On three image classification benchmarks, our batch size adaptation
yields faster optimization convergence, while simultaneously simplifying
learning rate tuning. A TensorFlow implementation is available.
Comments: In Proceedings of the 2016 IEEE 26th International Workshop on Machine Learning for Signal Processing (MLSP), Vietro sul Mare, Italy
Subjects:
Learning (cs.LG)
; Sound (cs.SD)
Chord recognition systems depend on robust feature extraction pipelines.
While these pipelines are traditionally hand-crafted, recent advances in
end-to-end machine learning have begun to inspire researchers to explore
data-driven methods for such tasks. In this paper, we present a chord
recognition system that uses a fully convolutional deep auditory model for
feature extraction. The extracted features are processed by a Conditional
Random Field that decodes the final chord sequence. Both processing stages are
trained automatically and do not require expert knowledge for optimising
parameters. We show that the learned auditory system extracts musically
interpretable features, and that the proposed chord recognition system achieves
results on par or better than state-of-the-art algorithms.
Comments: Published In Proceedings of the 17th International Society for Music Information Retrieval Conference (2016)
Subjects:
Learning (cs.LG)
; Computer Vision and Pattern Recognition (cs.CV)
This paper addresses the matching of short music audio snippets to the
corresponding pixel location in images of sheet music. A system is presented
that simultaneously learns to read notes, listens to music and matches the
currently played music to its corresponding notes in the sheet. It consists of
an end-to-end multi-modal convolutional neural network that takes as input
images of sheet music and spectrograms of the respective audio snippets. It
learns to predict, for a given unseen audio snippet (covering approximately one
bar of music), the corresponding position in the respective score line. Our
results suggest that with the use of (deep) neural networks — which have
proven to be powerful image processing models — working with sheet music
becomes feasible and a promising future research direction.
Comments: Accepted as an oral presentation in 42th IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP)
Subjects:
Learning (cs.LG)
; Information Theory (cs.IT)
Compressive sensing (CS) is a promising technology for realizing
energy-efficient wireless sensors for long-term health monitoring. However,
conventional model-driven CS frameworks suffer from limited compression ratio
and reconstruction quality when dealing with physiological signals due to
inaccurate models and the overlook of individual variability. In this paper, we
propose a data-driven CS framework that can learn signal characteristics and
personalized features from any individual recording of physiologic signals to
enhance CS performance with a minimized number of measurements. Such
improvements are accomplished by a co-training approach that optimizes the
sensing matrix and the dictionary towards improved restricted isometry property
and signal sparsity, respectively. Experimental results upon ECG signals show
that the proposed method, at a compression ratio of 10x, successfully reduces
the isometry constant of the trained sensing matrices by 86% against random
matrices and improves the overall reconstructed signal-to-noise ratio by 15dB
over conventional model-driven approaches.
Ian Dewancker , Michael McCourt , Scott Clark Subjects : Learning (cs.LG)
The engineering of machine learning systems is still a nascent field; relying
on a seemingly daunting collection of quickly evolving tools and best
practices. It is our hope that this guidebook will serve as a useful resource
for machine learning practitioners looking to take advantage of Bayesian
optimization techniques. We outline four example machine learning problems that
can be solved using open source machine learning libraries, and highlight the
benefits of using Bayesian optimization in the context of these common machine
learning applications.
Hoel Le Capitaine Subjects : Learning (cs.LG) ; Machine Learning (stat.ML)
A number of machine learning algorithms are using a metric, or a distance, in
order to compare individuals. The Euclidean distance is usually employed, but
it may be more efficient to learn a parametric distance such as Mahalanobis
metric. Learning such a metric is a hot topic since more than ten years now,
and a number of methods have been proposed to efficiently learn it. However,
the nature of the problem makes it quite difficult for large scale data, as
well as data for which classes overlap. This paper presents a simple way of
improving accuracy and scalability of any iterative metric learning algorithm,
where constraints are obtained prior to the algorithm. The proposed approach
relies on a loss-dependent weighted selection of constraints that are used for
learning the metric. Using the corresponding dedicated loss function, the
method clearly allows to obtain better results than state-of-the-art methods,
both in terms of accuracy and time complexity. Some experimental results on
real world, and potentially large, datasets are demonstrating the effectiveness
of our proposition.
Shripad Gade , Nitin H. Vaidya Subjects : Distributed, Parallel, and Cluster Computing (cs.DC) ; Learning (cs.LG); Optimization and Control (math.OC)
Continual data collection and widespread deployment of machine learning
algorithms, particularly the distributed variants, have raised new privacy
challenges. In a distributed machine learning scenario, the dataset is stored
among several machines and they solve a distributed optimization problem to
collectively learn the underlying model. We present a secure multi-party
computation inspired privacy preserving distributed algorithm for optimizing a
convex function consisting of several possibly non-convex functions. Each
individual objective function is privately stored with an agent while the
agents communicate model parameters with neighbor machines connected in a
network. We show that our algorithm can correctly optimize the overall
objective function and learn the underlying model accurately. We further prove
that under a vertex connectivity condition on the topology, our algorithm
preserves privacy of individual objective functions. We establish limits on the
what a coalition of adversaries can learn by observing the messages and states
shared over a network.
CSVideoNet: A Recurrent Convolutional Neural Network for Compressive Sensing Video Reconstruction
Comments: 10 pages, 6 pages, 2 tables
Subjects:
Computer Vision and Pattern Recognition (cs.CV)
; Learning (cs.LG)
In this paper, we develop a deep neural network architecture called
“CSVideoNet” that can learn visual representations from random measurements for
compressive sensing (CS) video reconstruction. CSVideoNet is an end-to-end
trainable and non-iterative model that combines convolutional neural networks
(CNNs) with a recurrent neural networks (RNN) to facilitate video
reconstruction by leveraging temporal-spatial features. The proposed network
can accept random measurements with a multi-level compression ratio (CR). The
lightly and aggressively compressed measurements offer background information
and object details, respectively. This is similar to the variable bit rate
techniques widely used in conventional video coding approaches. The RNN
employed by CSVideoNet can leverage temporal coherence that exists in adjacent
video frames to extrapolate motion features and merge them with spatial visual
features extracted by the CNNs to further enhance reconstruction quality,
especially at high CRs. We test our CSVideoNet on the UCF-101 dataset.
Experimental results show that CSVideoNet outperforms the existing video CS
reconstruction approaches. The results demonstrate that our method can preserve
relatively excellent visual details from original videos even at a 100x CR,
which is difficult to realize with the reference approaches. Also, the
non-iterative nature of CSVideoNet results in an decrease in runtime by three
orders of magnitude over iterative reconstruction algorithms. Furthermore,
CSVideoNet can enhance the CR of CS cameras beyond the limitation of
conventional approaches, ensuring a reduction in bandwidth for data
transmission. These benefits are especially favorable to high-frame-rate video
applications.
Comments: Proceedings of the 17th International Society for Music Information Retrieval Conference (ISMIR 2016), New York, NY
Subjects:
Sound (cs.SD)
; Learning (cs.LG)
In an attempt at exploring the limitations of simple approaches to the task
of piano transcription (as usually defined in MIR), we conduct an in-depth
analysis of neural network-based framewise transcription. We systematically
compare different popular input representations for transcription systems to
determine the ones most suitable for use with neural networks. Exploiting
recent advances in training techniques and new regularizers, and taking into
account hyper-parameter tuning, we show that it is possible, by simple
bottom-up frame-wise processing, to obtain a piano transcriber that outperforms
the current published state of the art on the publicly available MAPS dataset
— without any complex post-processing steps. Thus, we propose this simple
approach as a new baseline for this dataset, for future transcription research
to build on and improve.
Comments: In NIPS 2016 End-to-end Learning for Speech and Audio Processing Workshop, Barcelona, Spain
Subjects:
Sound (cs.SD)
; Information Retrieval (cs.IR); Learning (cs.LG)
This paper demonstrates the feasibility of learning to retrieve short
snippets of sheet music (images) when given a short query excerpt of music
(audio) — and vice versa –, without any symbolic representation of music or
scores. This would be highly useful in many content-based musical retrieval
scenarios. Our approach is based on Deep Canonical Correlation Analysis (DCCA)
and learns correlated latent spaces allowing for cross-modality retrieval in
both directions. Initial experiments with relatively simple monophonic music
show promising results.
Comments: In Proceedings of the 17th International Society for Music Information Retrieval Conference (ISMIR), New York, USA, 2016
Subjects:
Sound (cs.SD)
; Learning (cs.LG)
We explore frame-level audio feature learning for chord recognition using
artificial neural networks. We present the argument that chroma vectors
potentially hold enough information to model harmonic content of audio for
chord recognition, but that standard chroma extractors compute too noisy
features. This leads us to propose a learned chroma feature extractor based on
artificial neural networks. It is trained to compute chroma features that
encode harmonic information important for chord recognition, while being robust
to irrelevant interferences. We achieve this by feeding the network an audio
spectrum with context instead of a single frame as input. This way, the network
can learn to selectively compensate noise and resolve harmonic ambiguities.
We compare the resulting features to hand-crafted ones by using a simple
linear frame-wise classifier for chord recognition on various data sets. The
results show that the learned feature extractor produces superior chroma
vectors for chord recognition.
Ashish Bora , Sugato Basu , Joydeep Ghosh Subjects : Neural and Evolutionary Computing (cs.NE) ; Learning (cs.LG)
Many time series are generated by a set of entities that interact with one
another over time. This paper introduces a broad, flexible framework to learn
from multiple inter-dependent time series generated by such entities. Our
framework explicitly models the entities and their interactions through time.
It achieves this by building on the capabilities of Recurrent Neural Networks,
while also offering several ways to incorporate domain knowledge/constraints
into the model architecture. The capabilities of our approach are showcased
through an application to weather prediction, which shows gains over strong
baselines.
Comments: 4 pages, 11 pages of supplementary information
Subjects:
Statistical Mechanics (cond-mat.stat-mech)
; Learning (cs.LG); Data Analysis, Statistics and Probability (physics.data-an); Machine Learning (stat.ML)
Reconstruction of structure and parameters of a graphical model from binary
samples is a problem of practical importance in a variety of disciplines,
ranging from statistical physics and computational biology to image processing
and machine learning. The focus of the research community shifted towards
developing universal reconstruction algorithms which are both computationally
efficient and require the minimal amount of expensive data. We introduce a new
method, Interaction Screening, which accurately estimates the model parameters
using local optimization problems. The algorithm provably achieves perfect
graph structure recovery with an information-theoretically optimal number of
samples and outperforms state of the art techniques, especially in the
low-temperature regime which is known to be the hardest for learning. We assess
the efficacy of Interaction Screening through extensive numerical tests on
Ising models of various topologies and with different types of interactions,
ranging from ferromagnetic to spin-glass.
Comments: 11 pages, 8 figures
Subjects:
Social and Information Networks (cs.SI)
; Learning (cs.LG); Data Analysis, Statistics and Probability (physics.data-an); Machine Learning (stat.ML)
We address the problem of semi-supervised learning in relational networks,
networks in which nodes are entities and links are the relationships or
interactions between them. Typically this problem is confounded with the
problem of graph-based semi-supervised learning (GSSL), because both problems
represent the data as a graph and predict the missing class labels of nodes.
However, not all graphs are created equally. In GSSL a graph is constructed,
often from independent data, based on similarity. As such, edges tend to
connect instances with the same class label. Relational networks, however, can
be more heterogeneous and edges do not always indicate similarity. For
instance, instead of links being more likely to connect nodes with the same
class label, they may occur more frequently between nodes with different class
labels (link-heterogeneity). Or nodes with the same class label do not
necessarily have the same type of connectivity across the whole network
(class-heterogeneity), e.g. in a network of sexual interactions we may observe
links between opposite genders in some parts of the graph and links between the
same genders in others. Performing classification in networks with different
types of heterogeneity is a hard problem that is made harder still when we do
not know a-priori the type or level of heterogeneity. Here we present two
scalable approaches for graph-based semi-supervised learning for the more
general case of relational networks. We demonstrate these approaches on
synthetic and real-world networks that display different link patterns within
and between classes. Compared to state-of-the-art approaches, ours give better
classification performance without prior knowledge of how classes interact. In
particular, our two-step label propagation algorithm gives consistently good
accuracy and runs on networks of over 1.6 million nodes and 30 million edges in
around 12 seconds.
Comments: Pre-print. The final publication is available at Springer via this http URL
Subjects:
Neural and Evolutionary Computing (cs.NE)
; Learning (cs.LG)
In an attempt to solve the lengthy training times of neural networks, we
proposed Parallel Circuits (PCs), a biologically inspired architecture.
Previous work has shown that this approach fails to maintain generalization
performance in spite of achieving sharp speed gains. To address this issue, and
motivated by the way Dropout prevents node co-adaption, in this paper, we
suggest an improvement by extending Dropout to the PC architecture. The paper
provides multiple insights into this combination, including a variety of fusion
approaches. Experiments show promising results in which improved error rates
are achieved in most cases, whilst maintaining the speed advantage of the PC
approach.
Comments: Accepted for the proceedings of the Causation: Foundation to Application Workshop, UAI 2016
Subjects:
Machine Learning (stat.ML)
; Artificial Intelligence (cs.AI); Learning (cs.LG); Systems and Control (cs.SY)
We demonstrate the possibility of classifying causal systems into kinds that
share a common structure without first constructing an explicit dynamical model
or using prior knowledge of the system dynamics. The algorithmic ability to
determine whether arbitrary systems are governed by causal relations of the
same form offers significant practical applications in the development and
validation of dynamical models. It is also of theoretical interest as an
essential stage in the scientific inference of laws from empirical data. The
algorithm presented is based on the dynamical symmetry approach to dynamical
kinds. A dynamical symmetry with respect to time is an intervention on one or
more variables of a system that commutes with the time evolution of the system.
A dynamical kind is a class of systems sharing a set of dynamical symmetries.
The algorithm presented classifies deterministic, time-dependent causal systems
by directly comparing their exhibited symmetries. Using simulated, noisy data
from a variety of nonlinear systems, we show that this algorithm correctly
sorts systems into dynamical kinds. It is robust under significant sampling
error, is immune to violations of normality in sampling error, and fails
gracefully with increasing dynamical similarity. The algorithm we demonstrate
is the first to address this aspect of automated scientific discovery.
Semi-Supervised Phone Classification using Deep Neural Networks and Stochastic Graph-Based Entropic Regularization
Comments: InterSpeech Workshop on Machine Learning in Speech and Language Processing, 2016
Subjects:
Machine Learning (stat.ML)
; Learning (cs.LG)
We describe a graph-based semi-supervised learning framework in the context
of deep neural networks that uses a graph-based entropic regularizer to favor
smooth solutions over a graph induced by the data. The main contribution of
this work is a computationally efficient, stochastic graph-regularization
technique that uses mini-batches that are consistent with the graph structure,
but also provides enough stochasticity (in terms of mini-batch data diversity)
for convergence of stochastic gradient descent methods to good solutions. For
this work, we focus on results of frame-level phone classification accuracy on
the TIMIT speech corpus but our method is general and scalable to much larger
data sets. Results indicate that our method significantly improves
classification accuracy compared to the fully-supervised case when the fraction
of labeled data is low, and it is competitive with other methods in the fully
labeled case.
Efficient Distributed Semi-Supervised Learning using Stochastic Regularization over Affinity Graphs
Comments: NIPS 2016 Workshop on Machine Learning Systems
Subjects:
Machine Learning (stat.ML)
; Distributed, Parallel, and Cluster Computing (cs.DC); Learning (cs.LG)
We describe a computationally efficient, stochastic graph-regularization
technique that can be utilized for the semi-supervised training of deep neural
networks in a parallel or distributed setting. We utilize a technique, first
described in [13] for the construction of mini-batches for stochastic gradient
descent (SGD) based on synthesized partitions of an affinity graph that are
consistent with the graph structure, but also preserve enough stochasticity for
convergence of SGD to good local minima. We show how our technique allows a
graph-based semi-supervised loss function to be decomposed into a sum over
objectives, facilitating data parallelism for scalable training of machine
learning models. Empirical results indicate that our method significantly
improves classification accuracy compared to the fully-supervised case when the
fraction of labeled data is low, and in the parallel case, achieves significant
speed-up in terms of wall-clock time to convergence. We show the results for
both sequential and distributed-memory semi-supervised DNN training on a speech
corpus.
Deep learning is effective for the classification of OCT images of normal versus Age-related Macular Degeneration
Comments: 4 Figures, 1 Table
Subjects:
Machine Learning (stat.ML)
; Computer Vision and Pattern Recognition (cs.CV); Learning (cs.LG)
Objective: The advent of Electronic Medical Records (EMR) with large
electronic imaging databases along with advances in deep neural networks with
machine learning has provided a unique opportunity to achieve milestones in
automated image analysis. Optical coherence tomography (OCT) is the most
commonly obtained imaging modality in ophthalmology and represents a dense and
rich dataset when combined with labels derived from the EMR. We sought to
determine if deep learning could be utilized to distinguish normal OCT images
from images from patients with Age-related Macular Degeneration (AMD). Methods:
Automated extraction of an OCT imaging database was performed and linked to
clinical endpoints from the EMR. OCT macula scans were obtained by Heidelberg
Spectralis, and each OCT scan was linked to EMR clinical endpoints extracted
from EPIC. The central 11 images were selected from each OCT scan of two
cohorts of patients: normal and AMD. Cross-validation was performed using a
random subset of patients. Area under receiver operator curves (auROC) were
constructed at an independent image level, macular OCT level, and patient
level. Results: Of an extraction of 2.6 million OCT images linked to clinical
datapoints from the EMR, 52,690 normal and 48,312 AMD macular OCT images were
selected. A deep neural network was trained to categorize images as either
normal or AMD. At the image level, we achieved an auROC of 92.78% with an
accuracy of 87.63%. At the macula level, we achieved an auROC of 93.83% with an
accuracy of 88.98%. At a patient level, we achieved an auROC of 97.45% with an
accuracy of 93.45%. Peak sensitivity and specificity with optimal cutoffs were
92.64% and 93.69% respectively. Conclusions: Deep learning techniques are
effective for classifying OCT images. These findings have important
implications in utilizing OCT in automated screening and computer aided
diagnosis tools.
Interpretable Semantic Textual Similarity: Finding and explaining differences between sentences
Comments: Preprint version, Knowledge-Based Systems (ISSN: 0950-7051). (2016)
Subjects:
Computation and Language (cs.CL)
; Artificial Intelligence (cs.AI); Learning (cs.LG)
User acceptance of artificial intelligence agents might depend on their
ability to explain their reasoning, which requires adding an interpretability
layer that fa- cilitates users to understand their behavior. This paper focuses
on adding an in- terpretable layer on top of Semantic Textual Similarity (STS),
which measures the degree of semantic equivalence between two sentences. The
interpretability layer is formalized as the alignment between pairs of segments
across the two sentences, where the relation between the segments is labeled
with a relation type and a similarity score. We present a publicly available
dataset of sentence pairs annotated following the formalization. We then
develop a system trained on this dataset which, given a sentence pair, explains
what is similar and different, in the form of graded and typed segment
alignments. When evaluated on the dataset, the system performs better than an
informed baseline, showing that the dataset and task are well-defined and
feasible. Most importantly, two user studies show how the system output can be
used to automatically produce explanations in natural language. Users performed
better when having access to the explanations, pro- viding preliminary evidence
that our dataset and method to automatically produce explanations is useful in
real applications.
Comments: To appear in Tenth ACM International conference on Web Search and Data Mining (WSDM) in 2017
Subjects:
Social and Information Networks (cs.SI)
; Learning (cs.LG); Physics and Society (physics.soc-ph); Machine Learning (stat.ML)
Learning from the crowd has become increasingly popular in the Web and social
media. There is a wide variety of crowdlearning sites in which, on the one
hand, users learn from the knowledge that other users contribute to the site,
and, on the other hand, knowledge is reviewed and curated by the same users
using assessment measures such as upvotes or likes.
In this paper, we present a probabilistic modeling framework of
crowdlearning, which uncovers the evolution of a user’s expertise over time by
leveraging other users’ assessments of her contributions. The model allows for
both off-site and on-site learning and captures forgetting of knowledge. We
then develop a scalable estimation method to fit the model parameters from
millions of recorded learning and contributing events. We show the
effectiveness of our model by tracing activity of ~25 thousand users in Stack
Overflow over a 4.5 year period. We find that answers with high knowledge value
are rare. Newbies and experts tend to acquire less knowledge than users in the
middle range. Prolific learners tend to be also proficient contributors that
post answers with high knowledge value.
Lossy Transmission of Correlated Sources over a Multiple Access Channel: Necessary Conditions and Separation Results
Comments: Submitted to IEEE Transactions on Information Theory on Nov 30, 2016
Subjects:
Information Theory (cs.IT)
Lossy communication of correlated sources over a multiple access channel is
studied. First, lossy communication is investigated in the presence of
correlated decoder side information. An achievable joint source-channel coding
scheme is presented, and the conditions under which separate source and channel
coding is optimal are explored. It is shown that separation is optimal when the
encoders and the decoder have access to a common observation conditioned on
which the two sources are independent. Separation is shown to be optimal also
when only the encoders have access to such a common observation whose lossless
recovery is required at the decoder. Moreover, the optimality of separation is
shown for sources with a common part, and sources with reconstruction
constraints. Next, these results obtained for the system in presence of side
information are utilized to provide a set of necessary conditions for the
transmission of correlated sources over a multiple access channel without side
information. The identified necessary conditions are specialized to the case of
bivariate Gaussian sources over a Gaussian multiple access channel, and are
shown to be tighter than known results in the literature in certain cases. Our
results indicate that side information can have a significant impact on the
optimality of source-channel separation in lossy transmission, in addition to
being instrumental in identifying necessary conditions for the transmission of
correlated sources when no side information is present.
Comments: Pre-print, submitted for review
Subjects:
Information Theory (cs.IT)
; Optimization and Control (math.OC)
The roll-out of smart meters in electricity networks introduces risks for
consumer privacy due to increased measurement frequency and granularity.
Through various Non-Intrusive Load Monitoring techniques, consumer behavior may
be inferred from their metering data. In this paper, we propose an energy
management method that protects privacy through the minimization of information
leakage. The method is based on a Model Predictive Controller that utilizes
energy storage and local generation, and that predicts the effects of its
actions on the statistics of the actual energy consumption of a consumer and
that seen by the grid. Computationally, the method requires solving a
Mixed-Integer Quadratic Program of manageable size whenever new meter readings
are available. We simulate the controller on generated residential load
profiles with different privacy costs in a two-tier time-of-use energy pricing
environment. Results show that information leakage is effectively reduced at
the expense of increased energy cost. The results also show that, using the
proposed controller, the consumer load profile seen by the grid resembles a
mixture between that obtained with Non-Intrusive Load Leveling and Lazy
Stepping.
Jessalyn Bolkema , Heide Gluesing-Luerssen , Christine A. Kelley , Kristin Lauter , Beth Malmskog , Joachim Rosenthal Subjects : Information Theory (cs.IT) ; Cryptography and Security (cs.CR)
Two variations of the McEliece cryptosystem are presented. The first one is
based on a relaxation of the column permutation in the classical McEliece
scrambling process. This is done in such a way that the Hamming weight of the
error, added in the encryption process, can be controlled so that efficient
decryption remains possible. The second variation is based on the use of
spatially coupled moderate-density parity-check codes as secret codes. These
codes are known for their excellent error-correction performance and allow for
a relatively low key size in the cryptosystem. For both variants the security
with respect to known attacks is discussed.
Comments: submitted for publication
Subjects:
Information Theory (cs.IT)
In this paper, we consider a multi-user wireless system with one full duplex
(FD) base station (BS) serving a set of half duplex (HD) mobile users.To cope
with the in-band self-interference (SI) and co-channel interference, we
formulate a quality-of-service (QoS) based linear transceiver design problem.
The problem jointly optimizes the downlink (DL) and uplink (UL) beamforming
vectors of the BS and the transmission powers of UL users so as to provide both
the DL and UL users with guaranteed signal-to-interference-plus-noise ratio
performance, using a minimum UL and DL transmission sum power.The considered
system model not only takes into account noise caused by non-ideal RF circuits,
analog/digital SI cancellation but also constrains the maximum signal power at
the input of the analog-to-digital converter (ADC) for avoiding signal
distortion due to finite ADC precision. The formulated design problem is not
convex and challenging to solve in general. We first show that for a special
case where the SI channel estimation errors are independent and identically
distributed, the QoS-based linear transceiver design problem is globally
solvable by a polynomial-time bisection algorithm.For the general case, we
propose a suboptimal algorithm based on alternating optimization (AO). The AO
algorithm is guaranteed to converge to a Karush-Kuhn-Tucker solution.To reduce
the complexity of the AO algorithm, we further develop a fixed-point method by
extending the classical uplink-downlink duality in HD systems to the FD
system.Simulation results are presented to demonstrate the performance of the
proposed algorithms and the comparison with HD systems.
Comments: Submitted for possible journal publication
Subjects:
Information Theory (cs.IT)
This paper considers the joint antenna selection (AS) problem for a classical
two-user MIMO non-orthogonal multiple access (NOMA) system, where both the base
station (BS) and users (UEs) are equipped with multiple antennas. Specifically,
several computationally-efficient AS algorithms are developed for two
commonly-used NOMA scenarios: fixed power allocation NOMA (F-NOMA) and
cognitive radio-inspired NOMA (CR-NOMA). For the F-NOMA system, two novel AS
schemes, namely max-max-max AS (A(^3)-AS) and max-min-max AS (AIA-AS), are
proposed to maximize the system sum-rate, without and with the consideration of
user fairness, respectively. In the CR-NOMA network, a novel AS algorithm,
termed maximum-channel-gain-based AS (MCG-AS), is proposed to maximize the
achievable rate of the secondary user, under the condition that the primary
user’s quality of service requirement is satisfied. The asymptotic closed-form
expressions of the average sum-rate for A(^3)-AS and AIA-AS and that of the
average rate of the secondary user for MCG-AS are derived, respectively.
Numerical results demonstrate that the AIA-AS provides better user-fairness,
while the A(^3)-AS achieves a near-optimal sum-rate in F-NOMA systems. For the
CR-NOMA scenario, MCG-AS achieves a near-optimal performance in a wide SNR
regime. Furthermore, all the proposed AS algorithms yield a significant
computational complexity reduction, compared to exhaustive search-based
counterparts.
Hanaa Marshoud , Paschalis C. Sofotasios , Sami Muhaidat , Bayan S. Sharif , George K. Karagiannidis Subjects : Information Theory (cs.IT)
Multiple-input multiple-output (MIMO) techniques have recently demonstrated
significant potentials in visible light communications (VLC), as they can
overcome the modulation bandwidth limitation and provide substantial
improvement in terms of spectral efficiency and link reliability. However, MIMO
systems typically suffer from inter-channel interference, which causes severe
degradation to the system performance. In this context, we propose a novel
optical adaptive precoding (OAP) scheme for the downlink of MIMO VLC systems,
which exploits the knowledge of transmitted symbols to enhance the effective
signal-to-interference-plus-noise ratio. We also derive bit-error-rate
expressions for the OAP under perfect and outdated channel state information
(CSI). Our results demonstrate that the proposed scheme is more robust to both
CSI error and channel correlation, compared to conventional channel inversion
precoding.
Anastasios Tsiamis , Konstantinos Gatsis , George J. Pappas Subjects : Systems and Control (cs.SY) ; Cryptography and Security (cs.CR); Information Theory (cs.IT)
We study the problem of remote state estimation, in the presence of an
eavesdropper. An authorized user estimates the state of a linear plant, based
on the data received from a sensor, while the data may also be intercepted by
the eavesdropper. To maintain confidentiality with respect to state, we
introduce a novel control-theoretic definition of perfect secrecy requiring
that the user’s expected error remains bounded while the eavesdropper’s
expected error grows unbounded. We propose a secrecy mechanism which guarantees
perfect secrecy by randomly withholding sensor information, under the condition
that the user’s packet reception rate is larger than the eavesdropper’s
interception rate. Given this mechanism, we also explore the tradeoff between
user’s utility and confidentiality with respect to the eavesdropper, via an
optimization problem. Finally, some examples are studied to provide insights
about this tradeoff.
Comments: Accepted as an oral presentation in 42th IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP)
Subjects:
Learning (cs.LG)
; Information Theory (cs.IT)
Compressive sensing (CS) is a promising technology for realizing
energy-efficient wireless sensors for long-term health monitoring. However,
conventional model-driven CS frameworks suffer from limited compression ratio
and reconstruction quality when dealing with physiological signals due to
inaccurate models and the overlook of individual variability. In this paper, we
propose a data-driven CS framework that can learn signal characteristics and
personalized features from any individual recording of physiologic signals to
enhance CS performance with a minimized number of measurements. Such
improvements are accomplished by a co-training approach that optimizes the
sensing matrix and the dictionary towards improved restricted isometry property
and signal sparsity, respectively. Experimental results upon ECG signals show
that the proposed method, at a compression ratio of 10x, successfully reduces
the isometry constant of the trained sensing matrices by 86% against random
matrices and improves the overall reconstructed signal-to-noise ratio by 15dB
over conventional model-driven approaches.
微信扫一扫,关注我爱机器学习公众号
微博:我爱机器学习