转载

arXiv Paper Daily: Mon, 6 Feb 2017

Neural and Evolutionary Computing

Robust Particle Swarm Optimizer based on Chemomimicry

Casey Kneale , Karl S. Booksh

Comments: To be revised for formatting and submitted as a Letters style paper

Subjects

:

Neural and Evolutionary Computing (cs.NE)

A particle swarm optimizer (PSO) loosely based on the phenomena of

crystallization and a chaos factor which follows the complimentary error

function is described. The method features three phases: diffusion, directed

motion, and nucleation. During the diffusion phase random walk is the only

contributor to particle motion. As the algorithm progresses the contribution

from chaos decreases and movement toward global best locations is pursued until

convergence has occurred. The algorithm was found to be more robust to local

minima in multimodal test functions than a standard PSO algorithm and is

designed for problems which feature experimental precision.

Eye-Movement behavior identification for AD diagnosis

Juan Biondi , Gerardo Fernandez , Silvia Castro , Osvaldo Agamenonni Subjects : Neural and Evolutionary Computing (cs.NE) ; Neurons and Cognition (q-bio.NC)

In the present work, we develop a deep-learning approach for differentiating

the eye-movement behavior of people with neurodegenerative diseases over

healthy control subjects during reading well-defined sentences. We define an

information compaction of the eye-tracking data of subjects without and with

probable Alzheimer’s disease when reading a set of well-defined, previously

validated, sentences including high-, low-predictable sentences, and proverbs.

Using this information we train a set of denoising sparse-autoencoders and

build a deep neural network with these and a softmax classifier. Our results

are very promising and show that these models may help to understand the

dynamics of eye movement behavior and its relationship with underlying

neuropsychological correlates.

Optimal Experimental Design of Field Trials using Differential Evolution

Vitaliy Feoktistov , Stephane Pietravalle , Nicolas Heslot

Comments: 7 pages, 5 figures

Subjects

:

Neural and Evolutionary Computing (cs.NE)

; Quantitative Methods (q-bio.QM)

When setting up field experiments, to test and compare a range of genotypes

(e.g. maize hybrids), it is important to account for any possible field effect

that may otherwise bias performance estimates of genotypes. To do so, we

propose a model-based method aimed at optimizing the allocation of the tested

genotypes and checks between fields and placement within field, according to

their kinship. This task can be formulated as a combinatorial permutation-based

problem. We used Differential Evolution concept to solve this problem. We then

present results of optimal strategies for between-field and within-field

placements of genotypes and compare them to existing optimization strategies,

both in terms of convergence time and result quality. The new algorithm gives

promising results in terms of convergence and search space exploration.

Structured Attention Networks

Yoon Kim , Carl Denton , Luong Hoang , Alexander M. Rush Subjects : Computation and Language (cs.CL) ; Learning (cs.LG); Neural and Evolutionary Computing (cs.NE)

Attention networks have proven to be an effective approach for embedding

categorical inference within a deep neural network. However, for many tasks we

may want to model richer structural dependencies without abandoning end-to-end

training. In this work, we experiment with incorporating richer structural

distributions, encoded using graphical models, within deep networks. We show

that these structured attention networks are simple extensions of the basic

attention procedure, and that they allow for extending attention beyond the

standard soft-selection approach, such as attending to partial segmentations or

to subtrees. We experiment with two different classes of structured attention

networks: a linear-chain conditional random field and a graph-based parsing

model, and describe how these models can be practically implemented as neural

network layers. Experiments show that this approach is effective for

incorporating structural biases, and structured attention networks outperform

baseline attention models on a variety of synthetic and real tasks: tree

transduction, neural machine translation, question answering, and natural

language inference. We further find that models trained in this way learn

interesting unsupervised hidden representations that generalize simple

attention.

Computer Vision and Pattern Recognition

Joint 2D-3D-Semantic Data for Indoor Scene Understanding

Iro Armeni , Sasha Sax , Amir R. Zamir , Silvio Savarese

Comments: The dataset is available this http URL

Subjects

:

Computer Vision and Pattern Recognition (cs.CV)

; Robotics (cs.RO)

We present a dataset of large-scale indoor spaces that provides a variety of

mutually registered modalities from 2D, 2.5D and 3D domains, with

instance-level semantic and geometric annotations. The dataset covers over

6,000 m2 and contains over 102,000 RGB images, along with the corresponding

depths, surface normals, semantic annotations, global XYZ images (all in forms

of both regular and 360{deg} equirectangular images) as well as camera

information. It also includes registered raw and semantically an- notated 3D

meshes and point clouds. The dataset enables development of joint and

cross-modal learning models and potentially unsupervised approaches utilizing

the regularities present in large-scale indoor spaces. The dataset is available

here: this http URL

Deep Learning with Low Precision by Half-wave Gaussian Quantization

Zhaowei Cai , Xiaodong He , Jian Sun , Nuno Vasconcelos Subjects : Computer Vision and Pattern Recognition (cs.CV) ; Artificial Intelligence (cs.AI); Learning (cs.LG)

The problem of quantizing the activations of a deep neural network is

considered. An examination of the popular binary quantization approach shows

that this consists of approximating a classical non-linearity, the hyperbolic

tangent, by two functions: a piecewise constant sign function, which is used in

feedforward network computations, and a piecewise linear hard tanh function,

used in the backpropagation step during network learning. The problem of

approximating the ReLU non-linearity, widely used in the recent deep learning

literature, is then considered. An half-wave Gaussian quantizer (HWGQ) is

proposed for forward approximation and shown to have efficient implementation,

by exploiting the statistics of of network activations and batch normalization

operations commonly used in the literature. To overcome the problem of gradient

mismatch, due to the use of different forward and backward approximations,

several piece-wise backward approximators are then investigated. The

implementation of the resulting quantized network, denoted as HWGQ-Net, is

shown to achieve much closer performance to full precision networks, such as

AlexNet, ResNet, GoogLeNet and VGG-Net, than previously available low-precision

networks, with 1-bit binary weights and 2-bit quantized activations.

A method of limiting performance loss of CNNs in noisy environments

James R. Geraci , Parichay Kapoor Subjects : Computer Vision and Pattern Recognition (cs.CV)

Convolutional Neural Network (CNN) recognition rates drop in the presence of

noise. We demonstrate a novel method of counteracting this drop in recognition

rate by adjusting the biases of the neurons in the convolutional layers

according to the noise conditions encountered at runtime. We compare our

technique to training one network for all possible noise levels, dehazing via

preprocessing a signal with a denoising autoencoder, and training a network

specifically for each noise level. Our system compares favorably in terms of

robustness, computational complexity and recognition rate.

FCSS: Fully Convolutional Self-Similarity for Dense Semantic Correspondence

Seungryong Kim , Dongbo Min , Bumsub Ham , Sangryul Jeon , Stephen Lin , Kwanghoon Sohn Subjects : Computer Vision and Pattern Recognition (cs.CV)

We present a descriptor, called fully convolutional self-similarity (FCSS),

for dense semantic correspondence. To robustly match points among different

instances within the same object class, we formulate FCSS using local

self-similarity (LSS) within a fully convolutional network. In contrast to

existing CNN-based descriptors, FCSS is inherently insensitive to intra-class

appearance variations because of its LSS-based structure, while maintaining the

precise localization ability of deep neural networks. The sampling patterns of

local structure and the self-similarity measure are jointly learned within the

proposed network in an end-to-end and multi-scale manner. As training data for

semantic correspondence is rather limited, we propose to leverage object

candidate priors provided in existing image datasets and also correspondence

consistency between object pairs to enable weakly-supervised learning.

Experiments demonstrate that FCSS outperforms conventional handcrafted

descriptors and CNN-based descriptors on various benchmarks.

Seeded Laplaican: An Eigenfunction Solution for Scribble Based Interactive Image Segmentation

Ahmed Taha , Marwan Torki Subjects : Computer Vision and Pattern Recognition (cs.CV)

In this paper, we cast the scribble-based interactive image segmentation as a

semi-supervised learning problem. Our novel approach alleviates the need to

solve an expensive generalized eigenvector problem by approximating the

eigenvectors using efficiently computed eigenfunctions. The smoothness operator

defined on feature densities at the limit n tends to infinity recovers the

exact eigenvectors of the graph Laplacian, where n is the number of nodes in

the graph. To further reduce the computational complexity without scarifying

our accuracy, we select pivots pixels from user annotations. In our

experiments, we evaluate our approach using both human scribble and “robot

user” annotations to guide the foreground/background segmentation. We developed

a new unbiased collection of five annotated images datasets to standardize the

evaluation procedure for any scribble-based segmentation method. We

experimented with several variations, including different feature vectors,

pivot count and the number of eigenvectors. Experiments are carried out on

datasets that contain a wide variety of natural images. We achieve better

qualitative and quantitative results compared to state-of-the-art interactive

segmentation algorithms.

Deep Learning For Video Saliency Detection

Wenguan Wang , Jianbing Shen , Ling Shao Subjects : Computer Vision and Pattern Recognition (cs.CV)

This paper proposes a deep learning model to efficiently detect salient

regions in videos. It addresses two important issues: (1) deep video saliency

model training with the absence of sufficiently large and pixel-wise annotated

video data; and (2) fast video saliency training and detection. The proposed

deep video saliency network consists of two modules, for capturing the spatial

and temporal saliency stimuli, respectively. The dynamic saliency model,

explicitly incorporating saliency estimates from the static saliency model,

directly produces spatiotemporal saliency inference without time-consuming

optical flow computation. We further propose a novel data augmentation

technique that simulates video training data from existing annotated image

datasets, which enables our network to learn diverse saliency stimuli and

prevents overfitting with the limited number of training videos. Leveraging our

synthetic video data (150K video sequences) and real videos, our deep video

saliency model successfully learns both spatial and temporal saliency stimuli,

thus producing accurate spatiotemporal saliency estimate. We advance the

state-of-the-art on the DAVIS dataset (MAE of .06) and the FBMS dataset (MAE of

.07), and do so with much improved speed (2fps with all steps) on one GPU.

YouTube-BoundingBoxes: A Large High-Precision Human-Annotated Data Set for Object Detection in Video

Esteban Real , Jonathon Shlens , Stefano Mazzocchi , Xin Pan , Vincent Vanhoucke Subjects : Computer Vision and Pattern Recognition (cs.CV)

We introduce a new large-scale data set of video URLs with densely-sampled

object bounding box annotations called YouTube-BoundingBoxes (YT-BB). The data

set consists of approximately 380,000 video segments about 19s long,

automatically selected to feature objects in natural settings without editing

or post-processing, with a recording quality often akin to that of a hand-held

cell phone camera. The objects represent a subset of the MS COCO label set. All

video segments were human-annotated with high-precision classification labels

and bounding boxes at 1 frame per second. The use of a cascade of increasingly

precise human annotations ensures a label accuracy above 95% for every class

and tight bounding boxes. Finally, we train and evaluate well-known deep

network architectures and report baseline figures for per-frame classification

and localization to provide a point of comparison for future work. We also

demonstrate how the temporal contiguity of video can potentially be used to

improve such inferences. The data set can be found at

this https URL We hope the availability of such large

curated corpus will spur new advances in video object detection and tracking.

Intrinsic Grassmann Averages for Online Linear and Robust Subspace Learning

Rudrasis Chakraborty , Søren Hauberg , Baba C. Vemuri Subjects : Learning (cs.LG) ; Computer Vision and Pattern Recognition (cs.CV)

Principal Component Analysis (PCA) is a fundamental method for estimating a

linear subspace approximation to high-dimensional data. Many algorithms exist

in literature to achieve a statistically robust version of PCA called RPCA. In

this paper, we present a geometric framework for computing the principal linear

subspaces in both situations that amounts to computing the intrinsic average on

the space of all subspaces (the Grassmann manifold). Points on this manifold

are defined as the subspaces spanned by (K)-tuples of observations. We show

that the intrinsic Grassmann average of these subspaces coincide with the

principal components of the observations when they are drawn from a Gaussian

distribution. Similar results are also shown to hold for the RPCA. Further, we

propose an efficient online algorithm to do subspace averaging which is of

linear complexity in terms of number of samples and has a linear convergence

rate. When the data has outliers, our proposed online robust subspace averaging

algorithm shows significant performance (accuracy and computation time) gain

over a recently published RPCA methods with publicly accessible code. We have

demonstrated competitive performance of our proposed online subspace algorithm

method on one synthetic and two real data sets. Experimental results depicting

stability of our proposed method are also presented. Furthermore, on two real

outlier corrupted datasets, we present comparison experiments showing lower

reconstruction error using our online RPCA algorithm. In terms of

reconstruction error and time required, both our algorithms outperform the

competition.

Artificial Intelligence

The Value of Inferring the Internal State of Traffic Participants for Autonomous Freeway Driving

Zachary Sunberg , Christopher Ho , Mykel Kochenderfer Subjects : Artificial Intelligence (cs.AI)

Safe interaction with human drivers is one of the primary challenges for

autonomous vehicles. In order to plan driving maneuvers effectively, the

vehicle’s control system must infer and predict how humans will behave based on

their latent internal state (e.g., intentions and aggressiveness). This

research uses a simple model for human behavior with unknown parameters that

make up the internal states of the traffic participants and presents a method

for quantifying the value of estimating these states and planning with their

uncertainty explicitly modeled. An upper performance bound is established by an

omniscient Monte Carlo Tree Search (MCTS) planner that has perfect knowledge of

the internal states. A baseline lower bound is established by planning with

MCTS assuming that all drivers have the same internal state. MCTS variants are

then used to solve a partially observable Markov decision process (POMDP) that

models the internal state uncertainty to determine whether inferring the

internal state offers an advantage over the baseline. Applying this method to a

freeway lane changing scenario reveals that there is a significant performance

gap between the upper bound and baseline. POMDP planning techniques come close

to closing this gap, especially when important hidden model parameters are

correlated with measurable parameters.

On Robustness in Multilayer Interdependent Network

Joydeep Banerjee , Chenyang Zhou , Arunabha Sen

Comments: CRITIS 2015

Subjects

:

Networking and Internet Architecture (cs.NI)

; Artificial Intelligence (cs.AI)

Critical Infrastructures like power and communication networks are highly

interdependent on each other for their full functionality. Many significant

research have been pursued to model the interdependency and failure analysis of

these interdependent networks. However, most of these models fail to capture

the complex interdependencies that might actually exist between the

infrastructures. The emph{Implicative Interdependency Model} that utilizes

Boolean Logic to capture complex interdependencies was recently proposed which

overcome the limitations of the existing models. A number of problems were

studies based on this model. In this paper we study the extit{Robustness}

problem in Interdependent Power and Communication Network. The robustness is

defined with respect to two parameters (K in I^{+} cup {0}) and (

ho in

(0,1]). We utilized the emph{Implicative Interdependency Model} model to

capture the complex interdependency between the two networks. The model

classifies the interdependency relations into four cases. Computational

complexity of the problem is analyzed for each of these cases. A polynomial

time algorithm is designed for the first case that outputs the optimal

solution. All the other cases are proved to be NP-complete. An

in-approximability bound is provided for the third case. For the general case

we formulate an Integer Linear Program to get the optimal solution and a

polynomial time heuristic. The applicability of the heuristic is evaluated

using power and communication network data of Maricopa County, Arizona. The

experimental results showed that the heuristic almost always produced near

optimal value of parameter (K) for (

ho < 0.42).

Deep Learning with Low Precision by Half-wave Gaussian Quantization

Zhaowei Cai , Xiaodong He , Jian Sun , Nuno Vasconcelos Subjects : Computer Vision and Pattern Recognition (cs.CV) ; Artificial Intelligence (cs.AI); Learning (cs.LG)

The problem of quantizing the activations of a deep neural network is

considered. An examination of the popular binary quantization approach shows

that this consists of approximating a classical non-linearity, the hyperbolic

tangent, by two functions: a piecewise constant sign function, which is used in

feedforward network computations, and a piecewise linear hard tanh function,

used in the backpropagation step during network learning. The problem of

approximating the ReLU non-linearity, widely used in the recent deep learning

literature, is then considered. An half-wave Gaussian quantizer (HWGQ) is

proposed for forward approximation and shown to have efficient implementation,

by exploiting the statistics of of network activations and batch normalization

operations commonly used in the literature. To overcome the problem of gradient

mismatch, due to the use of different forward and backward approximations,

several piece-wise backward approximators are then investigated. The

implementation of the resulting quantized network, denoted as HWGQ-Net, is

shown to achieve much closer performance to full precision networks, such as

AlexNet, ResNet, GoogLeNet and VGG-Net, than previously available low-precision

networks, with 1-bit binary weights and 2-bit quantized activations.

Information Retrieval

Tempas: Temporal Archive Search Based on Tags

Helge Holzmann , Avishek Anand

Comments: WWW 2016, Montreal, Quebec, Canada

Subjects

:

Information Retrieval (cs.IR)

Limited search and access patterns over Web archives have been well

documented. One of the key reasons is the lack of understanding of the user

access patterns over such collections, which in turn is attributed to the lack

of effective search interfaces. Current search interfaces for Web archives are

(a) either purely navigational or (b) have sub-optimal search experience due to

ineffective retrieval models or query modeling. We identify that external

longitudinal resources, such as social bookmarking data, are crucial sources to

identify important and popular websites in the past. To this extent we present

Tempas, a tag-based temporal search engine for Web archives.

Websites are posted at specific times of interest on several external

platforms, such as bookmarking sites like Delicious. Attached tags not only act

as relevant descriptors useful for retrieval, but also encode the time of

relevance. With Tempas we tackle the challenge of temporally searching a Web

archive by indexing tags and time. We allow temporal selections for search

terms, rank documents based on their popularity and also provide meaningful

query recommendations by exploiting tag-tag and tag-document co-occurrence

statistics in arbitrary time windows. Finally, Tempas operates as a fairly

non-invasive indexing framework. By not dealing with contents from the actual

Web archive it constitutes an attractive and low-overhead approach for quick

access into Web archives.

Semi-Supervised Spam Detection in Twitter Stream

Surendra Sedhai , Aixin Sun

Comments: 9

Subjects

:

Information Retrieval (cs.IR)

; Cryptography and Security (cs.CR); Social and Information Networks (cs.SI)

Most existing techniques for spam detection on Twitter aim to identify and

block users who post spam tweets. In this paper, we propose a Semi-Supervised

Spam Detection (S3D) framework for spam detection at tweet-level. The proposed

framework consists of two main modules: spam detection module operating in

real-time mode, and model update module operating in batch mode. The spam

detection module consists of four light-weight detectors: (i) blacklisted

domain detector to label tweets containing blacklisted URLs, (ii)

near-duplicate detector to label tweets that are near-duplicates of confidently

pre-labeled tweets, (iii) reliable ham detector to label tweets that are posted

by trusted users and that do not contain spammy words, and (iv)

multi-classifier based detector labels the remaining tweets. The information

required by the detection module are updated in batch mode based on the tweets

that are labeled in the previous time window. Experiments on a large scale

dataset show that the framework adaptively learns patterns of new spam

activities and maintain good accuracy for spam detection in a tweet stream.

ReLiC: Entity Profiling by using Random Forest and Trustworthiness of a Source – Technical Report

Shubham Varma , Neyshith Sameer , C. Ravindranath Chowdary Subjects : Information Retrieval (cs.IR) ; Databases (cs.DB)

The digital revolution has brought most of the world on the world wide web.

The data available on WWW has increased many folds in the past decade. Social

networks, online clubs and organisations have come into existence. Information

is extracted from these venues about a real world entity like a person,

organisation, event, etc. However, this information may change over time, and

there is a need for the sources to be up-to-date. Therefore, it is desirable to

have a model to extract relevant data items from different sources and merge

them to build a complete profile of an entity (entity profiling). Further, this

model should be able to handle incorrect or obsolete data items. In this paper,

we propose a novel method for completing a profile. We have developed a two

phase method-1) The first phase (resolution phase) links records to the

queries. We have proposed and observed that the use of random forest for entity

resolution increases the performance of the system as this has resulted in more

records getting linked to the correct entity. Also, we used trustworthiness of

a source as a feature to the random forest. 2) The second phase selects the

appropriate values from records to complete a profile based on our proposed

selection criteria. We have used various metrics for measuring the performance

of the resolution phase as well as for the overall ReLiC framework. It is

established through our results that the use of biased sources has

significantly improved the performance of the ReLiC framework. Experimental

results show that our proposed system, ReLiC outperforms the state-of-the-art.

Neural Feature Embedding for User Response Prediction in Real-Time Bidding (RTB)

Enno Shioji , Masayuki Arai Subjects : Information Retrieval (cs.IR)

In the area of ad-targeting, predicting user responses is essential for many

applications such as Real-Time Bidding (RTB). Many of the features available in

this domain are sparse categorical features. This presents a challenge

especially when the user responses to be predicted is rare, because each

feature will only have very few positive examples. Recently, neural embedding

techniques such as word2vec which learn distributed representations of words

using occurrence statistics in the corpus have been shown to be effective in

many Natural Language Processing tasks. In this paper, we use real-world data

set to show that a similar technique can be used to learn distributed

representations of features from users’ web history, and that such

representations can be used to improve the accuracy of commonly used models for

predicting rare user responses.

Multi-level computational methods for interdisciplinary research in the HathiTrust Digital Library

Jaimie Murdock , Colin Allen , Katy Börner , Robert Light , Simon McAlister , Robert Rose , Doori Rose , Jun Otsuka , David Bourget , John Lawrence , Andrew Ravenscroft , Chris Reed

Comments: 23 pages, 3 figures

Subjects

:

Digital Libraries (cs.DL)

; Computation and Language (cs.CL); Information Retrieval (cs.IR)

We show how faceted search using a combination of traditional classification

systems and mixed-membership models can move beyond keyword search to inform

resource discovery, hypothesis formulation, and argument extraction for

interdisciplinary research. Our test domain is the history and philosophy of

scientific work on animal mind and cognition. We demonstrate an application of

our methods to the problem of identifying and extracting arguments about

anthropomorphism during a critical period in the development of comparative

psychology. We show how a combination of classification systems and

mixed-membership models trained over large digital libraries can inform

resource discovery in this domain, using methods that can be generalized to

other interdisciplinary research questions. Through a novel approach of

drill-down topic modeling, we are able to reduce a collection of 1,315 fulltext

volumes to 6 focal volumes that did not appear in the first ten search results

in the HathiTrust digital library. This ultimately supports a system for

semi-automatic identification of argument structures to augment the kind of

“close reading” that leads to novel interpretations at the heart of scholarly

work in the humanities, drilling down from massive quantities of text to very

specific passages. This multi-level view advances understanding of the

intellectual and societal contexts in which writings are interpreted.

Topic Modeling the Hàn diăn Ancient Classics

Colin Allen , Hongliang Luo , Jaimie Murdock , Jianghuai Pu , Xiaohong Wang , Yanjie Zhai , Kun Zhao

Comments: 24 pages; 14 pages supplemental

Subjects

:

Computation and Language (cs.CL)

; Computers and Society (cs.CY); Digital Libraries (cs.DL); Human-Computer Interaction (cs.HC); Information Retrieval (cs.IR)

Ancient Chinese texts present an area of enormous challenge and opportunity

for humanities scholars interested in exploiting computational methods to

assist in the development of new insights and interpretations of culturally

significant materials. In this paper we describe a collaborative effort between

Indiana University and Xi’an Jiaotong University to support exploration and

interpretation of a digital corpus of over 18,000 ancient Chinese documents,

which we refer to as the “Handian” ancient classics corpus (H`an diu{a}n

gu{u} j’i, i.e, the “Han canon” or “Chinese classics”). It contains classics

of ancient Chinese philosophy, documents of historical and biographical

significance, and literary works. We begin by describing the Digital Humanities

context of this joint project, and the advances in humanities computing that

made this project feasible. We describe the corpus and introduce our

application of probabilistic topic modeling to this corpus, with attention to

the particular challenges posed by modeling ancient Chinese documents. We give

a specific example of how the software we have developed can be used to aid

discovery and interpretation of themes in the corpus. We outline more advanced

forms of computer-aided interpretation that are also made possible by the

programming interface provided by our system, and the general implications of

these methods for understanding the nature of meaning in these texts.

Computation and Language

Multilingual Multi-modal Embeddings for Natural Language Processing

Iacer Calixto , Qun Liu , Nick Campbell

Comments: 4 pages (5 including references), no figures

Subjects

:

Computation and Language (cs.CL)

We propose a novel discriminative model that learns embeddings from

multilingual and multi-modal data, meaning that our model can take advantage of

images and descriptions in multiple languages to improve embedding quality. To

that end, we introduce a modification of a pairwise contrastive estimation

optimisation function as our training objective. We evaluate our embeddings on

an image-sentence ranking (ISR), a semantic textual similarity (STS), and a

neural machine translation (NMT) task. We find that the additional multilingual

signals lead to improvements on both the ISR and STS tasks, and the

discriminative cost can also be used in re-ranking (n)-best lists produced by

NMT models, yielding strong improvements.

Automatic Prediction of Discourse Connectives

Eric Malmi , Daniele Pighin , Sebastian Krause , Mikhail Kozhevnikov

Comments: 9 pages

Subjects

:

Computation and Language (cs.CL)

Accurate prediction of suitable discourse connectives (however, furthermore,

etc.) is a key component of any system aimed at building coherent and fluent

discourses from shorter sentences and passages. As an example, a dialog system

might assemble a long and informative answer by sampling passages extracted

from different documents retrieved from the web. We formulate the task of

discourse connective prediction and release a dataset of 2.9M sentence pairs

separated by discourse connectives for this task. Then, we evaluate the

hardness of the task for human raters, apply a recently proposed decomposable

attention (DA) model to this task and observe that the automatic predictor has

a higher F1 than human raters (32 vs. 30). Nevertheless, under specific

conditions the raters still outperform the DA model, suggesting that there is

headroom for future improvements. Finally, we further demonstrate the

usefulness of the connectives dataset by showing that it improves implicit

discourse relation prediction when used for model pre-training.

Structured Attention Networks

Yoon Kim , Carl Denton , Luong Hoang , Alexander M. Rush Subjects : Computation and Language (cs.CL) ; Learning (cs.LG); Neural and Evolutionary Computing (cs.NE)

Attention networks have proven to be an effective approach for embedding

categorical inference within a deep neural network. However, for many tasks we

may want to model richer structural dependencies without abandoning end-to-end

training. In this work, we experiment with incorporating richer structural

distributions, encoded using graphical models, within deep networks. We show

that these structured attention networks are simple extensions of the basic

attention procedure, and that they allow for extending attention beyond the

standard soft-selection approach, such as attending to partial segmentations or

to subtrees. We experiment with two different classes of structured attention

networks: a linear-chain conditional random field and a graph-based parsing

model, and describe how these models can be practically implemented as neural

network layers. Experiments show that this approach is effective for

incorporating structural biases, and structured attention networks outperform

baseline attention models on a variety of synthetic and real tasks: tree

transduction, neural machine translation, question answering, and natural

language inference. We further find that models trained in this way learn

interesting unsupervised hidden representations that generalize simple

attention.

Topic Modeling the Hàn diăn Ancient Classics

Colin Allen , Hongliang Luo , Jaimie Murdock , Jianghuai Pu , Xiaohong Wang , Yanjie Zhai , Kun Zhao

Comments: 24 pages; 14 pages supplemental

Subjects

:

Computation and Language (cs.CL)

; Computers and Society (cs.CY); Digital Libraries (cs.DL); Human-Computer Interaction (cs.HC); Information Retrieval (cs.IR)

Ancient Chinese texts present an area of enormous challenge and opportunity

for humanities scholars interested in exploiting computational methods to

assist in the development of new insights and interpretations of culturally

significant materials. In this paper we describe a collaborative effort between

Indiana University and Xi’an Jiaotong University to support exploration and

interpretation of a digital corpus of over 18,000 ancient Chinese documents,

which we refer to as the “Handian” ancient classics corpus (H`an diu{a}n

gu{u} j’i, i.e, the “Han canon” or “Chinese classics”). It contains classics

of ancient Chinese philosophy, documents of historical and biographical

significance, and literary works. We begin by describing the Digital Humanities

context of this joint project, and the advances in humanities computing that

made this project feasible. We describe the corpus and introduce our

application of probabilistic topic modeling to this corpus, with attention to

the particular challenges posed by modeling ancient Chinese documents. We give

a specific example of how the software we have developed can be used to aid

discovery and interpretation of themes in the corpus. We outline more advanced

forms of computer-aided interpretation that are also made possible by the

programming interface provided by our system, and the general implications of

these methods for understanding the nature of meaning in these texts.

Multi-level computational methods for interdisciplinary research in the HathiTrust Digital Library

Jaimie Murdock , Colin Allen , Katy Börner , Robert Light , Simon McAlister , Robert Rose , Doori Rose , Jun Otsuka , David Bourget , John Lawrence , Andrew Ravenscroft , Chris Reed

Comments: 23 pages, 3 figures

Subjects

:

Digital Libraries (cs.DL)

; Computation and Language (cs.CL); Information Retrieval (cs.IR)

We show how faceted search using a combination of traditional classification

systems and mixed-membership models can move beyond keyword search to inform

resource discovery, hypothesis formulation, and argument extraction for

interdisciplinary research. Our test domain is the history and philosophy of

scientific work on animal mind and cognition. We demonstrate an application of

our methods to the problem of identifying and extracting arguments about

anthropomorphism during a critical period in the development of comparative

psychology. We show how a combination of classification systems and

mixed-membership models trained over large digital libraries can inform

resource discovery in this domain, using methods that can be generalized to

other interdisciplinary research questions. Through a novel approach of

drill-down topic modeling, we are able to reduce a collection of 1,315 fulltext

volumes to 6 focal volumes that did not appear in the first ten search results

in the HathiTrust digital library. This ultimately supports a system for

semi-automatic identification of argument structures to augment the kind of

“close reading” that leads to novel interpretations at the heart of scholarly

work in the humanities, drilling down from massive quantities of text to very

specific passages. This multi-level view advances understanding of the

intellectual and societal contexts in which writings are interpreted.

KU-ISPL Speaker Recognition Systems under Language mismatch condition for NIST 2016 Speaker Recognition Evaluation

Suwon Shon , Hanseok Ko

Comments: SRE16, NIST SRE 2016 system description

Subjects

:

Sound (cs.SD)

; Computation and Language (cs.CL)

Korea University Intelligent Signal Processing Lab. (KU-ISPL) developed

speaker recognition system for SRE16 fixed training condition. Data for

evaluation trials are collected from outside North America, spoken in Tagalog

and Cantonese while training data only is spoken English. Thus, main issue for

SRE16 is compensating the discrepancy between different languages. As

development dataset which is spoken in Cebuano and Mandarin, we could prepare

the evaluation trials through preliminary experiments to compensate the

language mismatched condition. Our team developed 4 different approaches to

extract i-vectors and applied state-of-the-art techniques as backend. To

compensate language mismatch, we investigated and endeavored unique method such

as unsupervised language clustering, inter language variability compensation

and gender/language dependent score normalization.

Distributed, Parallel, and Cluster Computing

Distributed Optimization Using the Primal-Dual Method of Multipliers

G. Zhang , R. Heusdens Subjects : Distributed, Parallel, and Cluster Computing (cs.DC) ; Optimization and Control (math.OC)

In this paper, we propose the primal-dual method of multipliers (PDMM) for

distributed optimization over a graph. In particular, we optimize a sum of

convex functions defined over a graph, where every edge in the graph carries a

linear equality constraint. In designing the new algorithm, an augmented

primal-dual Lagrangian function is constructed which smoothly captures the

graph topology. It is shown that a saddle point of the constructed function

provides an optimal solution of the original problem. Further under both the

synchronous and asynchronous updating schemes, PDMM has the convergence rate of

O(1/K) (where K denotes the iteration index) for general closed, proper and

convex functions. Other properties of PDMM such as convergence speeds versus

different parameter- settings and resilience to transmission failure are also

investigated through the experiments of distributed averaging.

Distributed Approximation Algorithms for the Multiple Knapsack Problem

Ananth Murthy , Chandan Yeshwanth , Shrisha Rao

Comments: 18 pages

Subjects

:

Data Structures and Algorithms (cs.DS)

; Distributed, Parallel, and Cluster Computing (cs.DC); Discrete Mathematics (cs.DM)

We consider the distributed version of the Multiple Knapsack Problem (MKP),

where (m) items are to be distributed amongst (n) processors, each with a

knapsack. We propose different distributed approximation algorithms with a

tradeoff between time and message complexities. The algorithms are based on the

greedy approach of assigning the best item to the knapsack with the largest

capacity. These algorithms obtain a solution with a bound of (frac{1}{n+1})

times the optimum solution, with either (mathcal{O}left(mlog n

ight)) time

and (mathcal{O}left(m n

ight)) messages, or (mathcal{O}left(m

ight)) time

and (mathcal{O}left(mn^{2}

ight)) messages.

Learning

Intrinsic Grassmann Averages for Online Linear and Robust Subspace Learning

Rudrasis Chakraborty , Søren Hauberg , Baba C. Vemuri Subjects : Learning (cs.LG) ; Computer Vision and Pattern Recognition (cs.CV)

Principal Component Analysis (PCA) is a fundamental method for estimating a

linear subspace approximation to high-dimensional data. Many algorithms exist

in literature to achieve a statistically robust version of PCA called RPCA. In

this paper, we present a geometric framework for computing the principal linear

subspaces in both situations that amounts to computing the intrinsic average on

the space of all subspaces (the Grassmann manifold). Points on this manifold

are defined as the subspaces spanned by (K)-tuples of observations. We show

that the intrinsic Grassmann average of these subspaces coincide with the

principal components of the observations when they are drawn from a Gaussian

distribution. Similar results are also shown to hold for the RPCA. Further, we

propose an efficient online algorithm to do subspace averaging which is of

linear complexity in terms of number of samples and has a linear convergence

rate. When the data has outliers, our proposed online robust subspace averaging

algorithm shows significant performance (accuracy and computation time) gain

over a recently published RPCA methods with publicly accessible code. We have

demonstrated competitive performance of our proposed online subspace algorithm

method on one synthetic and two real data sets. Experimental results depicting

stability of our proposed method are also presented. Furthermore, on two real

outlier corrupted datasets, we present comparison experiments showing lower

reconstruction error using our online RPCA algorithm. In terms of

reconstruction error and time required, both our algorithms outperform the

competition.

Deep Learning with Low Precision by Half-wave Gaussian Quantization

Zhaowei Cai , Xiaodong He , Jian Sun , Nuno Vasconcelos Subjects : Computer Vision and Pattern Recognition (cs.CV) ; Artificial Intelligence (cs.AI); Learning (cs.LG)

The problem of quantizing the activations of a deep neural network is

considered. An examination of the popular binary quantization approach shows

that this consists of approximating a classical non-linearity, the hyperbolic

tangent, by two functions: a piecewise constant sign function, which is used in

feedforward network computations, and a piecewise linear hard tanh function,

used in the backpropagation step during network learning. The problem of

approximating the ReLU non-linearity, widely used in the recent deep learning

literature, is then considered. An half-wave Gaussian quantizer (HWGQ) is

proposed for forward approximation and shown to have efficient implementation,

by exploiting the statistics of of network activations and batch normalization

operations commonly used in the literature. To overcome the problem of gradient

mismatch, due to the use of different forward and backward approximations,

several piece-wise backward approximators are then investigated. The

implementation of the resulting quantized network, denoted as HWGQ-Net, is

shown to achieve much closer performance to full precision networks, such as

AlexNet, ResNet, GoogLeNet and VGG-Net, than previously available low-precision

networks, with 1-bit binary weights and 2-bit quantized activations.

Structured Attention Networks

Yoon Kim , Carl Denton , Luong Hoang , Alexander M. Rush Subjects : Computation and Language (cs.CL) ; Learning (cs.LG); Neural and Evolutionary Computing (cs.NE)

Attention networks have proven to be an effective approach for embedding

categorical inference within a deep neural network. However, for many tasks we

may want to model richer structural dependencies without abandoning end-to-end

training. In this work, we experiment with incorporating richer structural

distributions, encoded using graphical models, within deep networks. We show

that these structured attention networks are simple extensions of the basic

attention procedure, and that they allow for extending attention beyond the

standard soft-selection approach, such as attending to partial segmentations or

to subtrees. We experiment with two different classes of structured attention

networks: a linear-chain conditional random field and a graph-based parsing

model, and describe how these models can be practically implemented as neural

network layers. Experiments show that this approach is effective for

incorporating structural biases, and structured attention networks outperform

baseline attention models on a variety of synthetic and real tasks: tree

transduction, neural machine translation, question answering, and natural

language inference. We further find that models trained in this way learn

interesting unsupervised hidden representations that generalize simple

attention.

Recurrent Neural Networks for anomaly detection in the Post-Mortem time series of LHC superconducting magnets

Maciej Wielgosz , Andrzej Skoczeń , Matej Mertik

Comments: Related to arxiv: 1611.06241

Subjects

:

Instrumentation and Detectors (physics.ins-det)

; Learning (cs.LG); Accelerator Physics (physics.acc-ph)

This paper presents a model based on Deep Learning algorithms of LSTM and GRU

for facilitating an anomaly detection in Large Hadron Collider superconducting

magnets. We used high resolution data available in Post Mortem database to

train a set of models and chose the best possible set of their

hyper-parameters. Using Deep Learning approach allowed to examine a vast body

of data and extract the fragments which require further experts examination and

are regarded as anomalies. The presented method does not require tedious manual

threshold setting and operator attention at the stage of the system setup.

Instead, the automatic approach is proposed, which achieves according to our

experiments accuracy of 99%. This is reached for the largest dataset of 302 MB

and the following architecture of the network: single layer LSTM, 128 cells, 20

epochs of training, look_back=16, look_ahead=128, grid=100 and optimizer Adam.

All the experiments were run on GPU Nvidia Tesla K80

An Introduction to Machine Learning Communications Systems

Timothy J. O'Shea , Jakob Hoydis

Comments: 10 pages, 8 figures, 5 tables, under concurrent academic journal submission

Subjects

:

Information Theory (cs.IT)

; Learning (cs.LG); Networking and Internet Architecture (cs.NI)

We introduce and motivate machine learning (ML) communications systems that

aim to improve on and to even replace the vast expert knowledge in the field of

communications using modern machine learning techniques. These have recently

achieved breakthroughs in many different domains, but not yet in

communications. By interpreting a communications system as an autoencoder, we

develop a fundamental new way to think about radio communications system design

as an end-to-end reconstruction optimization task that seeks to jointly

optimize transmitter and receiver components in a single process. We further

present the concept of Radio Transformer Networks (RTNs) as a means to

incorporate expert domain knowledge in the ML model and study the application

of convolutional neural networks (CNNs) on raw IQ time-series data for

modulation classification. We conclude the paper with a deep discussion of open

challenges and areas for future investigation.

Skip Connections as Effective Symmetry-Breaking

A. Emin Orhan

Comments: 18 pages, 12 figures, 1 supplementary figure

Subjects

:

Neural and Evolutionary Computing (cs.NE)

; Learning (cs.LG)

Skip connections made the training of very deep neural networks possible and

have become an indispendable component in a variety of neural architectures. A

completely satisfactory explanation for their success remains elusive. Here, we

present a novel explanation for the benefits of skip connections in training

very deep neural networks. We argue that skip connections help break symmetries

inherent in the loss landscapes of deep networks, leading to drastically

simplified landscapes. In particular, skip connections between adjacent layers

in a multilayer network break the permutation symmetry of nodes in a given

layer, and the recently proposed DenseNet architecture, where each layer

projects skip connections to every layer above it, also breaks the rescaling

symmetry of connectivity matrices between different layers. This hypothesis is

supported by evidence from a toy model with binary weights and from experiments

with fully-connected networks suggesting (i) that skip connections do not

necessarily improve training unless they help break symmetries and (ii) that

alternative ways of breaking the symmetries also lead to significant

performance improvements in training deep networks, hence there is nothing

special about skip connections in this respect. We find, however, that skip

connections confer additional benefits over and above symmetry-breaking, such

as the ability to deal effectively with the vanishing gradients problem.

An Impossibility Result for Reconstruction in a Degree-Corrected Planted-Partition Model

Lennart Gulikers , Marc Lelarge , Laurent Massoulié

Comments: Made some simplifications

Subjects

:

Probability (math.PR)

; Learning (cs.LG); Social and Information Networks (cs.SI); Machine Learning (stat.ML)

We consider a Degree-Corrected Planted-Partition model: a random graph on (n)

nodes with two asymptotically equal-sized clusters. The model parameters are

two constants (a,b > 0) and an i.i.d. sequence of weights ((phi_u)_{u=1}^n),

with finite second moment (Phi^{(2)}). Vertices (u) and (v) are joined by an

edge with probability (frac{phi_u phi_v}{n}a) when they are in the same

class and with probability (frac{phi_u phi_v}{n}b) otherwise.

We prove that it is information-theoretically impossible to estimate the

spins in a way positively correlated with the true community structure when

((a-b)^2 Phi^{(2)} leq 2(a+b)).

A by-product of our proof is a precise coupling-result for

local-neighbourhoods in Degree-Corrected Planted-Partition models, which could

be of independent interest.

Information Theory

Polar Codes and Polar Lattices for the Heegard-Berger Problem

Jinwen Shi , Ling Liu , Deniz Gündüz , Cong Ling Subjects : Information Theory (cs.IT)

Explicit coding schemes are proposed to achieve the rate-distortion bound for

the Heegard-Berger problem using polar codes. Specifically, a nested polar code

construction is employed to achieve the rate-distortion bound for the binary

case. The nested structure contains two optimal polar codes for lossy source

coding and channel coding, respectively. Moreover, a similar nested polar

lattice construction is employed for the Gaussian case. The proposed polar

lattice is constructed by nesting a quantization polar lattice and an AWGN

capacity achieving polar lattice.

Stability and Instability Conditions for Slotted Aloha with Exponential Backoff

Luca Barletta , Flaminio Borgonovo

Comments: 22 pages, 1 figure. Submitted to the IEEE Trans. on Information Theory

Subjects

:

Information Theory (cs.IT)

This paper provides stability and instability conditions for slotted Aloha

under the exponential backoff (EB) model with geometric law (imapsto

b^{-i-i_0}), when transmission buffers are in saturation, i.e., always full. In

particular, we prove that for any number of users and for (b>1) the system is:

(i) ergodic for (i_0 >1), (ii) null recurrent for (0<i_0le 1), and (iii)

transient for (i_0=0). Furthermore, when referring to a system with queues and

Poisson arrivals, the system is shown to be stable whenever EB in saturation is

stable with throughput (lambda_0) and the system input rate is upper-bounded

as (lambda<lambda_0).

Relay Selection in Cooperative Power Line Communication: A Multi-Armed Bandit Approach

Babak Nikfar , A. J. Han Vinck Subjects : Information Theory (cs.IT)

Power line communication (PLC) exploits the existence of installed

infrastructure of power delivery system, in order to transmit data over power

lines. In PLC networks, different nodes of the network are interconnected via

power delivery transmission lines, and the data signal is flowing between them.

However, the attenuation and the harsh environment of the power line

communication channels, makes it difficult to establish a reliable

communication between two nodes of the network which are separated by a long

distance. Relaying and cooperative communication has been used to overcome this

problem. In this paper a two-hop cooperative PLC has been studied, where the

data is communicated between a transmitter and a receiver node, through a

single array node which has to be selected from a set of available arrays. The

relay selection problem can be solved by having channel state information (CSI)

at transmitter and selecting the relay which results in the best performance.

However, acquiring the channel state information at transmitter increases the

complexity of the communication system and introduces undesired overhead to the

system. We propose a class of machine learning schemes, namely multi-armed

bandit (MAB), to solve the relay selection problem without the knowledge of the

channel at the transmitter. Furthermore, we develop a new MAB algorithm which

exploits the periodicity of the synchronous impulsive noise of the PLC channel,

in order to improve the relay selection algorithm.

Stochastic Joint Radio and Computational Resource Management for Multi-User Mobile-Edge Computing Systems

Yuyi Mao , Jun Zhang , S.H. Song , Khaled B. Letaief

Comments: 33 pages, 7 figures, submitted to IEEE Transactions on Wireless Communications

Subjects

:

Information Theory (cs.IT)

Mobile-edge computing (MEC) has recently emerged as a prominent technology to

liberate mobile devices from computationally intensive workloads, by offloading

them to the proximate MEC server. To make offloading effective, the radio and

computational resources need to be dynamically managed, to cope with the

time-varying computation demands and wireless fading channels. In this paper,

we develop an online joint radio and computational resource management

algorithm for multi-user MEC systems, with the objective as minimizing the

long-term average weighted sum power consumption of the mobile devices and the

MEC server, subject to a task buffer stability constraint. Specifically, at

each time slot, the optimal CPU-cycle frequencies of the mobile devices are

obtained in closed forms, and the optimal transmit power and bandwidth

allocation for computation offloading are determined with the Gauss-Seidel

method; while for the MEC server, both the optimal frequencies of the CPU cores

and the optimal MEC server scheduling decision are derived in closed forms.

Besides, a delay-improved mechanism is proposed to reduce the execution delay.

Rigorous performance analysis is conducted for the proposed algorithm and its

delay-improved version, indicating that the weighted sum power consumption and

execution delay obey an (left[Oleft(1slash V

ight),Oleft(V

ight)

ight])

tradeoff with (V) as a control parameter. Simulation results are provided to

validate the theoretical analysis and demonstrate the impacts of various

parameters.

Guided Signal Reconstruction Theory

Andrew Knyazev , Akshay Gadde , Hassan Mansour , Dong Tian

Comments: 20 pages, 11 figures

Subjects

:

Information Theory (cs.IT)

; Functional Analysis (math.FA); Machine Learning (stat.ML)

An axiomatic approach to signal reconstruction is formulated, involving a

sample consistent set and a guiding set, describing desired reconstructions.

New frame-less reconstruction methods are proposed, based on a novel concept of

a reconstruction set, defined as a shortest pathway between the sample

consistent set and the guiding set. Existence and uniqueness of the

reconstruction set are investigated in a Hilbert space, where the guiding set

is a closed subspace and the sample consistent set is a closed plane, formed by

a sampling subspace. Connections to earlier known consistent, generalized, and

regularized reconstructions are clarified. New stability and reconstruction

error bounds are derived, using the largest nontrivial angle between the

sampling and guiding subspaces. Conjugate gradient iterative reconstruction

algorithms are proposed and illustrated numerically for image magnification.

An Introduction to Machine Learning Communications Systems

Timothy J. O'Shea , Jakob Hoydis

Comments: 10 pages, 8 figures, 5 tables, under concurrent academic journal submission

Subjects

:

Information Theory (cs.IT)

; Learning (cs.LG); Networking and Internet Architecture (cs.NI)

We introduce and motivate machine learning (ML) communications systems that

aim to improve on and to even replace the vast expert knowledge in the field of

communications using modern machine learning techniques. These have recently

achieved breakthroughs in many different domains, but not yet in

communications. By interpreting a communications system as an autoencoder, we

develop a fundamental new way to think about radio communications system design

as an end-to-end reconstruction optimization task that seeks to jointly

optimize transmitter and receiver components in a single process. We further

present the concept of Radio Transformer Networks (RTNs) as a means to

incorporate expert domain knowledge in the ML model and study the application

of convolutional neural networks (CNNs) on raw IQ time-series data for

modulation classification. We conclude the paper with a deep discussion of open

challenges and areas for future investigation.

Autocorrelation and Lower Bound on the 2-Adic Complexity of LSB Sequence of (p)-ary (m)-Sequence

Yuhua Sun , Qiang Wang , Tongjiang Yan

Comments: 28 pages

Subjects

:

Information Theory (cs.IT)

In modern stream cipher, there are many algorithms, such as ZUC, LTE

encryption algorithm and LTE integrity algorithm, using bit-component sequences

of (p)-ary (m)-sequences as the input of the algorithm. Therefore, analyzing

their statistical property (For example, autocorrelation, linear complexity and

2-adic complexity) of bit-component sequences of (p)-ary (m)-sequences is

becoming an important research topic. In this paper, we first derive some

autocorrelation properties of LSB (Least Significant Bit) sequences of (p)-ary

(m)-sequences, i.e., we convert the problem of computing autocorrelations of

LSB sequences of period (p^n-1) for any positive (ngeq2) to the problem of

determining autocorrelations of LSB sequence of period (p-1). Then, based on

this property and computer calculation, we list some autocorrelation

distributions of LSB sequences of (p)-ary (m)-sequences with order (n) for some

small primes (p)’s, such as (p=3,5,7,11,17,31). Additionally, using their

autocorrelation distributions and the method inspired by Hu, we give the lower

bounds on the 2-adic complexities of these LSB sequences. Our results show that

the main parts of all the lower bounds on the 2-adic complexity of these LSB

sequencesare larger than (frac{N}{2}), where (N) is the period of these

sequences. Therefor, these bounds are large enough to resist the analysis of

RAA (Rational Approximation Algorithm) for FCSR (Feedback with Carry Shift

Register). Especially, for a Mersenne prime (p=2^k-1), since all its

bit-component sequences of a (p)-ary (m)-sequence are shift equivalent, our

results hold for all its bit-component sequences.

Quantum Optimal Multiple Assignment Scheme for Realizing General Access Structure of Secret Sharing

Ryutaroh Matsumoto

Comments: ieice.cls, 3 pages, no figure

Journal-ref: IEICE Trans. Fundamentals, vol. E100-A, no. 2, pp. 726-728 (Feb.

2017)

Subjects

:

Quantum Physics (quant-ph)

; Cryptography and Security (cs.CR); Information Theory (cs.IT)

The multiple assignment scheme is to assign one or more shares to single

participant so that any kind of access structure can be realized by classical

secret sharing schemes. We propose its quantum version including ramp secret

sharing schemes. Then we propose an integer optimization approach to minimize

the average share size.

Scheduling and Power Allocation in Self-Backhauled Full Duplex Small Cells

Sanjay Goyal , Pei Liu , Shivendra Panwar

Comments: 7 pages, 7 figures, will appear in proceedings of IEEE ICC 2017

Subjects

:

Networking and Internet Architecture (cs.NI)

; Information Theory (cs.IT)

Full duplex (FD) communications, which increases spectral efficiency through

simultaneous transmission and reception on the same frequency band, is a

promising technology to meet the demand of next generation wireless networks.

In this paper, we consider the application of such FD communication to

self-backhauled small cells. We consider a FD capable small cell base station

(BS) being wirelessly backhauled by a FD capable macro-cell BS. FD

communication enables simultaneous backhaul and access transmissions at small

cell BSs, which reduces the need to orthogonalize allocated spectrum between

access and backhaul. However, in such simultaneous operations, all the links

experience higher interference, which significantly suppresses the gains of FD

operations. We propose an interference-aware scheduling method to maximize the

FD gain across multiple UEs in both uplink and downlink directions, while

maintaining a level of fairness between all UEs. It jointly schedules the

appropriate links and traffic based on the back-pressure algorithm, and

allocates appropriate transmission powers to the scheduled links using

Geometric Programming. Our simulation results show that the proposed scheduler

nearly doubles the throughput of small cells compared to traditional

half-duplex self-backhauling.

欢迎加入我爱机器学习QQ9群:173718917

arXiv Paper Daily: Mon, 6 Feb 2017

微信扫一扫,关注我爱机器学习公众号

微博:我爱机器学习

原文  https://www.52ml.net/21515.html
正文到此结束
Loading...