id
stringlengths 9
16
| submitter
stringlengths 3
64
⌀ | authors
stringlengths 5
6.63k
| title
stringlengths 7
245
| comments
stringlengths 1
482
⌀ | journal-ref
stringlengths 4
382
⌀ | doi
stringlengths 9
151
⌀ | report-no
stringclasses 984
values | categories
stringlengths 5
108
| license
stringclasses 9
values | abstract
stringlengths 83
3.41k
| versions
listlengths 1
20
| update_date
timestamp[s]date 2007-05-23 00:00:00
2025-04-11 00:00:00
| authors_parsed
sequencelengths 1
427
| prompt
stringlengths 166
3.49k
| label
stringclasses 2
values | prob
float64 0.5
0.98
|
---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
1603.01684 | Chenxing Xia | Hanling Zhang and Chenxing Xia | Saliency Detection combining Multi-layer Integration algorithm with
background prior and energy function | 25 pages, 8 figures. arXiv admin note: text overlap with
arXiv:1505.07192 by other authors | null | null | null | cs.CV | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | In this paper, we propose an improved mechanism for saliency detection.
Firstly,based on a neoteric background prior selecting four corners of an image
as background,we use color and spatial contrast with each superpixel to obtain
a salinecy map(CBP). Inspired by reverse-measurement methods to improve the
accuracy of measurement in Engineering,we employ the Objectness labels as
foreground prior based on part of information of CBP to construct a
map(OFP).Further,an original energy function is applied to optimize both of
them respectively and a single-layer saliency map(SLP)is formed by merging the
above twos.Finally,to deal with the scale problem,we obtain our multi-layer
map(MLP) by presenting an integration algorithm to take advantage of multiple
saliency maps. Quantitative and qualitative experiments on three datasets
demonstrate that our method performs favorably against the state-of-the-art
algorithm.
| [
{
"version": "v1",
"created": "Sat, 5 Mar 2016 06:12:44 GMT"
}
] | 2016-03-16T00:00:00 | [
[
"Zhang",
"Hanling",
""
],
[
"Xia",
"Chenxing",
""
]
] | TITLE: Saliency Detection combining Multi-layer Integration algorithm with
background prior and energy function
ABSTRACT: In this paper, we propose an improved mechanism for saliency detection.
Firstly,based on a neoteric background prior selecting four corners of an image
as background,we use color and spatial contrast with each superpixel to obtain
a salinecy map(CBP). Inspired by reverse-measurement methods to improve the
accuracy of measurement in Engineering,we employ the Objectness labels as
foreground prior based on part of information of CBP to construct a
map(OFP).Further,an original energy function is applied to optimize both of
them respectively and a single-layer saliency map(SLP)is formed by merging the
above twos.Finally,to deal with the scale problem,we obtain our multi-layer
map(MLP) by presenting an integration algorithm to take advantage of multiple
saliency maps. Quantitative and qualitative experiments on three datasets
demonstrate that our method performs favorably against the state-of-the-art
algorithm.
| no_new_dataset | 0.947527 |
1603.04522 | Yong Liu Stephen | Yong Liu, Peilin Zhao, Xin Liu, Min Wu, Xiao-Li Li | Learning Optimal Social Dependency for Recommendation | 8 pages, 2 figures | null | null | null | cs.IR cs.SI | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | Social recommender systems exploit users' social relationships to improve the
recommendation accuracy. Intuitively, a user tends to trust different subsets
of her social friends, regarding with different scenarios. Therefore, the main
challenge of social recommendation is to exploit the optimal social dependency
between users for a specific recommendation task. In this paper, we propose a
novel recommendation method, named probabilistic relational matrix
factorization (PRMF), which aims to learn the optimal social dependency between
users to improve the recommendation accuracy, with or without users' social
relationships. Specifically, in PRMF, the latent features of users are assumed
to follow a matrix variate normal (MVN) distribution. The positive and negative
dependency between users are modeled by the row precision matrix of the MVN
distribution. Moreover, we have also proposed an efficient alternating
algorithm to solve the optimization problem of PRMF. The experimental results
on real datasets demonstrate that the proposed PRMF method outperforms
state-of-the-art social recommendation approaches, in terms of root mean square
error (RMSE) and mean absolute error (MAE).
| [
{
"version": "v1",
"created": "Tue, 15 Mar 2016 01:41:11 GMT"
}
] | 2016-03-16T00:00:00 | [
[
"Liu",
"Yong",
""
],
[
"Zhao",
"Peilin",
""
],
[
"Liu",
"Xin",
""
],
[
"Wu",
"Min",
""
],
[
"Li",
"Xiao-Li",
""
]
] | TITLE: Learning Optimal Social Dependency for Recommendation
ABSTRACT: Social recommender systems exploit users' social relationships to improve the
recommendation accuracy. Intuitively, a user tends to trust different subsets
of her social friends, regarding with different scenarios. Therefore, the main
challenge of social recommendation is to exploit the optimal social dependency
between users for a specific recommendation task. In this paper, we propose a
novel recommendation method, named probabilistic relational matrix
factorization (PRMF), which aims to learn the optimal social dependency between
users to improve the recommendation accuracy, with or without users' social
relationships. Specifically, in PRMF, the latent features of users are assumed
to follow a matrix variate normal (MVN) distribution. The positive and negative
dependency between users are modeled by the row precision matrix of the MVN
distribution. Moreover, we have also proposed an efficient alternating
algorithm to solve the optimization problem of PRMF. The experimental results
on real datasets demonstrate that the proposed PRMF method outperforms
state-of-the-art social recommendation approaches, in terms of root mean square
error (RMSE) and mean absolute error (MAE).
| no_new_dataset | 0.942981 |
1603.04614 | Qingqun Ning | Qingqun Ning, Jianke Zhu, Zhiyuan Zhong, Steven C.H. Hoi, Chun Chen | Scalable Image Retrieval by Sparse Product Quantization | 12 pages | null | null | null | cs.CV | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | Fast Approximate Nearest Neighbor (ANN) search technique for high-dimensional
feature indexing and retrieval is the crux of large-scale image retrieval. A
recent promising technique is Product Quantization, which attempts to index
high-dimensional image features by decomposing the feature space into a
Cartesian product of low dimensional subspaces and quantizing each of them
separately. Despite the promising results reported, their quantization approach
follows the typical hard assignment of traditional quantization methods, which
may result in large quantization errors and thus inferior search performance.
Unlike the existing approaches, in this paper, we propose a novel approach
called Sparse Product Quantization (SPQ) to encoding the high-dimensional
feature vectors into sparse representation. We optimize the sparse
representations of the feature vectors by minimizing their quantization errors,
making the resulting representation is essentially close to the original data
in practice. Experiments show that the proposed SPQ technique is not only able
to compress data, but also an effective encoding technique. We obtain
state-of-the-art results for ANN search on four public image datasets and the
promising results of content-based image retrieval further validate the
efficacy of our proposed method.
| [
{
"version": "v1",
"created": "Tue, 15 Mar 2016 09:53:32 GMT"
}
] | 2016-03-16T00:00:00 | [
[
"Ning",
"Qingqun",
""
],
[
"Zhu",
"Jianke",
""
],
[
"Zhong",
"Zhiyuan",
""
],
[
"Hoi",
"Steven C. H.",
""
],
[
"Chen",
"Chun",
""
]
] | TITLE: Scalable Image Retrieval by Sparse Product Quantization
ABSTRACT: Fast Approximate Nearest Neighbor (ANN) search technique for high-dimensional
feature indexing and retrieval is the crux of large-scale image retrieval. A
recent promising technique is Product Quantization, which attempts to index
high-dimensional image features by decomposing the feature space into a
Cartesian product of low dimensional subspaces and quantizing each of them
separately. Despite the promising results reported, their quantization approach
follows the typical hard assignment of traditional quantization methods, which
may result in large quantization errors and thus inferior search performance.
Unlike the existing approaches, in this paper, we propose a novel approach
called Sparse Product Quantization (SPQ) to encoding the high-dimensional
feature vectors into sparse representation. We optimize the sparse
representations of the feature vectors by minimizing their quantization errors,
making the resulting representation is essentially close to the original data
in practice. Experiments show that the proposed SPQ technique is not only able
to compress data, but also an effective encoding technique. We obtain
state-of-the-art results for ANN search on four public image datasets and the
promising results of content-based image retrieval further validate the
efficacy of our proposed method.
| no_new_dataset | 0.947186 |
1507.03325 | Zhao Zhang | Zhao Zhang, Kyle Barbary, Frank Austin Nothaft, Evan Sparks, Oliver
Zahn, Michael J. Franklin, David A. Patterson, Saul Perlmutter | Scientific Computing Meets Big Data Technology: An Astronomy Use Case | null | null | null | null | cs.DC | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | Scientific analyses commonly compose multiple single-process programs into a
dataflow. An end-to-end dataflow of single-process programs is known as a
many-task application. Typically, tools from the HPC software stack are used to
parallelize these analyses. In this work, we investigate an alternate approach
that uses Apache Spark -- a modern big data platform -- to parallelize
many-task applications. We present Kira, a flexible and distributed astronomy
image processing toolkit using Apache Spark. We then use the Kira toolkit to
implement a Source Extractor application for astronomy images, called Kira SE.
With Kira SE as the use case, we study the programming flexibility, dataflow
richness, scheduling capacity and performance of Apache Spark running on the
EC2 cloud. By exploiting data locality, Kira SE achieves a 2.5x speedup over an
equivalent C program when analyzing a 1TB dataset using 512 cores on the Amazon
EC2 cloud. Furthermore, we show that by leveraging software originally designed
for big data infrastructure, Kira SE achieves competitive performance to the C
implementation running on the NERSC Edison supercomputer. Our experience with
Kira indicates that emerging Big Data platforms such as Apache Spark are a
performant alternative for many-task scientific applications.
| [
{
"version": "v1",
"created": "Mon, 13 Jul 2015 04:47:04 GMT"
},
{
"version": "v2",
"created": "Mon, 14 Mar 2016 19:04:09 GMT"
}
] | 2016-03-15T00:00:00 | [
[
"Zhang",
"Zhao",
""
],
[
"Barbary",
"Kyle",
""
],
[
"Nothaft",
"Frank Austin",
""
],
[
"Sparks",
"Evan",
""
],
[
"Zahn",
"Oliver",
""
],
[
"Franklin",
"Michael J.",
""
],
[
"Patterson",
"David A.",
""
],
[
"Perlmutter",
"Saul",
""
]
] | TITLE: Scientific Computing Meets Big Data Technology: An Astronomy Use Case
ABSTRACT: Scientific analyses commonly compose multiple single-process programs into a
dataflow. An end-to-end dataflow of single-process programs is known as a
many-task application. Typically, tools from the HPC software stack are used to
parallelize these analyses. In this work, we investigate an alternate approach
that uses Apache Spark -- a modern big data platform -- to parallelize
many-task applications. We present Kira, a flexible and distributed astronomy
image processing toolkit using Apache Spark. We then use the Kira toolkit to
implement a Source Extractor application for astronomy images, called Kira SE.
With Kira SE as the use case, we study the programming flexibility, dataflow
richness, scheduling capacity and performance of Apache Spark running on the
EC2 cloud. By exploiting data locality, Kira SE achieves a 2.5x speedup over an
equivalent C program when analyzing a 1TB dataset using 512 cores on the Amazon
EC2 cloud. Furthermore, we show that by leveraging software originally designed
for big data infrastructure, Kira SE achieves competitive performance to the C
implementation running on the NERSC Edison supercomputer. Our experience with
Kira indicates that emerging Big Data platforms such as Apache Spark are a
performant alternative for many-task scientific applications.
| no_new_dataset | 0.939137 |
1602.03418 | Swami Sankaranarayanan | Swami Sankaranarayanan, Azadeh Alavi, Rama Chellappa | Triplet Similarity Embedding for Face Verification | null | null | null | null | cs.CV | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | In this work, we present an unconstrained face verification algorithm and
evaluate it on the recently released IJB-A dataset that aims to push the
boundaries of face verification methods. The proposed algorithm couples a deep
CNN-based approach with a low-dimensional discriminative embedding learnt using
triplet similarity constraints in a large margin fashion. Aside from yielding
performance improvement, this embedding provides significant advantages in
terms of memory and post-processing operations like hashing and visualization.
Experiments on the IJB-A dataset show that the proposed algorithm outperforms
state of the art methods in verification and identification metrics, while
requiring less training time.
| [
{
"version": "v1",
"created": "Wed, 10 Feb 2016 15:48:47 GMT"
},
{
"version": "v2",
"created": "Sun, 13 Mar 2016 18:06:34 GMT"
}
] | 2016-03-15T00:00:00 | [
[
"Sankaranarayanan",
"Swami",
""
],
[
"Alavi",
"Azadeh",
""
],
[
"Chellappa",
"Rama",
""
]
] | TITLE: Triplet Similarity Embedding for Face Verification
ABSTRACT: In this work, we present an unconstrained face verification algorithm and
evaluate it on the recently released IJB-A dataset that aims to push the
boundaries of face verification methods. The proposed algorithm couples a deep
CNN-based approach with a low-dimensional discriminative embedding learnt using
triplet similarity constraints in a large margin fashion. Aside from yielding
performance improvement, this embedding provides significant advantages in
terms of memory and post-processing operations like hashing and visualization.
Experiments on the IJB-A dataset show that the proposed algorithm outperforms
state of the art methods in verification and identification metrics, while
requiring less training time.
| no_new_dataset | 0.878991 |
1602.04984 | Hyo-Eun Kim | Hyo-Eun Kim and Sangheum Hwang | Deconvolutional Feature Stacking for Weakly-Supervised Semantic
Segmentation | null | null | null | null | cs.CV | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | A weakly-supervised semantic segmentation framework with a tied
deconvolutional neural network is presented. Each deconvolution layer in the
framework consists of unpooling and deconvolution operations. 'Unpooling'
upsamples the input feature map based on unpooling switches defined by
corresponding convolution layer's pooling operation. 'Deconvolution' convolves
the input unpooled features by using convolutional weights tied with the
corresponding convolution layer's convolution operation. The
unpooling-deconvolution combination helps to eliminate less discriminative
features in a feature extraction stage, since output features of the
deconvolution layer are reconstructed from the most discriminative unpooled
features instead of the raw one. This results in reduction of false positives
in a pixel-level inference stage. All the feature maps restored from the entire
deconvolution layers can constitute a rich discriminative feature set according
to different abstraction levels. Those features are stacked to be selectively
used for generating class-specific activation maps. Under the weak supervision
(image-level labels), the proposed framework shows promising results on lesion
segmentation in medical images (chest X-rays) and achieves state-of-the-art
performance on the PASCAL VOC segmentation dataset in the same experimental
condition.
| [
{
"version": "v1",
"created": "Tue, 16 Feb 2016 11:05:24 GMT"
},
{
"version": "v2",
"created": "Thu, 18 Feb 2016 07:46:24 GMT"
},
{
"version": "v3",
"created": "Sat, 12 Mar 2016 08:22:30 GMT"
}
] | 2016-03-15T00:00:00 | [
[
"Kim",
"Hyo-Eun",
""
],
[
"Hwang",
"Sangheum",
""
]
] | TITLE: Deconvolutional Feature Stacking for Weakly-Supervised Semantic
Segmentation
ABSTRACT: A weakly-supervised semantic segmentation framework with a tied
deconvolutional neural network is presented. Each deconvolution layer in the
framework consists of unpooling and deconvolution operations. 'Unpooling'
upsamples the input feature map based on unpooling switches defined by
corresponding convolution layer's pooling operation. 'Deconvolution' convolves
the input unpooled features by using convolutional weights tied with the
corresponding convolution layer's convolution operation. The
unpooling-deconvolution combination helps to eliminate less discriminative
features in a feature extraction stage, since output features of the
deconvolution layer are reconstructed from the most discriminative unpooled
features instead of the raw one. This results in reduction of false positives
in a pixel-level inference stage. All the feature maps restored from the entire
deconvolution layers can constitute a rich discriminative feature set according
to different abstraction levels. Those features are stacked to be selectively
used for generating class-specific activation maps. Under the weak supervision
(image-level labels), the proposed framework shows promising results on lesion
segmentation in medical images (chest X-rays) and achieves state-of-the-art
performance on the PASCAL VOC segmentation dataset in the same experimental
condition.
| no_new_dataset | 0.950134 |
1603.01979 | Haewoon Kwak | Haewoon Kwak and Jisun An | Two Tales of the World: Comparison of Widely Used World News Datasets
GDELT and EventRegistry | To be appeared in ICWSM'16 | null | null | null | cs.DL cs.CY | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | In this work, we compare GDELT and Event Registry, which monitor news
articles worldwide and provide big data to researchers regarding scale, news
sources, and news geography. We found significant differences in scale and news
sources, but surprisingly, we observed high similarity in news geography
between the two datasets.
| [
{
"version": "v1",
"created": "Mon, 7 Mar 2016 09:25:57 GMT"
}
] | 2016-03-15T00:00:00 | [
[
"Kwak",
"Haewoon",
""
],
[
"An",
"Jisun",
""
]
] | TITLE: Two Tales of the World: Comparison of Widely Used World News Datasets
GDELT and EventRegistry
ABSTRACT: In this work, we compare GDELT and Event Registry, which monitor news
articles worldwide and provide big data to researchers regarding scale, news
sources, and news geography. We found significant differences in scale and news
sources, but surprisingly, we observed high similarity in news geography
between the two datasets.
| no_new_dataset | 0.943764 |
1603.03783 | Sachin Mehta | Sachin Mehta and Balakrishnan Prabhakaran | Region Graph Based Method for Multi-Object Detection and Tracking using
Depth Cameras | Accepted in IEEE Winter Conference in Computer Vision (WACV'16) | null | null | null | cs.CV | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | In this paper, we propose a multi-object detection and tracking method using
depth cameras. Depth maps are very noisy and obscure in object detection. We
first propose a region-based method to suppress high magnitude noise which
cannot be filtered using spatial filters. Second, the proposed method detect
Region of Interests by temporal learning which are then tracked using weighted
graph-based approach. We demonstrate the performance of the proposed method on
standard depth camera datasets with and without object occlusions. Experimental
results show that the proposed method is able to suppress high magnitude noise
in depth maps and detect/track the objects (with and without occlusion).
| [
{
"version": "v1",
"created": "Fri, 11 Mar 2016 21:06:35 GMT"
}
] | 2016-03-15T00:00:00 | [
[
"Mehta",
"Sachin",
""
],
[
"Prabhakaran",
"Balakrishnan",
""
]
] | TITLE: Region Graph Based Method for Multi-Object Detection and Tracking using
Depth Cameras
ABSTRACT: In this paper, we propose a multi-object detection and tracking method using
depth cameras. Depth maps are very noisy and obscure in object detection. We
first propose a region-based method to suppress high magnitude noise which
cannot be filtered using spatial filters. Second, the proposed method detect
Region of Interests by temporal learning which are then tracked using weighted
graph-based approach. We demonstrate the performance of the proposed method on
standard depth camera datasets with and without object occlusions. Experimental
results show that the proposed method is able to suppress high magnitude noise
in depth maps and detect/track the objects (with and without occlusion).
| no_new_dataset | 0.949856 |
1603.03827 | Franck Dernoncourt | Ji Young Lee, Franck Dernoncourt | Sequential Short-Text Classification with Recurrent and Convolutional
Neural Networks | Accepted as a conference paper at NAACL 2016 | null | null | null | cs.CL cs.AI cs.LG cs.NE stat.ML | http://creativecommons.org/licenses/by/4.0/ | Recent approaches based on artificial neural networks (ANNs) have shown
promising results for short-text classification. However, many short texts
occur in sequences (e.g., sentences in a document or utterances in a dialog),
and most existing ANN-based systems do not leverage the preceding short texts
when classifying a subsequent one. In this work, we present a model based on
recurrent neural networks and convolutional neural networks that incorporates
the preceding short texts. Our model achieves state-of-the-art results on three
different datasets for dialog act prediction.
| [
{
"version": "v1",
"created": "Sat, 12 Mar 2016 00:02:51 GMT"
}
] | 2016-03-15T00:00:00 | [
[
"Lee",
"Ji Young",
""
],
[
"Dernoncourt",
"Franck",
""
]
] | TITLE: Sequential Short-Text Classification with Recurrent and Convolutional
Neural Networks
ABSTRACT: Recent approaches based on artificial neural networks (ANNs) have shown
promising results for short-text classification. However, many short texts
occur in sequences (e.g., sentences in a document or utterances in a dialog),
and most existing ANN-based systems do not leverage the preceding short texts
when classifying a subsequent one. In this work, we present a model based on
recurrent neural networks and convolutional neural networks that incorporates
the preceding short texts. Our model achieves state-of-the-art results on three
different datasets for dialog act prediction.
| no_new_dataset | 0.953923 |
1603.03836 | Amirali Aghazadeh | Amirali Aghazadeh, Andrew Lan, Anshumali Shrivastava, Richard Baraniuk | Near-Isometric Binary Hashing for Large-scale Datasets | null | null | null | null | cs.DS | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | We develop a scalable algorithm to learn binary hash codes for indexing
large-scale datasets. Near-isometric binary hashing (NIBH) is a data-dependent
hashing scheme that quantizes the output of a learned low-dimensional embedding
to obtain a binary hash code. In contrast to conventional hashing schemes,
which typically rely on an $\ell_2$-norm (i.e., average distortion)
minimization, NIBH is based on a $\ell_{\infty}$-norm (i.e., worst-case
distortion) minimization that provides several benefits, including superior
distance, ranking, and near-neighbor preservation performance. We develop a
practical and efficient algorithm for NIBH based on column generation that
scales well to large datasets. A range of experimental evaluations demonstrate
the superiority of NIBH over ten state-of-the-art binary hashing schemes.
| [
{
"version": "v1",
"created": "Sat, 12 Mar 2016 01:04:50 GMT"
}
] | 2016-03-15T00:00:00 | [
[
"Aghazadeh",
"Amirali",
""
],
[
"Lan",
"Andrew",
""
],
[
"Shrivastava",
"Anshumali",
""
],
[
"Baraniuk",
"Richard",
""
]
] | TITLE: Near-Isometric Binary Hashing for Large-scale Datasets
ABSTRACT: We develop a scalable algorithm to learn binary hash codes for indexing
large-scale datasets. Near-isometric binary hashing (NIBH) is a data-dependent
hashing scheme that quantizes the output of a learned low-dimensional embedding
to obtain a binary hash code. In contrast to conventional hashing schemes,
which typically rely on an $\ell_2$-norm (i.e., average distortion)
minimization, NIBH is based on a $\ell_{\infty}$-norm (i.e., worst-case
distortion) minimization that provides several benefits, including superior
distance, ranking, and near-neighbor preservation performance. We develop a
practical and efficient algorithm for NIBH based on column generation that
scales well to large datasets. A range of experimental evaluations demonstrate
the superiority of NIBH over ten state-of-the-art binary hashing schemes.
| no_new_dataset | 0.949435 |
1603.03984 | Rudrasis Chakraborty Mr | Rudrasis Chakraborty, Dohyung Seo and Baba C. Vemuri | An efficient Exact-PGA algorithm for constant curvature manifolds | Accepted in CVPR (IEEE Conference on Computer Vision and Pattern
Recognition) 2016 | null | null | null | cs.CV | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | Manifold-valued datasets are widely encountered in many computer vision
tasks. A non-linear analog of the PCA, called the Principal Geodesic Analysis
(PGA) suited for data lying on Riemannian manifolds was reported in literature
a decade ago. Since the objective function in PGA is highly non-linear and hard
to solve efficiently in general, researchers have proposed a linear
approximation. Though this linear approximation is easy to compute, it lacks
accuracy especially when the data exhibits a large variance. Recently, an
alternative called exact PGA was proposed which tries to solve the optimization
without any linearization. For general Riemannian manifolds, though it gives
better accuracy than the original (linearized) PGA, for data that exhibit large
variance, the optimization is not computationally efficient. In this paper, we
propose an efficient exact PGA for constant curvature Riemannian manifolds
(CCM-EPGA). CCM-EPGA differs significantly from existing PGA algorithms in two
aspects, (i) the distance between a given manifold-valued data point and the
principal submanifold is computed analytically and thus no optimization is
required as in existing methods. (ii) Unlike the existing PGA algorithms, the
descent into codimension-1 submanifolds does not require any optimization but
is accomplished through the use of the Rimeannian inverse Exponential map and
the parallel transport operations. We present theoretical and experimental
results for constant curvature Riemannian manifolds depicting favorable
performance of CCM-EPGA compared to existing PGA algorithms. We also present
data reconstruction from principal components and directions which has not been
presented in literature in this setting.
| [
{
"version": "v1",
"created": "Sun, 13 Mar 2016 02:57:34 GMT"
}
] | 2016-03-15T00:00:00 | [
[
"Chakraborty",
"Rudrasis",
""
],
[
"Seo",
"Dohyung",
""
],
[
"Vemuri",
"Baba C.",
""
]
] | TITLE: An efficient Exact-PGA algorithm for constant curvature manifolds
ABSTRACT: Manifold-valued datasets are widely encountered in many computer vision
tasks. A non-linear analog of the PCA, called the Principal Geodesic Analysis
(PGA) suited for data lying on Riemannian manifolds was reported in literature
a decade ago. Since the objective function in PGA is highly non-linear and hard
to solve efficiently in general, researchers have proposed a linear
approximation. Though this linear approximation is easy to compute, it lacks
accuracy especially when the data exhibits a large variance. Recently, an
alternative called exact PGA was proposed which tries to solve the optimization
without any linearization. For general Riemannian manifolds, though it gives
better accuracy than the original (linearized) PGA, for data that exhibit large
variance, the optimization is not computationally efficient. In this paper, we
propose an efficient exact PGA for constant curvature Riemannian manifolds
(CCM-EPGA). CCM-EPGA differs significantly from existing PGA algorithms in two
aspects, (i) the distance between a given manifold-valued data point and the
principal submanifold is computed analytically and thus no optimization is
required as in existing methods. (ii) Unlike the existing PGA algorithms, the
descent into codimension-1 submanifolds does not require any optimization but
is accomplished through the use of the Rimeannian inverse Exponential map and
the parallel transport operations. We present theoretical and experimental
results for constant curvature Riemannian manifolds depicting favorable
performance of CCM-EPGA compared to existing PGA algorithms. We also present
data reconstruction from principal components and directions which has not been
presented in literature in this setting.
| no_new_dataset | 0.952131 |
1603.04002 | Richard Nock | Giorgio Patrini, Richard Nock, Stephen Hardy, Tiberio Caetano | Fast Learning from Distributed Datasets without Entity Matching | null | null | null | null | cs.LG cs.DC | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | Consider the following data fusion scenario: two datasets/peers contain the
same real-world entities described using partially shared features, e.g.
banking and insurance company records of the same customer base. Our goal is to
learn a classifier in the cross product space of the two domains, in the hard
case in which no shared ID is available -- e.g. due to anonymization.
Traditionally, the problem is approached by first addressing entity matching
and subsequently learning the classifier in a standard manner. We present an
end-to-end solution which bypasses matching entities, based on the recently
introduced concept of Rademacher observations (rados). Informally, we replace
the minimisation of a loss over examples, which requires to solve entity
resolution, by the equivalent minimisation of a (different) loss over rados.
Among others, key properties we show are (i) a potentially huge subset of these
rados does not require to perform entity matching, and (ii) the algorithm that
provably minimizes the rado loss over these rados has time and space
complexities smaller than the algorithm minimizing the equivalent example loss.
Last, we relax a key assumption of the model, that the data is vertically
partitioned among peers --- in this case, we would not even know the existence
of a solution to entity resolution. In this more general setting, experiments
validate the possibility of significantly beating even the optimal peer in
hindsight.
| [
{
"version": "v1",
"created": "Sun, 13 Mar 2016 06:03:39 GMT"
}
] | 2016-03-15T00:00:00 | [
[
"Patrini",
"Giorgio",
""
],
[
"Nock",
"Richard",
""
],
[
"Hardy",
"Stephen",
""
],
[
"Caetano",
"Tiberio",
""
]
] | TITLE: Fast Learning from Distributed Datasets without Entity Matching
ABSTRACT: Consider the following data fusion scenario: two datasets/peers contain the
same real-world entities described using partially shared features, e.g.
banking and insurance company records of the same customer base. Our goal is to
learn a classifier in the cross product space of the two domains, in the hard
case in which no shared ID is available -- e.g. due to anonymization.
Traditionally, the problem is approached by first addressing entity matching
and subsequently learning the classifier in a standard manner. We present an
end-to-end solution which bypasses matching entities, based on the recently
introduced concept of Rademacher observations (rados). Informally, we replace
the minimisation of a loss over examples, which requires to solve entity
resolution, by the equivalent minimisation of a (different) loss over rados.
Among others, key properties we show are (i) a potentially huge subset of these
rados does not require to perform entity matching, and (ii) the algorithm that
provably minimizes the rado loss over these rados has time and space
complexities smaller than the algorithm minimizing the equivalent example loss.
Last, we relax a key assumption of the model, that the data is vertically
partitioned among peers --- in this case, we would not even know the existence
of a solution to entity resolution. In this more general setting, experiments
validate the possibility of significantly beating even the optimal peer in
hindsight.
| no_new_dataset | 0.948489 |
1603.04010 | Sofiane Abbar | Sofiane Abbar, Tahar Zanouda, Laure Berti-Equille, Javier
Borge-Holthoefer | Using Twitter to Understand Public Interest in Climate Change: The case
of Qatar | Will appear in the proceedings of the International Workshop on
Social Media for Environment and Ecological Monitoring (SWEEM'16) | null | null | null | cs.SI cs.CY | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | Climate change has received an extensive attention from public opinion in the
last couple of years, after being considered for decades as an exclusive
scientific debate. Governments and world-wide organizations such as the United
Nations are working more than ever on raising and maintaining public awareness
toward this global issue. In the present study, we examine and analyze Climate
Change conversations in Qatar's Twittersphere, and sense public awareness
towards this global and shared problem in general, and its various related
topics in particular. Such topics include but are not limited to politics,
economy, disasters, energy and sandstorms. To address this concern, we collect
and analyze a large dataset of 109 million tweets posted by 98K distinct users
living in Qatar -- one of the largest emitters of CO2 worldwide. We use a
taxonomy of climate change topics created as part of the United Nations Pulse
project to capture the climate change discourse in more than 36K tweets. We
also examine which topics people refer to when they discuss climate change, and
perform different analysis to understand the temporal dynamics of public
interest toward these topics.
| [
{
"version": "v1",
"created": "Sun, 13 Mar 2016 08:39:19 GMT"
}
] | 2016-03-15T00:00:00 | [
[
"Abbar",
"Sofiane",
""
],
[
"Zanouda",
"Tahar",
""
],
[
"Berti-Equille",
"Laure",
""
],
[
"Borge-Holthoefer",
"Javier",
""
]
] | TITLE: Using Twitter to Understand Public Interest in Climate Change: The case
of Qatar
ABSTRACT: Climate change has received an extensive attention from public opinion in the
last couple of years, after being considered for decades as an exclusive
scientific debate. Governments and world-wide organizations such as the United
Nations are working more than ever on raising and maintaining public awareness
toward this global issue. In the present study, we examine and analyze Climate
Change conversations in Qatar's Twittersphere, and sense public awareness
towards this global and shared problem in general, and its various related
topics in particular. Such topics include but are not limited to politics,
economy, disasters, energy and sandstorms. To address this concern, we collect
and analyze a large dataset of 109 million tweets posted by 98K distinct users
living in Qatar -- one of the largest emitters of CO2 worldwide. We use a
taxonomy of climate change topics created as part of the United Nations Pulse
project to capture the climate change discourse in more than 36K tweets. We
also examine which topics people refer to when they discuss climate change, and
perform different analysis to understand the temporal dynamics of public
interest toward these topics.
| new_dataset | 0.845113 |
1603.04026 | Huamin Ren | Huamin Ren, Hong Pan, S{\o}ren Ingvor Olsen, Thomas B. Moeslund | A comprehensive study of sparse codes on abnormality detection | 7 pages | null | null | null | cs.CV | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | Sparse representation has been applied successfully in abnormal event
detection, in which the baseline is to learn a dictionary accompanied by sparse
codes. While much emphasis is put on discriminative dictionary construction,
there are no comparative studies of sparse codes regarding abnormality
detection. We comprehensively study two types of sparse codes solutions -
greedy algorithms and convex L1-norm solutions - and their impact on
abnormality detection performance. We also propose our framework of combining
sparse codes with different detection methods. Our comparative experiments are
carried out from various angles to better understand the applicability of
sparse codes, including computation time, reconstruction error, sparsity,
detection accuracy, and their performance combining various detection methods.
Experiments show that combining OMP codes with maximum coordinate detection
could achieve state-of-the-art performance on the UCSD dataset [14].
| [
{
"version": "v1",
"created": "Sun, 13 Mar 2016 13:13:10 GMT"
}
] | 2016-03-15T00:00:00 | [
[
"Ren",
"Huamin",
""
],
[
"Pan",
"Hong",
""
],
[
"Olsen",
"Søren Ingvor",
""
],
[
"Moeslund",
"Thomas B.",
""
]
] | TITLE: A comprehensive study of sparse codes on abnormality detection
ABSTRACT: Sparse representation has been applied successfully in abnormal event
detection, in which the baseline is to learn a dictionary accompanied by sparse
codes. While much emphasis is put on discriminative dictionary construction,
there are no comparative studies of sparse codes regarding abnormality
detection. We comprehensively study two types of sparse codes solutions -
greedy algorithms and convex L1-norm solutions - and their impact on
abnormality detection performance. We also propose our framework of combining
sparse codes with different detection methods. Our comparative experiments are
carried out from various angles to better understand the applicability of
sparse codes, including computation time, reconstruction error, sparsity,
detection accuracy, and their performance combining various detection methods.
Experiments show that combining OMP codes with maximum coordinate detection
could achieve state-of-the-art performance on the UCSD dataset [14].
| no_new_dataset | 0.940898 |
1603.04042 | Ning Xu | Ning Xu, Brian Price, Scott Cohen, Jimei Yang, Thomas Huang | Deep Interactive Object Selection | Computer Vision and Pattern Recognition | null | null | null | cs.CV | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | Interactive object selection is a very important research problem and has
many applications. Previous algorithms require substantial user interactions to
estimate the foreground and background distributions. In this paper, we present
a novel deep learning based algorithm which has a much better understanding of
objectness and thus can reduce user interactions to just a few clicks. Our
algorithm transforms user provided positive and negative clicks into two
Euclidean distance maps which are then concatenated with the RGB channels of
images to compose (image, user interactions) pairs. We generate many of such
pairs by combining several random sampling strategies to model user click
patterns and use them to fine tune deep Fully Convolutional Networks (FCNs).
Finally the output probability maps of our FCN 8s model is integrated with
graph cut optimization to refine the boundary segments. Our model is trained on
the PASCAL segmentation dataset and evaluated on other datasets with different
object classes. Experimental results on both seen and unseen objects clearly
demonstrate that our algorithm has a good generalization ability and is
superior to all existing interactive object selection approaches.
| [
{
"version": "v1",
"created": "Sun, 13 Mar 2016 15:42:34 GMT"
}
] | 2016-03-15T00:00:00 | [
[
"Xu",
"Ning",
""
],
[
"Price",
"Brian",
""
],
[
"Cohen",
"Scott",
""
],
[
"Yang",
"Jimei",
""
],
[
"Huang",
"Thomas",
""
]
] | TITLE: Deep Interactive Object Selection
ABSTRACT: Interactive object selection is a very important research problem and has
many applications. Previous algorithms require substantial user interactions to
estimate the foreground and background distributions. In this paper, we present
a novel deep learning based algorithm which has a much better understanding of
objectness and thus can reduce user interactions to just a few clicks. Our
algorithm transforms user provided positive and negative clicks into two
Euclidean distance maps which are then concatenated with the RGB channels of
images to compose (image, user interactions) pairs. We generate many of such
pairs by combining several random sampling strategies to model user click
patterns and use them to fine tune deep Fully Convolutional Networks (FCNs).
Finally the output probability maps of our FCN 8s model is integrated with
graph cut optimization to refine the boundary segments. Our model is trained on
the PASCAL segmentation dataset and evaluated on other datasets with different
object classes. Experimental results on both seen and unseen objects clearly
demonstrate that our algorithm has a good generalization ability and is
superior to all existing interactive object selection approaches.
| no_new_dataset | 0.950041 |
1603.04265 | Tae Hyun Kim | Tae Hyun Kim, Seungjun Nah, and Kyoung Mu Lee | Dynamic Scene Deblurring using a Locally Adaptive Linear Blur Model | null | null | null | null | cs.CV | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | State-of-the-art video deblurring methods cannot handle blurry videos
recorded in dynamic scenes, since they are built under a strong assumption that
the captured scenes are static. Contrary to the existing methods, we propose a
video deblurring algorithm that can deal with general blurs inherent in dynamic
scenes. To handle general and locally varying blurs caused by various sources,
such as moving objects, camera shake, depth variation, and defocus, we estimate
pixel-wise non-uniform blur kernels. We infer bidirectional optical flows to
handle motion blurs, and also estimate Gaussian blur maps to remove optical
blur from defocus in our new blur model. Therefore, we propose a single energy
model that jointly estimates optical flows, defocus blur maps and latent
frames. We also provide a framework and efficient solvers to minimize the
proposed energy model. By optimizing the energy model, we achieve significant
improvements in removing general blurs, estimating optical flows, and extending
depth-of-field in blurry frames. Moreover, in this work, to evaluate the
performance of non-uniform deblurring methods objectively, we have constructed
a new realistic dataset with ground truths. In addition, extensive experimental
on publicly available challenging video data demonstrate that the proposed
method produces qualitatively superior performance than the state-of-the-art
methods which often fail in either deblurring or optical flow estimation.
| [
{
"version": "v1",
"created": "Mon, 14 Mar 2016 13:57:19 GMT"
}
] | 2016-03-15T00:00:00 | [
[
"Kim",
"Tae Hyun",
""
],
[
"Nah",
"Seungjun",
""
],
[
"Lee",
"Kyoung Mu",
""
]
] | TITLE: Dynamic Scene Deblurring using a Locally Adaptive Linear Blur Model
ABSTRACT: State-of-the-art video deblurring methods cannot handle blurry videos
recorded in dynamic scenes, since they are built under a strong assumption that
the captured scenes are static. Contrary to the existing methods, we propose a
video deblurring algorithm that can deal with general blurs inherent in dynamic
scenes. To handle general and locally varying blurs caused by various sources,
such as moving objects, camera shake, depth variation, and defocus, we estimate
pixel-wise non-uniform blur kernels. We infer bidirectional optical flows to
handle motion blurs, and also estimate Gaussian blur maps to remove optical
blur from defocus in our new blur model. Therefore, we propose a single energy
model that jointly estimates optical flows, defocus blur maps and latent
frames. We also provide a framework and efficient solvers to minimize the
proposed energy model. By optimizing the energy model, we achieve significant
improvements in removing general blurs, estimating optical flows, and extending
depth-of-field in blurry frames. Moreover, in this work, to evaluate the
performance of non-uniform deblurring methods objectively, we have constructed
a new realistic dataset with ground truths. In addition, extensive experimental
on publicly available challenging video data demonstrate that the proposed
method produces qualitatively superior performance than the state-of-the-art
methods which often fail in either deblurring or optical flow estimation.
| new_dataset | 0.955899 |
1603.04319 | Jalal Etesami | Jalal Etesami, Negar Kiyavash, Kun Zhang, Kushagra Singhal | Learning Network of Multivariate Hawkes Processes: A Time Series
Approach | null | null | null | null | cs.LG cs.AI stat.ML | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | Learning the influence structure of multiple time series data is of great
interest to many disciplines. This paper studies the problem of recovering the
causal structure in network of multivariate linear Hawkes processes. In such
processes, the occurrence of an event in one process affects the probability of
occurrence of new events in some other processes. Thus, a natural notion of
causality exists between such processes captured by the support of the
excitation matrix. We show that the resulting causal influence network is
equivalent to the Directed Information graph (DIG) of the processes, which
encodes the causal factorization of the joint distribution of the processes.
Furthermore, we present an algorithm for learning the support of excitation
matrix (or equivalently the DIG). The performance of the algorithm is evaluated
on synthesized multivariate Hawkes networks as well as a stock market and
MemeTracker real-world dataset.
| [
{
"version": "v1",
"created": "Mon, 14 Mar 2016 16:08:26 GMT"
}
] | 2016-03-15T00:00:00 | [
[
"Etesami",
"Jalal",
""
],
[
"Kiyavash",
"Negar",
""
],
[
"Zhang",
"Kun",
""
],
[
"Singhal",
"Kushagra",
""
]
] | TITLE: Learning Network of Multivariate Hawkes Processes: A Time Series
Approach
ABSTRACT: Learning the influence structure of multiple time series data is of great
interest to many disciplines. This paper studies the problem of recovering the
causal structure in network of multivariate linear Hawkes processes. In such
processes, the occurrence of an event in one process affects the probability of
occurrence of new events in some other processes. Thus, a natural notion of
causality exists between such processes captured by the support of the
excitation matrix. We show that the resulting causal influence network is
equivalent to the Directed Information graph (DIG) of the processes, which
encodes the causal factorization of the joint distribution of the processes.
Furthermore, we present an algorithm for learning the support of excitation
matrix (or equivalently the DIG). The performance of the algorithm is evaluated
on synthesized multivariate Hawkes networks as well as a stock market and
MemeTracker real-world dataset.
| no_new_dataset | 0.949809 |
1603.04327 | Ibrahim Sadek | Ibrahim Sadek | Automatic Discrimination of Color Retinal Images using the Bag of Words
Approach | null | null | null | null | cs.CV | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | Diabetic retinopathy (DR) and age related macular degeneration (ARMD) are
among the major causes of visual impairment worldwide. DR is mainly
characterized by red spots, namely microaneurysms and bright lesions,
specifically exudates whereas ARMD is mainly identified by tiny yellow or white
deposits called drusen. Since exudates might be the only manifestation of the
early diabetic retinopathy, there is an increase demand for automatic
retinopathy diagnosis. Exudates and drusen may share similar appearances, thus
discriminating between them is of interest to enhance screening performance. In
this research, we investigative the role of bag of words approach in the
automatic diagnosis of retinopathy diabetes. We proposed to use a single based
and multiple based methods for the construction of the visual dictionary by
combining the histogram of word occurrences from each dictionary and building a
single histogram. The introduced approach is evaluated for automatic diagnosis
of normal and abnormal color fundus images with bright lesions. This approach
has been implemented on 430 fundus images, including six publicly available
datasets, in addition to one local dataset. The mean accuracies reported are
97.2% and 99.77% for single based and multiple based dictionaries respectively.
| [
{
"version": "v1",
"created": "Mon, 14 Mar 2016 16:26:32 GMT"
}
] | 2016-03-15T00:00:00 | [
[
"Sadek",
"Ibrahim",
""
]
] | TITLE: Automatic Discrimination of Color Retinal Images using the Bag of Words
Approach
ABSTRACT: Diabetic retinopathy (DR) and age related macular degeneration (ARMD) are
among the major causes of visual impairment worldwide. DR is mainly
characterized by red spots, namely microaneurysms and bright lesions,
specifically exudates whereas ARMD is mainly identified by tiny yellow or white
deposits called drusen. Since exudates might be the only manifestation of the
early diabetic retinopathy, there is an increase demand for automatic
retinopathy diagnosis. Exudates and drusen may share similar appearances, thus
discriminating between them is of interest to enhance screening performance. In
this research, we investigative the role of bag of words approach in the
automatic diagnosis of retinopathy diabetes. We proposed to use a single based
and multiple based methods for the construction of the visual dictionary by
combining the histogram of word occurrences from each dictionary and building a
single histogram. The introduced approach is evaluated for automatic diagnosis
of normal and abnormal color fundus images with bright lesions. This approach
has been implemented on 430 fundus images, including six publicly available
datasets, in addition to one local dataset. The mean accuracies reported are
97.2% and 99.77% for single based and multiple based dictionaries respectively.
| no_new_dataset | 0.883286 |
1512.02895 | Shaoting Zhang | Xiaofan Zhang and Feng Zhou and Yuanqing Lin and Shaoting Zhang | Embedding Label Structures for Fine-Grained Feature Representation | null | null | null | null | cs.CV | http://creativecommons.org/licenses/by-nc-sa/4.0/ | Recent algorithms in convolutional neural networks (CNN) considerably advance
the fine-grained image classification, which aims to differentiate subtle
differences among subordinate classes. However, previous studies have rarely
focused on learning a fined-grained and structured feature representation that
is able to locate similar images at different levels of relevance, e.g.,
discovering cars from the same make or the same model, both of which require
high precision. In this paper, we propose two main contributions to tackle this
problem. 1) A multi-task learning framework is designed to effectively learn
fine-grained feature representations by jointly optimizing both classification
and similarity constraints. 2) To model the multi-level relevance, label
structures such as hierarchy or shared attributes are seamlessly embedded into
the framework by generalizing the triplet loss. Extensive and thorough
experiments have been conducted on three fine-grained datasets, i.e., the
Stanford car, the car-333, and the food datasets, which contain either
hierarchical labels or shared attributes. Our proposed method has achieved very
competitive performance, i.e., among state-of-the-art classification accuracy.
More importantly, it significantly outperforms previous fine-grained feature
representations for image retrieval at different levels of relevance.
| [
{
"version": "v1",
"created": "Wed, 9 Dec 2015 15:22:26 GMT"
},
{
"version": "v2",
"created": "Fri, 11 Mar 2016 02:59:36 GMT"
}
] | 2016-03-14T00:00:00 | [
[
"Zhang",
"Xiaofan",
""
],
[
"Zhou",
"Feng",
""
],
[
"Lin",
"Yuanqing",
""
],
[
"Zhang",
"Shaoting",
""
]
] | TITLE: Embedding Label Structures for Fine-Grained Feature Representation
ABSTRACT: Recent algorithms in convolutional neural networks (CNN) considerably advance
the fine-grained image classification, which aims to differentiate subtle
differences among subordinate classes. However, previous studies have rarely
focused on learning a fined-grained and structured feature representation that
is able to locate similar images at different levels of relevance, e.g.,
discovering cars from the same make or the same model, both of which require
high precision. In this paper, we propose two main contributions to tackle this
problem. 1) A multi-task learning framework is designed to effectively learn
fine-grained feature representations by jointly optimizing both classification
and similarity constraints. 2) To model the multi-level relevance, label
structures such as hierarchy or shared attributes are seamlessly embedded into
the framework by generalizing the triplet loss. Extensive and thorough
experiments have been conducted on three fine-grained datasets, i.e., the
Stanford car, the car-333, and the food datasets, which contain either
hierarchical labels or shared attributes. Our proposed method has achieved very
competitive performance, i.e., among state-of-the-art classification accuracy.
More importantly, it significantly outperforms previous fine-grained feature
representations for image retrieval at different levels of relevance.
| no_new_dataset | 0.947137 |
1603.03541 | Chenxia Wu | Chenxia Wu, Jiemi Zhang, Ozan Sener, Bart Selman, Silvio Savarese,
Ashutosh Saxena | Watch-n-Patch: Unsupervised Learning of Actions and Relations | arXiv admin note: text overlap with arXiv:1512.04208 | null | null | null | cs.CV cs.LG cs.RO | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | There is a large variation in the activities that humans perform in their
everyday lives. We consider modeling these composite human activities which
comprises multiple basic level actions in a completely unsupervised setting.
Our model learns high-level co-occurrence and temporal relations between the
actions. We consider the video as a sequence of short-term action clips, which
contains human-words and object-words. An activity is about a set of
action-topics and object-topics indicating which actions are present and which
objects are interacting with. We then propose a new probabilistic model
relating the words and the topics. It allows us to model long-range action
relations that commonly exist in the composite activities, which is challenging
in previous works. We apply our model to the unsupervised action segmentation
and clustering, and to a novel application that detects forgotten actions,
which we call action patching. For evaluation, we contribute a new challenging
RGB-D activity video dataset recorded by the new Kinect v2, which contains
several human daily activities as compositions of multiple actions interacting
with different objects. Moreover, we develop a robotic system that watches
people and reminds people by applying our action patching algorithm. Our
robotic setup can be easily deployed on any assistive robot.
| [
{
"version": "v1",
"created": "Fri, 11 Mar 2016 07:13:59 GMT"
}
] | 2016-03-14T00:00:00 | [
[
"Wu",
"Chenxia",
""
],
[
"Zhang",
"Jiemi",
""
],
[
"Sener",
"Ozan",
""
],
[
"Selman",
"Bart",
""
],
[
"Savarese",
"Silvio",
""
],
[
"Saxena",
"Ashutosh",
""
]
] | TITLE: Watch-n-Patch: Unsupervised Learning of Actions and Relations
ABSTRACT: There is a large variation in the activities that humans perform in their
everyday lives. We consider modeling these composite human activities which
comprises multiple basic level actions in a completely unsupervised setting.
Our model learns high-level co-occurrence and temporal relations between the
actions. We consider the video as a sequence of short-term action clips, which
contains human-words and object-words. An activity is about a set of
action-topics and object-topics indicating which actions are present and which
objects are interacting with. We then propose a new probabilistic model
relating the words and the topics. It allows us to model long-range action
relations that commonly exist in the composite activities, which is challenging
in previous works. We apply our model to the unsupervised action segmentation
and clustering, and to a novel application that detects forgotten actions,
which we call action patching. For evaluation, we contribute a new challenging
RGB-D activity video dataset recorded by the new Kinect v2, which contains
several human daily activities as compositions of multiple actions interacting
with different objects. Moreover, we develop a robotic system that watches
people and reminds people by applying our action patching algorithm. Our
robotic setup can be easily deployed on any assistive robot.
| new_dataset | 0.955026 |
1603.03627 | Roberto Luis Shinmoto Torres | Roberto L. Shinmoto Torres and Damith C. Ranasinghe and Qinfeng Shi
and Anton van den Hengel | Learning from Imbalanced Multiclass Sequential Data Streams Using
Dynamically Weighted Conditional Random Fields | 28 pages, 8 figures, 1 table | null | null | null | cs.LG | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | The present study introduces a method for improving the classification
performance of imbalanced multiclass data streams from wireless body worn
sensors. Data imbalance is an inherent problem in activity recognition caused
by the irregular time distribution of activities, which are sequential and
dependent on previous movements. We use conditional random fields (CRF), a
graphical model for structured classification, to take advantage of
dependencies between activities in a sequence. However, CRFs do not consider
the negative effects of class imbalance during training. We propose a
class-wise dynamically weighted CRF (dWCRF) where weights are automatically
determined during training by maximizing the expected overall F-score. Our
results based on three case studies from a healthcare application using a
batteryless body worn sensor, demonstrate that our method, in general, improves
overall and minority class F-score when compared to other CRF based classifiers
and achieves similar or better overall and class-wise performance when compared
to SVM based classifiers under conditions of limited training data. We also
confirm the performance of our approach using an additional battery powered
body worn sensor dataset, achieving similar results in cases of high class
imbalance.
| [
{
"version": "v1",
"created": "Fri, 11 Mar 2016 13:51:37 GMT"
}
] | 2016-03-14T00:00:00 | [
[
"Torres",
"Roberto L. Shinmoto",
""
],
[
"Ranasinghe",
"Damith C.",
""
],
[
"Shi",
"Qinfeng",
""
],
[
"Hengel",
"Anton van den",
""
]
] | TITLE: Learning from Imbalanced Multiclass Sequential Data Streams Using
Dynamically Weighted Conditional Random Fields
ABSTRACT: The present study introduces a method for improving the classification
performance of imbalanced multiclass data streams from wireless body worn
sensors. Data imbalance is an inherent problem in activity recognition caused
by the irregular time distribution of activities, which are sequential and
dependent on previous movements. We use conditional random fields (CRF), a
graphical model for structured classification, to take advantage of
dependencies between activities in a sequence. However, CRFs do not consider
the negative effects of class imbalance during training. We propose a
class-wise dynamically weighted CRF (dWCRF) where weights are automatically
determined during training by maximizing the expected overall F-score. Our
results based on three case studies from a healthcare application using a
batteryless body worn sensor, demonstrate that our method, in general, improves
overall and minority class F-score when compared to other CRF based classifiers
and achieves similar or better overall and class-wise performance when compared
to SVM based classifiers under conditions of limited training data. We also
confirm the performance of our approach using an additional battery powered
body worn sensor dataset, achieving similar results in cases of high class
imbalance.
| no_new_dataset | 0.949995 |
1603.03724 | Niharika Gauraha Niharika Gauraha | Niharika Gauraha, Swapan K. Parui | Efficient Clustering of Correlated Variables and Variable Selection in
High-Dimensional Linear Models | null | null | null | null | stat.ML cs.LG | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | In this paper, we introduce Adaptive Cluster Lasso(ACL) method for variable
selection in high dimensional sparse regression models with strongly correlated
variables. To handle correlated variables, the concept of clustering or
grouping variables and then pursuing model fitting is widely accepted. When the
dimension is very high, finding an appropriate group structure is as difficult
as the original problem. The ACL is a three-stage procedure where, at the first
stage, we use the Lasso(or its adaptive or thresholded version) to do initial
selection, then we also include those variables which are not selected by the
Lasso but are strongly correlated with the variables selected by the Lasso. At
the second stage we cluster the variables based on the reduced set of
predictors and in the third stage we perform sparse estimation such as Lasso on
cluster representatives or the group Lasso based on the structures generated by
clustering procedure. We show that our procedure is consistent and efficient in
finding true underlying population group structure(under assumption of
irrepresentable and beta-min conditions). We also study the group selection
consistency of our method and we support the theory using simulated and
pseudo-real dataset examples.
| [
{
"version": "v1",
"created": "Fri, 11 Mar 2016 19:06:33 GMT"
}
] | 2016-03-14T00:00:00 | [
[
"Gauraha",
"Niharika",
""
],
[
"Parui",
"Swapan K.",
""
]
] | TITLE: Efficient Clustering of Correlated Variables and Variable Selection in
High-Dimensional Linear Models
ABSTRACT: In this paper, we introduce Adaptive Cluster Lasso(ACL) method for variable
selection in high dimensional sparse regression models with strongly correlated
variables. To handle correlated variables, the concept of clustering or
grouping variables and then pursuing model fitting is widely accepted. When the
dimension is very high, finding an appropriate group structure is as difficult
as the original problem. The ACL is a three-stage procedure where, at the first
stage, we use the Lasso(or its adaptive or thresholded version) to do initial
selection, then we also include those variables which are not selected by the
Lasso but are strongly correlated with the variables selected by the Lasso. At
the second stage we cluster the variables based on the reduced set of
predictors and in the third stage we perform sparse estimation such as Lasso on
cluster representatives or the group Lasso based on the structures generated by
clustering procedure. We show that our procedure is consistent and efficient in
finding true underlying population group structure(under assumption of
irrepresentable and beta-min conditions). We also study the group selection
consistency of our method and we support the theory using simulated and
pseudo-real dataset examples.
| no_new_dataset | 0.951594 |
1406.5370 | Alexandre d'Aspremont | Fajwel Fogel, Alexandre d'Aspremont, Milan Vojnovic | Spectral Ranking using Seriation | Substantially revised. Accepted by JMLR | null | null | null | cs.LG cs.AI stat.ML | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | We describe a seriation algorithm for ranking a set of items given pairwise
comparisons between these items. Intuitively, the algorithm assigns similar
rankings to items that compare similarly with all others. It does so by
constructing a similarity matrix from pairwise comparisons, using seriation
methods to reorder this matrix and construct a ranking. We first show that this
spectral seriation algorithm recovers the true ranking when all pairwise
comparisons are observed and consistent with a total order. We then show that
ranking reconstruction is still exact when some pairwise comparisons are
corrupted or missing, and that seriation based spectral ranking is more robust
to noise than classical scoring methods. Finally, we bound the ranking error
when only a random subset of the comparions are observed. An additional benefit
of the seriation formulation is that it allows us to solve semi-supervised
ranking problems. Experiments on both synthetic and real datasets demonstrate
that seriation based spectral ranking achieves competitive and in some cases
superior performance compared to classical ranking methods.
| [
{
"version": "v1",
"created": "Fri, 20 Jun 2014 12:58:46 GMT"
},
{
"version": "v2",
"created": "Thu, 25 Jun 2015 10:21:50 GMT"
},
{
"version": "v3",
"created": "Mon, 18 Jan 2016 18:09:07 GMT"
},
{
"version": "v4",
"created": "Thu, 10 Mar 2016 18:15:19 GMT"
}
] | 2016-03-11T00:00:00 | [
[
"Fogel",
"Fajwel",
""
],
[
"d'Aspremont",
"Alexandre",
""
],
[
"Vojnovic",
"Milan",
""
]
] | TITLE: Spectral Ranking using Seriation
ABSTRACT: We describe a seriation algorithm for ranking a set of items given pairwise
comparisons between these items. Intuitively, the algorithm assigns similar
rankings to items that compare similarly with all others. It does so by
constructing a similarity matrix from pairwise comparisons, using seriation
methods to reorder this matrix and construct a ranking. We first show that this
spectral seriation algorithm recovers the true ranking when all pairwise
comparisons are observed and consistent with a total order. We then show that
ranking reconstruction is still exact when some pairwise comparisons are
corrupted or missing, and that seriation based spectral ranking is more robust
to noise than classical scoring methods. Finally, we bound the ranking error
when only a random subset of the comparions are observed. An additional benefit
of the seriation formulation is that it allows us to solve semi-supervised
ranking problems. Experiments on both synthetic and real datasets demonstrate
that seriation based spectral ranking achieves competitive and in some cases
superior performance compared to classical ranking methods.
| no_new_dataset | 0.946498 |
1410.6751 | Christoph Riedl | Christoph Riedl, Richard Zanibbi, Marti A. Hearst, Siyu Zhu, Michael
Menietti, Jason Crusan, Ivan Metelsky, Karim R. Lakhani | Detecting Figures and Part Labels in Patents: Competition-Based
Development of Image Processing Algorithms | null | null | 10.1007/s10032-016-0260-8 | null | cs.CV cs.IR | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | We report the findings of a month-long online competition in which
participants developed algorithms for augmenting the digital version of patent
documents published by the United States Patent and Trademark Office (USPTO).
The goal was to detect figures and part labels in U.S. patent drawing pages.
The challenge drew 232 teams of two, of which 70 teams (30%) submitted
solutions. Collectively, teams submitted 1,797 solutions that were compiled on
the competition servers. Participants reported spending an average of 63 hours
developing their solutions, resulting in a total of 5,591 hours of development
time. A manually labeled dataset of 306 patents was used for training, online
system tests, and evaluation. The design and performance of the top-5 systems
are presented, along with a system developed after the competition which
illustrates that winning teams produced near state-of-the-art results under
strict time and computation constraints. For the 1st place system, the harmonic
mean of recall and precision (f-measure) was 88.57% for figure region
detection, 78.81% for figure regions with correctly recognized figure titles,
and 70.98% for part label detection and character recognition. Data and
software from the competition are available through the online UCI Machine
Learning repository to inspire follow-on work by the image processing
community.
| [
{
"version": "v1",
"created": "Fri, 24 Oct 2014 17:45:36 GMT"
},
{
"version": "v2",
"created": "Mon, 27 Oct 2014 10:54:17 GMT"
},
{
"version": "v3",
"created": "Tue, 11 Nov 2014 14:33:11 GMT"
}
] | 2016-03-11T00:00:00 | [
[
"Riedl",
"Christoph",
""
],
[
"Zanibbi",
"Richard",
""
],
[
"Hearst",
"Marti A.",
""
],
[
"Zhu",
"Siyu",
""
],
[
"Menietti",
"Michael",
""
],
[
"Crusan",
"Jason",
""
],
[
"Metelsky",
"Ivan",
""
],
[
"Lakhani",
"Karim R.",
""
]
] | TITLE: Detecting Figures and Part Labels in Patents: Competition-Based
Development of Image Processing Algorithms
ABSTRACT: We report the findings of a month-long online competition in which
participants developed algorithms for augmenting the digital version of patent
documents published by the United States Patent and Trademark Office (USPTO).
The goal was to detect figures and part labels in U.S. patent drawing pages.
The challenge drew 232 teams of two, of which 70 teams (30%) submitted
solutions. Collectively, teams submitted 1,797 solutions that were compiled on
the competition servers. Participants reported spending an average of 63 hours
developing their solutions, resulting in a total of 5,591 hours of development
time. A manually labeled dataset of 306 patents was used for training, online
system tests, and evaluation. The design and performance of the top-5 systems
are presented, along with a system developed after the competition which
illustrates that winning teams produced near state-of-the-art results under
strict time and computation constraints. For the 1st place system, the harmonic
mean of recall and precision (f-measure) was 88.57% for figure region
detection, 78.81% for figure regions with correctly recognized figure titles,
and 70.98% for part label detection and character recognition. Data and
software from the competition are available through the online UCI Machine
Learning repository to inspire follow-on work by the image processing
community.
| new_dataset | 0.938576 |
1510.07867 | Rasmus Rothe | Rasmus Rothe and Radu Timofte and Luc Van Gool | Some like it hot - visual guidance for preference prediction | accepted for publication at CVPR 2016 | null | null | null | cs.CV | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | For people first impressions of someone are of determining importance. They
are hard to alter through further information. This begs the question if a
computer can reach the same judgement. Earlier research has already pointed out
that age, gender, and average attractiveness can be estimated with reasonable
precision. We improve the state-of-the-art, but also predict - based on
someone's known preferences - how much that particular person is attracted to a
novel face. Our computational pipeline comprises a face detector, convolutional
neural networks for the extraction of deep features, standard support vector
regression for gender, age and facial beauty, and - as the main novelties -
visual regularized collaborative filtering to infer inter-person preferences as
well as a novel regression technique for handling visual queries without rating
history. We validate the method using a very large dataset from a dating site
as well as images from celebrities. Our experiments yield convincing results,
i.e. we predict 76% of the ratings correctly solely based on an image, and
reveal some sociologically relevant conclusions. We also validate our
collaborative filtering solution on the standard MovieLens rating dataset,
augmented with movie posters, to predict an individual's movie rating. We
demonstrate our algorithms on howhot.io which went viral around the Internet
with more than 50 million pictures evaluated in the first month.
| [
{
"version": "v1",
"created": "Tue, 27 Oct 2015 11:17:46 GMT"
},
{
"version": "v2",
"created": "Thu, 10 Mar 2016 15:02:19 GMT"
}
] | 2016-03-11T00:00:00 | [
[
"Rothe",
"Rasmus",
""
],
[
"Timofte",
"Radu",
""
],
[
"Van Gool",
"Luc",
""
]
] | TITLE: Some like it hot - visual guidance for preference prediction
ABSTRACT: For people first impressions of someone are of determining importance. They
are hard to alter through further information. This begs the question if a
computer can reach the same judgement. Earlier research has already pointed out
that age, gender, and average attractiveness can be estimated with reasonable
precision. We improve the state-of-the-art, but also predict - based on
someone's known preferences - how much that particular person is attracted to a
novel face. Our computational pipeline comprises a face detector, convolutional
neural networks for the extraction of deep features, standard support vector
regression for gender, age and facial beauty, and - as the main novelties -
visual regularized collaborative filtering to infer inter-person preferences as
well as a novel regression technique for handling visual queries without rating
history. We validate the method using a very large dataset from a dating site
as well as images from celebrities. Our experiments yield convincing results,
i.e. we predict 76% of the ratings correctly solely based on an image, and
reveal some sociologically relevant conclusions. We also validate our
collaborative filtering solution on the standard MovieLens rating dataset,
augmented with movie posters, to predict an individual's movie rating. We
demonstrate our algorithms on howhot.io which went viral around the Internet
with more than 50 million pictures evaluated in the first month.
| no_new_dataset | 0.934275 |
1511.04317 | Mansour Ahmadi | Mansour Ahmadi, Dmitry Ulyanov, Stanislav Semenov, Mikhail Trofimov,
Giorgio Giacinto | Novel Feature Extraction, Selection and Fusion for Effective Malware
Family Classification | null | null | null | null | cs.CR cs.AI | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | Modern malware is designed with mutation characteristics, namely polymorphism
and metamorphism, which causes an enormous growth in the number of variants of
malware samples. Categorization of malware samples on the basis of their
behaviors is essential for the computer security community, because they
receive huge number of malware everyday, and the signature extraction process
is usually based on malicious parts characterizing malware families. Microsoft
released a malware classification challenge in 2015 with a huge dataset of near
0.5 terabytes of data, containing more than 20K malware samples. The analysis
of this dataset inspired the development of a novel paradigm that is effective
in categorizing malware variants into their actual family groups. This paradigm
is presented and discussed in the present paper, where emphasis has been given
to the phases related to the extraction, and selection of a set of novel
features for the effective representation of malware samples. Features can be
grouped according to different characteristics of malware behavior, and their
fusion is performed according to a per-class weighting paradigm. The proposed
method achieved a very high accuracy ($\approx$ 0.998) on the Microsoft Malware
Challenge dataset.
| [
{
"version": "v1",
"created": "Fri, 13 Nov 2015 15:33:02 GMT"
},
{
"version": "v2",
"created": "Thu, 10 Mar 2016 10:21:15 GMT"
}
] | 2016-03-11T00:00:00 | [
[
"Ahmadi",
"Mansour",
""
],
[
"Ulyanov",
"Dmitry",
""
],
[
"Semenov",
"Stanislav",
""
],
[
"Trofimov",
"Mikhail",
""
],
[
"Giacinto",
"Giorgio",
""
]
] | TITLE: Novel Feature Extraction, Selection and Fusion for Effective Malware
Family Classification
ABSTRACT: Modern malware is designed with mutation characteristics, namely polymorphism
and metamorphism, which causes an enormous growth in the number of variants of
malware samples. Categorization of malware samples on the basis of their
behaviors is essential for the computer security community, because they
receive huge number of malware everyday, and the signature extraction process
is usually based on malicious parts characterizing malware families. Microsoft
released a malware classification challenge in 2015 with a huge dataset of near
0.5 terabytes of data, containing more than 20K malware samples. The analysis
of this dataset inspired the development of a novel paradigm that is effective
in categorizing malware variants into their actual family groups. This paradigm
is presented and discussed in the present paper, where emphasis has been given
to the phases related to the extraction, and selection of a set of novel
features for the effective representation of malware samples. Features can be
grouped according to different characteristics of malware behavior, and their
fusion is performed according to a per-class weighting paradigm. The proposed
method achieved a very high accuracy ($\approx$ 0.998) on the Microsoft Malware
Challenge dataset.
| new_dataset | 0.845815 |
1512.07372 | Pin-Yu Chen | Pin-Yu Chen, Sutanay Choudhury, Alfred O. Hero | Multi-centrality Graph Spectral Decompositions and their Application to
Cyber Intrusion Detection | To appear in ICASSP 2016 | null | null | null | cs.SI cs.CR stat.ML | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | Many modern datasets can be represented as graphs and hence spectral
decompositions such as graph principal component analysis (PCA) can be useful.
Distinct from previous graph decomposition approaches based on subspace
projection of a single topological feature, e.g., the Fiedler vector of
centered graph adjacency matrix (graph Laplacian), we propose spectral
decomposition approaches to graph PCA and graph dictionary learning that
integrate multiple features, including graph walk statistics, centrality
measures and graph distances to reference nodes. In this paper we propose a new
PCA method for single graph analysis, called multi-centrality graph PCA
(MC-GPCA), and a new dictionary learning method for ensembles of graphs, called
multi-centrality graph dictionary learning (MC-GDL), both based on spectral
decomposition of multi-centrality matrices. As an application to cyber
intrusion detection, MC-GPCA can be an effective indicator of anomalous
connectivity pattern and MC-GDL can provide discriminative basis for attack
classification.
| [
{
"version": "v1",
"created": "Wed, 23 Dec 2015 07:13:27 GMT"
},
{
"version": "v2",
"created": "Thu, 10 Mar 2016 03:31:25 GMT"
}
] | 2016-03-11T00:00:00 | [
[
"Chen",
"Pin-Yu",
""
],
[
"Choudhury",
"Sutanay",
""
],
[
"Hero",
"Alfred O.",
""
]
] | TITLE: Multi-centrality Graph Spectral Decompositions and their Application to
Cyber Intrusion Detection
ABSTRACT: Many modern datasets can be represented as graphs and hence spectral
decompositions such as graph principal component analysis (PCA) can be useful.
Distinct from previous graph decomposition approaches based on subspace
projection of a single topological feature, e.g., the Fiedler vector of
centered graph adjacency matrix (graph Laplacian), we propose spectral
decomposition approaches to graph PCA and graph dictionary learning that
integrate multiple features, including graph walk statistics, centrality
measures and graph distances to reference nodes. In this paper we propose a new
PCA method for single graph analysis, called multi-centrality graph PCA
(MC-GPCA), and a new dictionary learning method for ensembles of graphs, called
multi-centrality graph dictionary learning (MC-GDL), both based on spectral
decomposition of multi-centrality matrices. As an application to cyber
intrusion detection, MC-GPCA can be an effective indicator of anomalous
connectivity pattern and MC-GDL can provide discriminative basis for attack
classification.
| no_new_dataset | 0.94474 |
1603.03097 | Yu Wang | Yu Wang, Yuncheng Li, Jiebo Luo | Deciphering the 2016 U.S. Presidential Campaign in the Twitter Sphere: A
Comparison of the Trumpists and Clintonists | 4 pages, to appear in the 10th International AAAI Conference on Web
and Social Media | null | null | null | cs.SI | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | In this paper, we study follower demographics of Donald Trump and Hillary
Clinton, the two leading candidates in the 2016 U.S. presidential race. We
build a unique dataset US2016, which includes the number of followers for each
candidate from September 17, 2015 to December 22, 2015. US2016 also includes
the geographical location of these followers, the number of their own followers
and, very importantly, the profile image of each follower. We use individuals'
number of followers and profile images to analyze four dimensions of follower
demographics: social status, gender, race and age. Our study shows that in
terms of social influence, the Trumpists are more polarized than the
Clintonists: they tend to have either a lot of influence or little influence.
We also find that compared with the Clintonists, the Trumpists are more likely
to be either very young or very old. Our study finds no gender affinity effect
for Clinton in the Twitter sphere, but we do find that the Clintonists are more
racially diverse.
| [
{
"version": "v1",
"created": "Wed, 9 Mar 2016 23:36:26 GMT"
}
] | 2016-03-11T00:00:00 | [
[
"Wang",
"Yu",
""
],
[
"Li",
"Yuncheng",
""
],
[
"Luo",
"Jiebo",
""
]
] | TITLE: Deciphering the 2016 U.S. Presidential Campaign in the Twitter Sphere: A
Comparison of the Trumpists and Clintonists
ABSTRACT: In this paper, we study follower demographics of Donald Trump and Hillary
Clinton, the two leading candidates in the 2016 U.S. presidential race. We
build a unique dataset US2016, which includes the number of followers for each
candidate from September 17, 2015 to December 22, 2015. US2016 also includes
the geographical location of these followers, the number of their own followers
and, very importantly, the profile image of each follower. We use individuals'
number of followers and profile images to analyze four dimensions of follower
demographics: social status, gender, race and age. Our study shows that in
terms of social influence, the Trumpists are more polarized than the
Clintonists: they tend to have either a lot of influence or little influence.
We also find that compared with the Clintonists, the Trumpists are more likely
to be either very young or very old. Our study finds no gender affinity effect
for Clinton in the Twitter sphere, but we do find that the Clintonists are more
racially diverse.
| new_dataset | 0.956796 |
1603.03101 | Chen-Yu Lee | Chen-Yu Lee and Simon Osindero | Recursive Recurrent Nets with Attention Modeling for OCR in the Wild | accepted at CVPR 2016 | null | null | null | cs.CV | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | We present recursive recurrent neural networks with attention modeling
(R$^2$AM) for lexicon-free optical character recognition in natural scene
images. The primary advantages of the proposed method are: (1) use of recursive
convolutional neural networks (CNNs), which allow for parametrically efficient
and effective image feature extraction; (2) an implicitly learned
character-level language model, embodied in a recurrent neural network which
avoids the need to use N-grams; and (3) the use of a soft-attention mechanism,
allowing the model to selectively exploit image features in a coordinated way,
and allowing for end-to-end training within a standard backpropagation
framework. We validate our method with state-of-the-art performance on
challenging benchmark datasets: Street View Text, IIIT5k, ICDAR and Synth90k.
| [
{
"version": "v1",
"created": "Wed, 9 Mar 2016 23:49:51 GMT"
}
] | 2016-03-11T00:00:00 | [
[
"Lee",
"Chen-Yu",
""
],
[
"Osindero",
"Simon",
""
]
] | TITLE: Recursive Recurrent Nets with Attention Modeling for OCR in the Wild
ABSTRACT: We present recursive recurrent neural networks with attention modeling
(R$^2$AM) for lexicon-free optical character recognition in natural scene
images. The primary advantages of the proposed method are: (1) use of recursive
convolutional neural networks (CNNs), which allow for parametrically efficient
and effective image feature extraction; (2) an implicitly learned
character-level language model, embodied in a recurrent neural network which
avoids the need to use N-grams; and (3) the use of a soft-attention mechanism,
allowing the model to selectively exploit image features in a coordinated way,
and allowing for end-to-end training within a standard backpropagation
framework. We validate our method with state-of-the-art performance on
challenging benchmark datasets: Street View Text, IIIT5k, ICDAR and Synth90k.
| no_new_dataset | 0.953923 |
1603.03281 | UshaRani Yelipe | Yelipe UshaRani, P. Sammulal | An Innovative Imputation and Classification Approach for Accurate
Disease Prediction | Special Issue of Journal IJCSIS indexed in Web of Science and Thomson
Reuters ISI. https://sites.google.com/site/ijcsis/vol-14-s1-feb-2016 | null | null | null | cs.DB cs.IR cs.LG | http://creativecommons.org/licenses/by/4.0/ | Imputation of missing attribute values in medical datasets for extracting
hidden knowledge from medical datasets is an interesting research topic of
interest which is very challenging. One cannot eliminate missing values in
medical records. The reason may be because some tests may not been conducted as
they are cost effective, values missed when conducting clinical trials, values
may not have been recorded to name some of the reasons. Data mining researchers
have been proposing various approaches to find and impute missing values to
increase classification accuracies so that disease may be predicted accurately.
In this paper, we propose a novel imputation approach for imputation of missing
values and performing classification after fixing missing values. The approach
is based on clustering concept and aims at dimensionality reduction of the
records. The case study discussed shows that missing values can be fixed and
imputed efficiently by achieving dimensionality reduction. The importance of
proposed approach for classification is visible in the case study which assigns
single class label in contrary to multi-label assignment if dimensionality
reduction is not performed.
| [
{
"version": "v1",
"created": "Thu, 10 Mar 2016 14:31:33 GMT"
}
] | 2016-03-11T00:00:00 | [
[
"UshaRani",
"Yelipe",
""
],
[
"Sammulal",
"P.",
""
]
] | TITLE: An Innovative Imputation and Classification Approach for Accurate
Disease Prediction
ABSTRACT: Imputation of missing attribute values in medical datasets for extracting
hidden knowledge from medical datasets is an interesting research topic of
interest which is very challenging. One cannot eliminate missing values in
medical records. The reason may be because some tests may not been conducted as
they are cost effective, values missed when conducting clinical trials, values
may not have been recorded to name some of the reasons. Data mining researchers
have been proposing various approaches to find and impute missing values to
increase classification accuracies so that disease may be predicted accurately.
In this paper, we propose a novel imputation approach for imputation of missing
values and performing classification after fixing missing values. The approach
is based on clustering concept and aims at dimensionality reduction of the
records. The case study discussed shows that missing values can be fixed and
imputed efficiently by achieving dimensionality reduction. The importance of
proposed approach for classification is visible in the case study which assigns
single class label in contrary to multi-label assignment if dimensionality
reduction is not performed.
| no_new_dataset | 0.947575 |
1603.03303 | Rahmtin Rotabi | Rahmtin Rotabi and Jon Kleinberg | The Status Gradient of Trends in Social Media | null | null | null | null | cs.SI cs.CY physics.soc-ph | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | An active line of research has studied the detection and representation of
trends in social media content. There is still relatively little understanding,
however, of methods to characterize the early adopters of these trends: who
picks up on these trends at different points in time, and what is their role in
the system? We develop a framework for analyzing the population of users who
participate in trending topics over the course of these topics' lifecycles.
Central to our analysis is the notion of a "status gradient", describing how
users of different activity levels adopt a trend at different points in time.
Across multiple datasets, we find that this methodology reveals key differences
in the nature of the early adopters in different domains.
| [
{
"version": "v1",
"created": "Thu, 10 Mar 2016 15:46:07 GMT"
}
] | 2016-03-11T00:00:00 | [
[
"Rotabi",
"Rahmtin",
""
],
[
"Kleinberg",
"Jon",
""
]
] | TITLE: The Status Gradient of Trends in Social Media
ABSTRACT: An active line of research has studied the detection and representation of
trends in social media content. There is still relatively little understanding,
however, of methods to characterize the early adopters of these trends: who
picks up on these trends at different points in time, and what is their role in
the system? We develop a framework for analyzing the population of users who
participate in trending topics over the course of these topics' lifecycles.
Central to our analysis is the notion of a "status gradient", describing how
users of different activity levels adopt a trend at different points in time.
Across multiple datasets, we find that this methodology reveals key differences
in the nature of the early adopters in different domains.
| no_new_dataset | 0.943971 |
1504.07947 | Le Hou | Le Hou, Dimitris Samaras, Tahsin M. Kurc, Yi Gao, James E. Davis, Joel
H. Saltz | Patch-based Convolutional Neural Network for Whole Slide Tissue Image
Classification | null | null | null | null | cs.CV | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | Convolutional Neural Networks (CNN) are state-of-the-art models for many
image classification tasks. However, to recognize cancer subtypes
automatically, training a CNN on gigapixel resolution Whole Slide Tissue Images
(WSI) is currently computationally impossible. The differentiation of cancer
subtypes is based on cellular-level visual features observed on image patch
scale. Therefore, we argue that in this situation, training a patch-level
classifier on image patches will perform better than or similar to an
image-level classifier. The challenge becomes how to intelligently combine
patch-level classification results and model the fact that not all patches will
be discriminative. We propose to train a decision fusion model to aggregate
patch-level predictions given by patch-level CNNs, which to the best of our
knowledge has not been shown before. Furthermore, we formulate a novel
Expectation-Maximization (EM) based method that automatically locates
discriminative patches robustly by utilizing the spatial relationships of
patches. We apply our method to the classification of glioma and non-small-cell
lung carcinoma cases into subtypes. The classification accuracy of our method
is similar to the inter-observer agreement between pathologists. Although it is
impossible to train CNNs on WSIs, we experimentally demonstrate using a
comparable non-cancer dataset of smaller images that a patch-based CNN can
outperform an image-based CNN.
| [
{
"version": "v1",
"created": "Wed, 29 Apr 2015 18:15:22 GMT"
},
{
"version": "v2",
"created": "Mon, 11 May 2015 01:55:55 GMT"
},
{
"version": "v3",
"created": "Tue, 19 May 2015 21:01:11 GMT"
},
{
"version": "v4",
"created": "Tue, 8 Mar 2016 18:07:03 GMT"
},
{
"version": "v5",
"created": "Wed, 9 Mar 2016 14:26:16 GMT"
}
] | 2016-03-10T00:00:00 | [
[
"Hou",
"Le",
""
],
[
"Samaras",
"Dimitris",
""
],
[
"Kurc",
"Tahsin M.",
""
],
[
"Gao",
"Yi",
""
],
[
"Davis",
"James E.",
""
],
[
"Saltz",
"Joel H.",
""
]
] | TITLE: Patch-based Convolutional Neural Network for Whole Slide Tissue Image
Classification
ABSTRACT: Convolutional Neural Networks (CNN) are state-of-the-art models for many
image classification tasks. However, to recognize cancer subtypes
automatically, training a CNN on gigapixel resolution Whole Slide Tissue Images
(WSI) is currently computationally impossible. The differentiation of cancer
subtypes is based on cellular-level visual features observed on image patch
scale. Therefore, we argue that in this situation, training a patch-level
classifier on image patches will perform better than or similar to an
image-level classifier. The challenge becomes how to intelligently combine
patch-level classification results and model the fact that not all patches will
be discriminative. We propose to train a decision fusion model to aggregate
patch-level predictions given by patch-level CNNs, which to the best of our
knowledge has not been shown before. Furthermore, we formulate a novel
Expectation-Maximization (EM) based method that automatically locates
discriminative patches robustly by utilizing the spatial relationships of
patches. We apply our method to the classification of glioma and non-small-cell
lung carcinoma cases into subtypes. The classification accuracy of our method
is similar to the inter-observer agreement between pathologists. Although it is
impossible to train CNNs on WSIs, we experimentally demonstrate using a
comparable non-cancer dataset of smaller images that a patch-based CNN can
outperform an image-based CNN.
| no_new_dataset | 0.951818 |
1506.01273 | David Martins de Matos | Marta Apar\'icio, Paulo Figueiredo, Francisco Raposo, David Martins de
Matos, Ricardo Ribeiro, Lu\'is Marujo | Summarization of Films and Documentaries Based on Subtitles and Scripts | 7 pages, 9 tables, 4 figures, submitted to Pattern Recognition
Letters (Elsevier) | Pattern Recognition Letters, Volume 73, 1 April 2016, Pages 7-12 | 10.1016/j.patrec.2015.12.016 | null | cs.CL cs.AI cs.IR | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | We assess the performance of generic text summarization algorithms applied to
films and documentaries, using the well-known behavior of summarization of news
articles as reference. We use three datasets: (i) news articles, (ii) film
scripts and subtitles, and (iii) documentary subtitles. Standard ROUGE metrics
are used for comparing generated summaries against news abstracts, plot
summaries, and synopses. We show that the best performing algorithms are LSA,
for news articles and documentaries, and LexRank and Support Sets, for films.
Despite the different nature of films and documentaries, their relative
behavior is in accordance with that obtained for news articles.
| [
{
"version": "v1",
"created": "Wed, 3 Jun 2015 15:07:14 GMT"
},
{
"version": "v2",
"created": "Thu, 4 Jun 2015 12:41:55 GMT"
},
{
"version": "v3",
"created": "Wed, 9 Mar 2016 16:50:43 GMT"
}
] | 2016-03-10T00:00:00 | [
[
"Aparício",
"Marta",
""
],
[
"Figueiredo",
"Paulo",
""
],
[
"Raposo",
"Francisco",
""
],
[
"de Matos",
"David Martins",
""
],
[
"Ribeiro",
"Ricardo",
""
],
[
"Marujo",
"Luís",
""
]
] | TITLE: Summarization of Films and Documentaries Based on Subtitles and Scripts
ABSTRACT: We assess the performance of generic text summarization algorithms applied to
films and documentaries, using the well-known behavior of summarization of news
articles as reference. We use three datasets: (i) news articles, (ii) film
scripts and subtitles, and (iii) documentary subtitles. Standard ROUGE metrics
are used for comparing generated summaries against news abstracts, plot
summaries, and synopses. We show that the best performing algorithms are LSA,
for news articles and documentaries, and LexRank and Support Sets, for films.
Despite the different nature of films and documentaries, their relative
behavior is in accordance with that obtained for news articles.
| no_new_dataset | 0.947962 |
1602.00991 | Peter Ondruska | Peter Ondruska and Ingmar Posner | Deep Tracking: Seeing Beyond Seeing Using Recurrent Neural Networks | Published in The Thirtieth AAAI Conference on Artificial Intelligence
(AAAI-16), Video: https://youtu.be/cdeWCpfUGWc, Code:
http://mrg.robots.ox.ac.uk/mrg_people/peter-ondruska/ | null | null | null | cs.LG cs.AI cs.CV cs.NE cs.RO | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | This paper presents to the best of our knowledge the first end-to-end object
tracking approach which directly maps from raw sensor input to object tracks in
sensor space without requiring any feature engineering or system identification
in the form of plant or sensor models. Specifically, our system accepts a
stream of raw sensor data at one end and, in real-time, produces an estimate of
the entire environment state at the output including even occluded objects. We
achieve this by framing the problem as a deep learning task and exploit
sequence models in the form of recurrent neural networks to learn a mapping
from sensor measurements to object tracks. In particular, we propose a learning
method based on a form of input dropout which allows learning in an
unsupervised manner, only based on raw, occluded sensor data without access to
ground-truth annotations. We demonstrate our approach using a synthetic dataset
designed to mimic the task of tracking objects in 2D laser data -- as commonly
encountered in robotics applications -- and show that it learns to track many
dynamic objects despite occlusions and the presence of sensor noise.
| [
{
"version": "v1",
"created": "Tue, 2 Feb 2016 16:10:16 GMT"
},
{
"version": "v2",
"created": "Tue, 8 Mar 2016 22:09:05 GMT"
}
] | 2016-03-10T00:00:00 | [
[
"Ondruska",
"Peter",
""
],
[
"Posner",
"Ingmar",
""
]
] | TITLE: Deep Tracking: Seeing Beyond Seeing Using Recurrent Neural Networks
ABSTRACT: This paper presents to the best of our knowledge the first end-to-end object
tracking approach which directly maps from raw sensor input to object tracks in
sensor space without requiring any feature engineering or system identification
in the form of plant or sensor models. Specifically, our system accepts a
stream of raw sensor data at one end and, in real-time, produces an estimate of
the entire environment state at the output including even occluded objects. We
achieve this by framing the problem as a deep learning task and exploit
sequence models in the form of recurrent neural networks to learn a mapping
from sensor measurements to object tracks. In particular, we propose a learning
method based on a form of input dropout which allows learning in an
unsupervised manner, only based on raw, occluded sensor data without access to
ground-truth annotations. We demonstrate our approach using a synthetic dataset
designed to mimic the task of tracking objects in 2D laser data -- as commonly
encountered in robotics applications -- and show that it learns to track many
dynamic objects despite occlusions and the presence of sensor noise.
| new_dataset | 0.965381 |
1603.02869 | Emanuele Lindo Secco | Daniel Elstob, Emanuele Lindo Secco | A Low Cost Eeg Based Bci Prosthetic Using Motor Imagery | International Journal of Information Technology Convergence and
Services (IJITCS) Vol.6, No.1,February 2016 | null | null | null | cs.HC | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | Brain Computer Interfaces (BCI) provide the opportunity to control external
devices using the brain ElectroEncephaloGram (EEG) signals. In this paper we
propose two software framework in order to control a 5 degree of freedom
robotic and prosthetic hand. Results are presented where an Emotiv Cognitive
Suite (i.e. the 1st framework) combined with an embedded software system (i.e.
an open source Arduino board) is able to control the hand through character
input associated with the taught actions of the suite. This system provides
evidence of the feasibility of brain signals being a viable approach to
controlling the chosen prosthetic. Results are then presented in the second
framework. This latter one allowed for the training and classification of EEG
signals for motor imagery tasks. When analysing the system, clear visual
representations of the performance and accuracy are presented in the results
using a confusion matrix, accuracy measurement and a feedback bar signifying
signal strength. Experiments with various acquisition datasets were carried out
and with a critical evaluation of the results given. Finally depending on the
classification of the brain signal a Python script outputs the driving command
to the Arduino to control the prosthetic. The proposed architecture performs
overall good results for the design and implementation of economically
convenient BCI and prosthesis.
| [
{
"version": "v1",
"created": "Wed, 9 Mar 2016 12:39:06 GMT"
}
] | 2016-03-10T00:00:00 | [
[
"Elstob",
"Daniel",
""
],
[
"Secco",
"Emanuele Lindo",
""
]
] | TITLE: A Low Cost Eeg Based Bci Prosthetic Using Motor Imagery
ABSTRACT: Brain Computer Interfaces (BCI) provide the opportunity to control external
devices using the brain ElectroEncephaloGram (EEG) signals. In this paper we
propose two software framework in order to control a 5 degree of freedom
robotic and prosthetic hand. Results are presented where an Emotiv Cognitive
Suite (i.e. the 1st framework) combined with an embedded software system (i.e.
an open source Arduino board) is able to control the hand through character
input associated with the taught actions of the suite. This system provides
evidence of the feasibility of brain signals being a viable approach to
controlling the chosen prosthetic. Results are then presented in the second
framework. This latter one allowed for the training and classification of EEG
signals for motor imagery tasks. When analysing the system, clear visual
representations of the performance and accuracy are presented in the results
using a confusion matrix, accuracy measurement and a feedback bar signifying
signal strength. Experiments with various acquisition datasets were carried out
and with a critical evaluation of the results given. Finally depending on the
classification of the brain signal a Python script outputs the driving command
to the Arduino to control the prosthetic. The proposed architecture performs
overall good results for the design and implementation of economically
convenient BCI and prosthesis.
| no_new_dataset | 0.949248 |
1312.6947 | Sourish Dasgupta | Sourish Dasgupta, Ankur Padia, Kushal Shah, Prasenjit Majumder | Formal Ontology Learning on Factual IS-A Corpus in English using
Description Logics | This paper has been withdrawn by the author due to requirement of
re-evaluation of results | null | null | null | cs.CL cs.AI | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | Ontology Learning (OL) is the computational task of generating a knowledge
base in the form of an ontology given an unstructured corpus whose content is
in natural language (NL). Several works can be found in this area most of which
are limited to statistical and lexico-syntactic pattern matching based
techniques Light-Weight OL. These techniques do not lead to very accurate
learning mostly because of several linguistic nuances in NL. Formal OL is an
alternative (less explored) methodology were deep linguistics analysis is made
using theory and tools found in computational linguistics to generate formal
axioms and definitions instead simply inducing a taxonomy. In this paper we
propose "Description Logic (DL)" based formal OL framework for learning factual
IS-A type sentences in English. We claim that semantic construction of IS-A
sentences is non trivial. Hence, we also claim that such sentences requires
special studies in the context of OL before any truly formal OL can be
proposed. We introduce a learner tool, called DLOL_IS-A, that generated such
ontologies in the owl format. We have adopted "Gold Standard" based OL
evaluation on IS-A rich WCL v.1.1 dataset and our own Community representative
IS-A dataset. We observed significant improvement of DLOL_IS-A when compared to
the light-weight OL tool Text2Onto and formal OL tool FRED.
| [
{
"version": "v1",
"created": "Wed, 25 Dec 2013 09:17:28 GMT"
},
{
"version": "v2",
"created": "Tue, 8 Mar 2016 05:20:47 GMT"
}
] | 2016-03-09T00:00:00 | [
[
"Dasgupta",
"Sourish",
""
],
[
"Padia",
"Ankur",
""
],
[
"Shah",
"Kushal",
""
],
[
"Majumder",
"Prasenjit",
""
]
] | TITLE: Formal Ontology Learning on Factual IS-A Corpus in English using
Description Logics
ABSTRACT: Ontology Learning (OL) is the computational task of generating a knowledge
base in the form of an ontology given an unstructured corpus whose content is
in natural language (NL). Several works can be found in this area most of which
are limited to statistical and lexico-syntactic pattern matching based
techniques Light-Weight OL. These techniques do not lead to very accurate
learning mostly because of several linguistic nuances in NL. Formal OL is an
alternative (less explored) methodology were deep linguistics analysis is made
using theory and tools found in computational linguistics to generate formal
axioms and definitions instead simply inducing a taxonomy. In this paper we
propose "Description Logic (DL)" based formal OL framework for learning factual
IS-A type sentences in English. We claim that semantic construction of IS-A
sentences is non trivial. Hence, we also claim that such sentences requires
special studies in the context of OL before any truly formal OL can be
proposed. We introduce a learner tool, called DLOL_IS-A, that generated such
ontologies in the owl format. We have adopted "Gold Standard" based OL
evaluation on IS-A rich WCL v.1.1 dataset and our own Community representative
IS-A dataset. We observed significant improvement of DLOL_IS-A when compared to
the light-weight OL tool Text2Onto and formal OL tool FRED.
| new_dataset | 0.878783 |
1601.04814 | Gianmarco De Francisci Morales | Gianmarco De Francisci Morales and Aristides Gionis | Streaming Similarity Self-Join | null | null | null | null | cs.DB cs.DS | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | We introduce and study the problem of computing the similarity self-join in a
streaming context (SSSJ), where the input is an unbounded stream of items
arriving continuously. The goal is to find all pairs of items in the stream
whose similarity is greater than a given threshold. The simplest formulation of
the problem requires unbounded memory, and thus, it is intractable. To make the
problem feasible, we introduce the notion of time-dependent similarity: the
similarity of two items decreases with the difference in their arrival time. By
leveraging the properties of this time-dependent similarity function, we design
two algorithmic frameworks to solve the sssj problem. The first one, MiniBatch
(MB), uses existing index-based filtering techniques for the static version of
the problem, and combines them in a pipeline. The second framework, Streaming
(STR), adds time filtering to the existing indexes, and integrates new
time-based bounds deeply in the working of the algorithms. We also introduce a
new indexing technique (L2), which is based on an existing state-of-the-art
indexing technique (L2AP), but is optimized for the streaming case. Extensive
experiments show that the STR algorithm, when instantiated with the L2 index,
is the most scalable option across a wide array of datasets and parameters.
| [
{
"version": "v1",
"created": "Tue, 19 Jan 2016 07:34:17 GMT"
},
{
"version": "v2",
"created": "Tue, 8 Mar 2016 09:14:27 GMT"
}
] | 2016-03-09T00:00:00 | [
[
"Morales",
"Gianmarco De Francisci",
""
],
[
"Gionis",
"Aristides",
""
]
] | TITLE: Streaming Similarity Self-Join
ABSTRACT: We introduce and study the problem of computing the similarity self-join in a
streaming context (SSSJ), where the input is an unbounded stream of items
arriving continuously. The goal is to find all pairs of items in the stream
whose similarity is greater than a given threshold. The simplest formulation of
the problem requires unbounded memory, and thus, it is intractable. To make the
problem feasible, we introduce the notion of time-dependent similarity: the
similarity of two items decreases with the difference in their arrival time. By
leveraging the properties of this time-dependent similarity function, we design
two algorithmic frameworks to solve the sssj problem. The first one, MiniBatch
(MB), uses existing index-based filtering techniques for the static version of
the problem, and combines them in a pipeline. The second framework, Streaming
(STR), adds time filtering to the existing indexes, and integrates new
time-based bounds deeply in the working of the algorithms. We also introduce a
new indexing technique (L2), which is based on an existing state-of-the-art
indexing technique (L2AP), but is optimized for the streaming case. Extensive
experiments show that the STR algorithm, when instantiated with the L2 index,
is the most scalable option across a wide array of datasets and parameters.
| no_new_dataset | 0.946745 |
1601.06892 | Kuldeep S Kulkarni Mr. | Kuldeep Kulkarni, Suhas Lohit, Pavan Turaga, Ronan Kerviche, Amit
Ashok | ReconNet: Non-Iterative Reconstruction of Images from Compressively
Sensed Random Measurements | Accepted at IEEE International Conference on Computer Vision and
Pattern Recognition (CVPR), 2016 | null | null | null | cs.CV | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | The goal of this paper is to present a non-iterative and more importantly an
extremely fast algorithm to reconstruct images from compressively sensed (CS)
random measurements. To this end, we propose a novel convolutional neural
network (CNN) architecture which takes in CS measurements of an image as input
and outputs an intermediate reconstruction. We call this network, ReconNet. The
intermediate reconstruction is fed into an off-the-shelf denoiser to obtain the
final reconstructed image. On a standard dataset of images we show significant
improvements in reconstruction results (both in terms of PSNR and time
complexity) over state-of-the-art iterative CS reconstruction algorithms at
various measurement rates. Further, through qualitative experiments on real
data collected using our block single pixel camera (SPC), we show that our
network is highly robust to sensor noise and can recover visually better
quality images than competitive algorithms at extremely low sensing rates of
0.1 and 0.04. To demonstrate that our algorithm can recover semantically
informative images even at a low measurement rate of 0.01, we present a very
robust proof of concept real-time visual tracking application.
| [
{
"version": "v1",
"created": "Tue, 26 Jan 2016 05:17:14 GMT"
},
{
"version": "v2",
"created": "Mon, 7 Mar 2016 23:31:08 GMT"
}
] | 2016-03-09T00:00:00 | [
[
"Kulkarni",
"Kuldeep",
""
],
[
"Lohit",
"Suhas",
""
],
[
"Turaga",
"Pavan",
""
],
[
"Kerviche",
"Ronan",
""
],
[
"Ashok",
"Amit",
""
]
] | TITLE: ReconNet: Non-Iterative Reconstruction of Images from Compressively
Sensed Random Measurements
ABSTRACT: The goal of this paper is to present a non-iterative and more importantly an
extremely fast algorithm to reconstruct images from compressively sensed (CS)
random measurements. To this end, we propose a novel convolutional neural
network (CNN) architecture which takes in CS measurements of an image as input
and outputs an intermediate reconstruction. We call this network, ReconNet. The
intermediate reconstruction is fed into an off-the-shelf denoiser to obtain the
final reconstructed image. On a standard dataset of images we show significant
improvements in reconstruction results (both in terms of PSNR and time
complexity) over state-of-the-art iterative CS reconstruction algorithms at
various measurement rates. Further, through qualitative experiments on real
data collected using our block single pixel camera (SPC), we show that our
network is highly robust to sensor noise and can recover visually better
quality images than competitive algorithms at extremely low sensing rates of
0.1 and 0.04. To demonstrate that our algorithm can recover semantically
informative images even at a low measurement rate of 0.01, we present a very
robust proof of concept real-time visual tracking application.
| no_new_dataset | 0.950227 |
1602.06866 | Hao Wu | Huijuan Shao, K.S.M. Tozammel Hossain, Hao Wu, Maleq Khan, Anil
Vullikanti, B. Aditya Prakash, Madhav Marathe and Naren Ramakrishnan | Forecasting the Flu: Designing Social Network Sensors for Epidemics | The conference version of the paper is submitted for publication | null | null | null | cs.SI physics.soc-ph | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | Early detection and modeling of a contagious epidemic can provide important
guidance about quelling the contagion, controlling its spread, or the effective
design of countermeasures. A topic of recent interest has been to design social
network sensors, i.e., identifying a small set of people who can be monitored
to provide insight into the emergence of an epidemic in a larger population. We
formally pose the problem of designing social network sensors for flu epidemics
and identify two different objectives that could be targeted in such sensor
design problems. Using the graph theoretic notion of dominators we develop an
efficient and effective heuristic for forecasting epidemics at lead time. Using
six city-scale datasets generated by extensive microscopic epidemiological
simulations involving millions of individuals, we illustrate the practical
applicability of our methods and show significant benefits (up to twenty-two
days more lead time) compared to other competitors. Most importantly, we
demonstrate the use of surrogates or proxies for policy makers for designing
social network sensors that require from nonintrusive knowledge of people to
more information on the relationship among people. The results show that the
more intrusive information we obtain, the longer lead time to predict the flu
outbreak up to nine days.
| [
{
"version": "v1",
"created": "Mon, 22 Feb 2016 17:32:31 GMT"
},
{
"version": "v2",
"created": "Thu, 25 Feb 2016 17:20:49 GMT"
},
{
"version": "v3",
"created": "Tue, 8 Mar 2016 16:46:37 GMT"
}
] | 2016-03-09T00:00:00 | [
[
"Shao",
"Huijuan",
""
],
[
"Hossain",
"K. S. M. Tozammel",
""
],
[
"Wu",
"Hao",
""
],
[
"Khan",
"Maleq",
""
],
[
"Vullikanti",
"Anil",
""
],
[
"Prakash",
"B. Aditya",
""
],
[
"Marathe",
"Madhav",
""
],
[
"Ramakrishnan",
"Naren",
""
]
] | TITLE: Forecasting the Flu: Designing Social Network Sensors for Epidemics
ABSTRACT: Early detection and modeling of a contagious epidemic can provide important
guidance about quelling the contagion, controlling its spread, or the effective
design of countermeasures. A topic of recent interest has been to design social
network sensors, i.e., identifying a small set of people who can be monitored
to provide insight into the emergence of an epidemic in a larger population. We
formally pose the problem of designing social network sensors for flu epidemics
and identify two different objectives that could be targeted in such sensor
design problems. Using the graph theoretic notion of dominators we develop an
efficient and effective heuristic for forecasting epidemics at lead time. Using
six city-scale datasets generated by extensive microscopic epidemiological
simulations involving millions of individuals, we illustrate the practical
applicability of our methods and show significant benefits (up to twenty-two
days more lead time) compared to other competitors. Most importantly, we
demonstrate the use of surrogates or proxies for policy makers for designing
social network sensors that require from nonintrusive knowledge of people to
more information on the relationship among people. The results show that the
more intrusive information we obtain, the longer lead time to predict the flu
outbreak up to nine days.
| no_new_dataset | 0.9463 |
1603.00893 | Boxiang Dong | Boxiang Dong, Hui Wendy Wang | Frequency-hiding Dependency-preserving Encryption for Outsourced
Databases | null | null | null | null | cs.DB cs.CR | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | The cloud paradigm enables users to outsource their data to computationally
powerful third-party service providers for data management. Many data
management tasks rely on the data dependencies in the outsourced data. This
raises an important issue of how the data owner can protect the sensitive
information in the outsourced data while preserving the data dependencies. In
this paper, we consider functional dependency FD, an important type of data
dependency. We design a FD-preserving encryption scheme, named F2, that enables
the service provider to discover the FDs from the encrypted dataset. We
consider the frequency analysis attack, and show that the F2 encryption scheme
can defend against the attack under Kerckhoff's principle with provable
guarantee. Our empirical study demonstrates the efficiency and effectiveness of
F2.
| [
{
"version": "v1",
"created": "Wed, 2 Mar 2016 21:20:16 GMT"
},
{
"version": "v2",
"created": "Tue, 8 Mar 2016 15:16:00 GMT"
}
] | 2016-03-09T00:00:00 | [
[
"Dong",
"Boxiang",
""
],
[
"Wang",
"Hui Wendy",
""
]
] | TITLE: Frequency-hiding Dependency-preserving Encryption for Outsourced
Databases
ABSTRACT: The cloud paradigm enables users to outsource their data to computationally
powerful third-party service providers for data management. Many data
management tasks rely on the data dependencies in the outsourced data. This
raises an important issue of how the data owner can protect the sensitive
information in the outsourced data while preserving the data dependencies. In
this paper, we consider functional dependency FD, an important type of data
dependency. We design a FD-preserving encryption scheme, named F2, that enables
the service provider to discover the FDs from the encrypted dataset. We
consider the frequency analysis attack, and show that the F2 encryption scheme
can defend against the attack under Kerckhoff's principle with provable
guarantee. Our empirical study demonstrates the efficiency and effectiveness of
F2.
| no_new_dataset | 0.949576 |
1603.01670 | Tao Wei | Tao Wei, Changhu Wang, Yong Rui, Chang Wen Chen | Network Morphism | Under review for ICML 2016 | null | null | null | cs.LG cs.CV cs.NE | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | We present in this paper a systematic study on how to morph a well-trained
neural network to a new one so that its network function can be completely
preserved. We define this as \emph{network morphism} in this research. After
morphing a parent network, the child network is expected to inherit the
knowledge from its parent network and also has the potential to continue
growing into a more powerful one with much shortened training time. The first
requirement for this network morphism is its ability to handle diverse morphing
types of networks, including changes of depth, width, kernel size, and even
subnet. To meet this requirement, we first introduce the network morphism
equations, and then develop novel morphing algorithms for all these morphing
types for both classic and convolutional neural networks. The second
requirement for this network morphism is its ability to deal with non-linearity
in a network. We propose a family of parametric-activation functions to
facilitate the morphing of any continuous non-linear activation neurons.
Experimental results on benchmark datasets and typical neural networks
demonstrate the effectiveness of the proposed network morphism scheme.
| [
{
"version": "v1",
"created": "Sat, 5 Mar 2016 02:06:43 GMT"
},
{
"version": "v2",
"created": "Tue, 8 Mar 2016 16:36:00 GMT"
}
] | 2016-03-09T00:00:00 | [
[
"Wei",
"Tao",
""
],
[
"Wang",
"Changhu",
""
],
[
"Rui",
"Yong",
""
],
[
"Chen",
"Chang Wen",
""
]
] | TITLE: Network Morphism
ABSTRACT: We present in this paper a systematic study on how to morph a well-trained
neural network to a new one so that its network function can be completely
preserved. We define this as \emph{network morphism} in this research. After
morphing a parent network, the child network is expected to inherit the
knowledge from its parent network and also has the potential to continue
growing into a more powerful one with much shortened training time. The first
requirement for this network morphism is its ability to handle diverse morphing
types of networks, including changes of depth, width, kernel size, and even
subnet. To meet this requirement, we first introduce the network morphism
equations, and then develop novel morphing algorithms for all these morphing
types for both classic and convolutional neural networks. The second
requirement for this network morphism is its ability to deal with non-linearity
in a network. We propose a family of parametric-activation functions to
facilitate the morphing of any continuous non-linear activation neurons.
Experimental results on benchmark datasets and typical neural networks
demonstrate the effectiveness of the proposed network morphism scheme.
| no_new_dataset | 0.953923 |
1603.02494 | Tapesh Santra | Tapesh Santra | A Bayesian non-parametric method for clustering high-dimensional binary
data | null | null | null | null | stat.AP cs.LG stat.ML | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | In many real life problems, objects are described by large number of binary
features. For instance, documents are characterized by presence or absence of
certain keywords; cancer patients are characterized by presence or absence of
certain mutations etc. In such cases, grouping together similar
objects/profiles based on such high dimensional binary features is desirable,
but challenging. Here, I present a Bayesian non parametric algorithm for
clustering high dimensional binary data. It uses a Dirichlet Process (DP)
mixture model and simulated annealing to not only cluster binary data, but also
find optimal number of clusters in the data. The performance of the algorithm
was evaluated and compared with other algorithms using simulated datasets. It
outperformed all other clustering methods that were tested in the simulation
studies. It was also used to cluster real datasets arising from document
analysis, handwritten image analysis and cancer research. It successfully
divided a set of documents based on their topics, hand written images based on
different styles of writing digits and identified tissue and mutation
specificity of chemotherapy treatments.
| [
{
"version": "v1",
"created": "Tue, 8 Mar 2016 12:02:59 GMT"
}
] | 2016-03-09T00:00:00 | [
[
"Santra",
"Tapesh",
""
]
] | TITLE: A Bayesian non-parametric method for clustering high-dimensional binary
data
ABSTRACT: In many real life problems, objects are described by large number of binary
features. For instance, documents are characterized by presence or absence of
certain keywords; cancer patients are characterized by presence or absence of
certain mutations etc. In such cases, grouping together similar
objects/profiles based on such high dimensional binary features is desirable,
but challenging. Here, I present a Bayesian non parametric algorithm for
clustering high dimensional binary data. It uses a Dirichlet Process (DP)
mixture model and simulated annealing to not only cluster binary data, but also
find optimal number of clusters in the data. The performance of the algorithm
was evaluated and compared with other algorithms using simulated datasets. It
outperformed all other clustering methods that were tested in the simulation
studies. It was also used to cluster real datasets arising from document
analysis, handwritten image analysis and cancer research. It successfully
divided a set of documents based on their topics, hand written images based on
different styles of writing digits and identified tissue and mutation
specificity of chemotherapy treatments.
| no_new_dataset | 0.953923 |
1505.06236 | Le Lu | Amal Farag, Le Lu, Holger R. Roth, Jiamin Liu, Evrim Turkbey, Ronald
M. Summers | A Bottom-up Approach for Pancreas Segmentation using Cascaded
Superpixels and (Deep) Image Patch Labeling | 14 pages, 14 figures, 2 tables | null | null | null | cs.CV | http://creativecommons.org/publicdomain/zero/1.0/ | Robust automated organ segmentation is a prerequisite for computer-aided
diagnosis (CAD), quantitative imaging analysis and surgical assistance. For
high-variability organs such as the pancreas, previous approaches report
undesirably low accuracies. We present a bottom-up approach for pancreas
segmentation in abdominal CT scans that is based on a hierarchy of information
propagation by classifying image patches at different resolutions; and
cascading superpixels. There are four stages: 1) decomposing CT slice images as
a set of disjoint boundary-preserving superpixels; 2) computing pancreas class
probability maps via dense patch labeling; 3) classifying superpixels by
pooling both intensity and probability features to form empirical statistics in
cascaded random forest frameworks; and 4) simple connectivity based
post-processing. The dense image patch labeling are conducted by: efficient
random forest classifier on image histogram, location and texture features; and
more expensive (but with better specificity) deep convolutional neural network
classification on larger image windows (with more spatial contexts). Evaluation
of the approach is performed on a database of 80 manually segmented CT volumes
in six-fold cross-validation (CV). Our achieved results are comparable, or
better than the state-of-the-art methods (evaluated by
"leave-one-patient-out"), with Dice 70.7% and Jaccard 57.9%. The computational
efficiency has been drastically improved in the order of 6~8 minutes, comparing
with others of ~10 hours per case. Finally, we implement a multi-atlas label
fusion (MALF) approach for pancreas segmentation using the same datasets. Under
six-fold CV, our bottom-up segmentation method significantly outperforms its
MALF counterpart: (70.7 +/- 13.0%) versus (52.5 +/- 20.8%) in Dice. Deep CNN
patch labeling confidences offer more numerical stability, reflected by smaller
standard deviations.
| [
{
"version": "v1",
"created": "Fri, 22 May 2015 21:59:45 GMT"
},
{
"version": "v2",
"created": "Mon, 7 Mar 2016 18:24:43 GMT"
}
] | 2016-03-08T00:00:00 | [
[
"Farag",
"Amal",
""
],
[
"Lu",
"Le",
""
],
[
"Roth",
"Holger R.",
""
],
[
"Liu",
"Jiamin",
""
],
[
"Turkbey",
"Evrim",
""
],
[
"Summers",
"Ronald M.",
""
]
] | TITLE: A Bottom-up Approach for Pancreas Segmentation using Cascaded
Superpixels and (Deep) Image Patch Labeling
ABSTRACT: Robust automated organ segmentation is a prerequisite for computer-aided
diagnosis (CAD), quantitative imaging analysis and surgical assistance. For
high-variability organs such as the pancreas, previous approaches report
undesirably low accuracies. We present a bottom-up approach for pancreas
segmentation in abdominal CT scans that is based on a hierarchy of information
propagation by classifying image patches at different resolutions; and
cascading superpixels. There are four stages: 1) decomposing CT slice images as
a set of disjoint boundary-preserving superpixels; 2) computing pancreas class
probability maps via dense patch labeling; 3) classifying superpixels by
pooling both intensity and probability features to form empirical statistics in
cascaded random forest frameworks; and 4) simple connectivity based
post-processing. The dense image patch labeling are conducted by: efficient
random forest classifier on image histogram, location and texture features; and
more expensive (but with better specificity) deep convolutional neural network
classification on larger image windows (with more spatial contexts). Evaluation
of the approach is performed on a database of 80 manually segmented CT volumes
in six-fold cross-validation (CV). Our achieved results are comparable, or
better than the state-of-the-art methods (evaluated by
"leave-one-patient-out"), with Dice 70.7% and Jaccard 57.9%. The computational
efficiency has been drastically improved in the order of 6~8 minutes, comparing
with others of ~10 hours per case. Finally, we implement a multi-atlas label
fusion (MALF) approach for pancreas segmentation using the same datasets. Under
six-fold CV, our bottom-up segmentation method significantly outperforms its
MALF counterpart: (70.7 +/- 13.0%) versus (52.5 +/- 20.8%) in Dice. Deep CNN
patch labeling confidences offer more numerical stability, reflected by smaller
standard deviations.
| no_new_dataset | 0.954095 |
1506.07285 | Richard Socher | Ankit Kumar and Ozan Irsoy and Peter Ondruska and Mohit Iyyer and
James Bradbury and Ishaan Gulrajani and Victor Zhong and Romain Paulus and
Richard Socher | Ask Me Anything: Dynamic Memory Networks for Natural Language Processing | null | null | null | null | cs.CL cs.LG cs.NE | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | Most tasks in natural language processing can be cast into question answering
(QA) problems over language input. We introduce the dynamic memory network
(DMN), a neural network architecture which processes input sequences and
questions, forms episodic memories, and generates relevant answers. Questions
trigger an iterative attention process which allows the model to condition its
attention on the inputs and the result of previous iterations. These results
are then reasoned over in a hierarchical recurrent sequence model to generate
answers. The DMN can be trained end-to-end and obtains state-of-the-art results
on several types of tasks and datasets: question answering (Facebook's bAbI
dataset), text classification for sentiment analysis (Stanford Sentiment
Treebank) and sequence modeling for part-of-speech tagging (WSJ-PTB). The
training for these different tasks relies exclusively on trained word vector
representations and input-question-answer triplets.
| [
{
"version": "v1",
"created": "Wed, 24 Jun 2015 08:27:02 GMT"
},
{
"version": "v2",
"created": "Fri, 24 Jul 2015 22:21:29 GMT"
},
{
"version": "v3",
"created": "Tue, 29 Sep 2015 05:02:29 GMT"
},
{
"version": "v4",
"created": "Tue, 9 Feb 2016 08:19:30 GMT"
},
{
"version": "v5",
"created": "Sat, 5 Mar 2016 20:18:55 GMT"
}
] | 2016-03-08T00:00:00 | [
[
"Kumar",
"Ankit",
""
],
[
"Irsoy",
"Ozan",
""
],
[
"Ondruska",
"Peter",
""
],
[
"Iyyer",
"Mohit",
""
],
[
"Bradbury",
"James",
""
],
[
"Gulrajani",
"Ishaan",
""
],
[
"Zhong",
"Victor",
""
],
[
"Paulus",
"Romain",
""
],
[
"Socher",
"Richard",
""
]
] | TITLE: Ask Me Anything: Dynamic Memory Networks for Natural Language Processing
ABSTRACT: Most tasks in natural language processing can be cast into question answering
(QA) problems over language input. We introduce the dynamic memory network
(DMN), a neural network architecture which processes input sequences and
questions, forms episodic memories, and generates relevant answers. Questions
trigger an iterative attention process which allows the model to condition its
attention on the inputs and the result of previous iterations. These results
are then reasoned over in a hierarchical recurrent sequence model to generate
answers. The DMN can be trained end-to-end and obtains state-of-the-art results
on several types of tasks and datasets: question answering (Facebook's bAbI
dataset), text classification for sentiment analysis (Stanford Sentiment
Treebank) and sequence modeling for part-of-speech tagging (WSJ-PTB). The
training for these different tasks relies exclusively on trained word vector
representations and input-question-answer triplets.
| no_new_dataset | 0.948537 |
1508.07630 | Vural Aksakalli | Vural Aksakalli and Milad Malekipirbazari | Feature Selection via Binary Simultaneous Perturbation Stochastic
Approximation | This is the Istanbul Sehir University Technical Report
#SHR-ISE-2016.01. A short version of this report has been accepted for
publication at Pattern Recognition Letters | null | null | SHR-ISE-2016.01 | stat.ML cs.LG | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | Feature selection (FS) has become an indispensable task in dealing with
today's highly complex pattern recognition problems with massive number of
features. In this study, we propose a new wrapper approach for FS based on
binary simultaneous perturbation stochastic approximation (BSPSA). This
pseudo-gradient descent stochastic algorithm starts with an initial feature
vector and moves toward the optimal feature vector via successive iterations.
In each iteration, the current feature vector's individual components are
perturbed simultaneously by random offsets from a qualified probability
distribution. We present computational experiments on datasets with numbers of
features ranging from a few dozens to thousands using three widely-used
classifiers as wrappers: nearest neighbor, decision tree, and linear support
vector machine. We compare our methodology against the full set of features as
well as a binary genetic algorithm and sequential FS methods using
cross-validated classification error rate and AUC as the performance criteria.
Our results indicate that features selected by BSPSA compare favorably to
alternative methods in general and BSPSA can yield superior feature sets for
datasets with tens of thousands of features by examining an extremely small
fraction of the solution space. We are not aware of any other wrapper FS
methods that are computationally feasible with good convergence properties for
such large datasets.
| [
{
"version": "v1",
"created": "Sun, 30 Aug 2015 20:03:53 GMT"
},
{
"version": "v2",
"created": "Thu, 14 Jan 2016 08:02:42 GMT"
},
{
"version": "v3",
"created": "Sat, 5 Mar 2016 19:42:14 GMT"
}
] | 2016-03-08T00:00:00 | [
[
"Aksakalli",
"Vural",
""
],
[
"Malekipirbazari",
"Milad",
""
]
] | TITLE: Feature Selection via Binary Simultaneous Perturbation Stochastic
Approximation
ABSTRACT: Feature selection (FS) has become an indispensable task in dealing with
today's highly complex pattern recognition problems with massive number of
features. In this study, we propose a new wrapper approach for FS based on
binary simultaneous perturbation stochastic approximation (BSPSA). This
pseudo-gradient descent stochastic algorithm starts with an initial feature
vector and moves toward the optimal feature vector via successive iterations.
In each iteration, the current feature vector's individual components are
perturbed simultaneously by random offsets from a qualified probability
distribution. We present computational experiments on datasets with numbers of
features ranging from a few dozens to thousands using three widely-used
classifiers as wrappers: nearest neighbor, decision tree, and linear support
vector machine. We compare our methodology against the full set of features as
well as a binary genetic algorithm and sequential FS methods using
cross-validated classification error rate and AUC as the performance criteria.
Our results indicate that features selected by BSPSA compare favorably to
alternative methods in general and BSPSA can yield superior feature sets for
datasets with tens of thousands of features by examining an extremely small
fraction of the solution space. We are not aware of any other wrapper FS
methods that are computationally feasible with good convergence properties for
such large datasets.
| no_new_dataset | 0.947914 |
1509.03611 | Ella Rabinovich | Ella Rabinovich, Shuly Wintner, Ofek Luis Lewinsohn | A Parallel Corpus of Translationese | null | null | null | null | cs.CL | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | We describe a set of bilingual English--French and English--German parallel
corpora in which the direction of translation is accurately and reliably
annotated. The corpora are diverse, consisting of parliamentary proceedings,
literary works, transcriptions of TED talks and political commentary. They will
be instrumental for research of translationese and its applications to (human
and machine) translation; specifically, they can be used for the task of
translationese identification, a research direction that enjoys a growing
interest in recent years. To validate the quality and reliability of the
corpora, we replicated previous results of supervised and unsupervised
identification of translationese, and further extended the experiments to
additional datasets and languages.
| [
{
"version": "v1",
"created": "Fri, 11 Sep 2015 19:07:49 GMT"
},
{
"version": "v2",
"created": "Sun, 6 Mar 2016 13:41:11 GMT"
}
] | 2016-03-08T00:00:00 | [
[
"Rabinovich",
"Ella",
""
],
[
"Wintner",
"Shuly",
""
],
[
"Lewinsohn",
"Ofek Luis",
""
]
] | TITLE: A Parallel Corpus of Translationese
ABSTRACT: We describe a set of bilingual English--French and English--German parallel
corpora in which the direction of translation is accurately and reliably
annotated. The corpora are diverse, consisting of parliamentary proceedings,
literary works, transcriptions of TED talks and political commentary. They will
be instrumental for research of translationese and its applications to (human
and machine) translation; specifically, they can be used for the task of
translationese identification, a research direction that enjoys a growing
interest in recent years. To validate the quality and reliability of the
corpora, we replicated previous results of supervised and unsupervised
identification of translationese, and further extended the experiments to
additional datasets and languages.
| no_new_dataset | 0.717507 |
1511.02258 | Ze Jia Zhang | Z. Zhang, K. Duraisamy, N. A. Gumerov | Efficient Multiscale Gaussian Process Regression using Hierarchical
Clustering | 22 pages, 9 figures. Preprint. Submitted to Machine Learning Mar.
2016 | null | null | null | cs.LG stat.ML | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | Standard Gaussian Process (GP) regression, a powerful machine learning tool,
is computationally expensive when it is applied to large datasets, and
potentially inaccurate when data points are sparsely distributed in a
high-dimensional feature space. To address these challenges, a new multiscale,
sparsified GP algorithm is formulated, with the goal of application to large
scientific computing datasets. In this approach, the data is partitioned into
clusters and the cluster centers are used to define a reduced training set,
resulting in an improvement over standard GPs in terms of training and
evaluation costs. Further, a hierarchical technique is used to adaptively map
the local covariance representation to the underlying sparsity of the feature
space, leading to improved prediction accuracy when the data distribution is
highly non-uniform. A theoretical investigation of the computational complexity
of the algorithm is presented. The efficacy of this method is then demonstrated
on smooth and discontinuous analytical functions and on data from a direct
numerical simulation of turbulent combustion.
| [
{
"version": "v1",
"created": "Fri, 6 Nov 2015 23:18:13 GMT"
},
{
"version": "v2",
"created": "Mon, 7 Mar 2016 04:20:37 GMT"
}
] | 2016-03-08T00:00:00 | [
[
"Zhang",
"Z.",
""
],
[
"Duraisamy",
"K.",
""
],
[
"Gumerov",
"N. A.",
""
]
] | TITLE: Efficient Multiscale Gaussian Process Regression using Hierarchical
Clustering
ABSTRACT: Standard Gaussian Process (GP) regression, a powerful machine learning tool,
is computationally expensive when it is applied to large datasets, and
potentially inaccurate when data points are sparsely distributed in a
high-dimensional feature space. To address these challenges, a new multiscale,
sparsified GP algorithm is formulated, with the goal of application to large
scientific computing datasets. In this approach, the data is partitioned into
clusters and the cluster centers are used to define a reduced training set,
resulting in an improvement over standard GPs in terms of training and
evaluation costs. Further, a hierarchical technique is used to adaptively map
the local covariance representation to the underlying sparsity of the feature
space, leading to improved prediction accuracy when the data distribution is
highly non-uniform. A theoretical investigation of the computational complexity
of the algorithm is presented. The efficacy of this method is then demonstrated
on smooth and discontinuous analytical functions and on data from a direct
numerical simulation of turbulent combustion.
| no_new_dataset | 0.952574 |
1603.01716 | Balint Antal | B\'alint Antal | Classifier ensemble creation via false labelling | null | Knowledge-based Systems 89: pp. 278-287. (2015) | 10.1016/j.knosys.2015.07.009 | null | cs.LG | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | In this paper, a novel approach to classifier ensemble creation is presented.
While other ensemble creation techniques are based on careful selection of
existing classifiers or preprocessing of the data, the presented approach
automatically creates an optimal labelling for a number of classifiers, which
are then assigned to the original data instances and fed to classifiers. The
approach has been evaluated on high-dimensional biomedical datasets. The
results show that the approach outperformed individual approaches in all cases.
| [
{
"version": "v1",
"created": "Sat, 5 Mar 2016 12:01:00 GMT"
}
] | 2016-03-08T00:00:00 | [
[
"Antal",
"Bálint",
""
]
] | TITLE: Classifier ensemble creation via false labelling
ABSTRACT: In this paper, a novel approach to classifier ensemble creation is presented.
While other ensemble creation techniques are based on careful selection of
existing classifiers or preprocessing of the data, the presented approach
automatically creates an optimal labelling for a number of classifiers, which
are then assigned to the original data instances and fed to classifiers. The
approach has been evaluated on high-dimensional biomedical datasets. The
results show that the approach outperformed individual approaches in all cases.
| no_new_dataset | 0.955152 |
1603.01870 | Sougata Chaudhuri | Sougata Chaudhuri and Georgios Theocharous and Mohammad Ghavamzadeh | Personalized Advertisement Recommendation: A Ranking Approach to Address
the Ubiquitous Click Sparsity Problem | Under review | null | null | null | cs.LG cs.IR | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | We study the problem of personalized advertisement recommendation (PAR),
which consist of a user visiting a system (website) and the system displaying
one of $K$ ads to the user. The system uses an internal ad recommendation
policy to map the user's profile (context) to one of the ads. The user either
clicks or ignores the ad and correspondingly, the system updates its
recommendation policy. PAR problem is usually tackled by scalable
\emph{contextual bandit} algorithms, where the policies are generally based on
classifiers. A practical problem in PAR is extreme click sparsity, due to very
few users actually clicking on ads. We systematically study the drawback of
using contextual bandit algorithms based on classifier-based policies, in face
of extreme click sparsity. We then suggest an alternate policy, based on
rankers, learnt by optimizing the Area Under the Curve (AUC) ranking loss,
which can significantly alleviate the problem of click sparsity. We conduct
extensive experiments on public datasets, as well as three industry proprietary
datasets, to illustrate the improvement in click-through-rate (CTR) obtained by
using the ranker-based policy over classifier-based policies.
| [
{
"version": "v1",
"created": "Sun, 6 Mar 2016 20:26:41 GMT"
}
] | 2016-03-08T00:00:00 | [
[
"Chaudhuri",
"Sougata",
""
],
[
"Theocharous",
"Georgios",
""
],
[
"Ghavamzadeh",
"Mohammad",
""
]
] | TITLE: Personalized Advertisement Recommendation: A Ranking Approach to Address
the Ubiquitous Click Sparsity Problem
ABSTRACT: We study the problem of personalized advertisement recommendation (PAR),
which consist of a user visiting a system (website) and the system displaying
one of $K$ ads to the user. The system uses an internal ad recommendation
policy to map the user's profile (context) to one of the ads. The user either
clicks or ignores the ad and correspondingly, the system updates its
recommendation policy. PAR problem is usually tackled by scalable
\emph{contextual bandit} algorithms, where the policies are generally based on
classifiers. A practical problem in PAR is extreme click sparsity, due to very
few users actually clicking on ads. We systematically study the drawback of
using contextual bandit algorithms based on classifier-based policies, in face
of extreme click sparsity. We then suggest an alternate policy, based on
rankers, learnt by optimizing the Area Under the Curve (AUC) ranking loss,
which can significantly alleviate the problem of click sparsity. We conduct
extensive experiments on public datasets, as well as three industry proprietary
datasets, to illustrate the improvement in click-through-rate (CTR) obtained by
using the ranker-based policy over classifier-based policies.
| no_new_dataset | 0.944944 |
1603.01929 | Junting Ye | Junting Ye, Santhosh Kumar, Leman Akoglu | Temporal Opinion Spam Detection by Multivariate Indicative Signals | 10 pages, 29 figures | null | null | null | cs.SI | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | Online consumer reviews reflect the testimonials of real people, unlike
advertisements. As such, they have critical impact on potential consumers, and
indirectly on businesses. According to a Harvard study (Luca 2011), +1 rise in
star-rating increases revenue by 5-9%. Problematically, such financial
incentives have created a market for spammers to fabricate reviews, to unjustly
promote or demote businesses, activities known as opinion spam (Jindal and Liu
2008). A vast majority of existing work on this problem have formulations based
on static review data, with respective techniques operating in an offline
fashion. Spam campaigns, however, are intended to make most impact during their
course. Abnormal events triggered by spammers' activities could be masked in
the load of future events, which static analysis would fail to identify. In
this work, we approach the opinion spam problem with a temporal formulation.
Specifically, we monitor a list of carefully selected indicative signals of
opinion spam over time and design efficient techniques to both detect and
characterize abnormal events in real-time. Experiments on datasets from two
different review sites show that our approach is fast, effective, and practical
to be deployed in real-world systems.
| [
{
"version": "v1",
"created": "Mon, 7 Mar 2016 04:18:06 GMT"
}
] | 2016-03-08T00:00:00 | [
[
"Ye",
"Junting",
""
],
[
"Kumar",
"Santhosh",
""
],
[
"Akoglu",
"Leman",
""
]
] | TITLE: Temporal Opinion Spam Detection by Multivariate Indicative Signals
ABSTRACT: Online consumer reviews reflect the testimonials of real people, unlike
advertisements. As such, they have critical impact on potential consumers, and
indirectly on businesses. According to a Harvard study (Luca 2011), +1 rise in
star-rating increases revenue by 5-9%. Problematically, such financial
incentives have created a market for spammers to fabricate reviews, to unjustly
promote or demote businesses, activities known as opinion spam (Jindal and Liu
2008). A vast majority of existing work on this problem have formulations based
on static review data, with respective techniques operating in an offline
fashion. Spam campaigns, however, are intended to make most impact during their
course. Abnormal events triggered by spammers' activities could be masked in
the load of future events, which static analysis would fail to identify. In
this work, we approach the opinion spam problem with a temporal formulation.
Specifically, we monitor a list of carefully selected indicative signals of
opinion spam over time and design efficient techniques to both detect and
characterize abnormal events in real-time. Experiments on datasets from two
different review sites show that our approach is fast, effective, and practical
to be deployed in real-world systems.
| no_new_dataset | 0.937096 |
1603.02175 | Chunfeng Yang | Chunfeng Yang, Yipeng Zhou, Dah Ming Chiu | Who are Like-minded: Mining User Interest Similarity in Online Social
Networks | null | null | null | null | cs.SI | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | In this paper, we mine and learn to predict how similar a pair of users'
interests towards videos are, based on demographic (age, gender and location)
and social (friendship, interaction and group membership) information of these
users. We use the video access patterns of active users as ground truth (a form
of benchmark). We adopt tag-based user profiling to establish this ground
truth, and justify why it is used instead of video-based methods, or many
latent topic models such as LDA and Collaborative Filtering approaches. We then
show the effectiveness of the different demographic and social features, and
their combinations and derivatives, in predicting user interest similarity,
based on different machine-learning methods for combining multiple features. We
propose a hybrid tree-encoded linear model for combining the features, and show
that it out-performs other linear and treebased models. Our methods can be used
to predict user interest similarity when the ground-truth is not available,
e.g. for new users, or inactive users whose interests may have changed from old
access data, and is useful for video recommendation. Our study is based on a
rich dataset from Tencent, a popular service provider of social networks, video
services, and various other services in China.
| [
{
"version": "v1",
"created": "Mon, 7 Mar 2016 17:51:42 GMT"
}
] | 2016-03-08T00:00:00 | [
[
"Yang",
"Chunfeng",
""
],
[
"Zhou",
"Yipeng",
""
],
[
"Chiu",
"Dah Ming",
""
]
] | TITLE: Who are Like-minded: Mining User Interest Similarity in Online Social
Networks
ABSTRACT: In this paper, we mine and learn to predict how similar a pair of users'
interests towards videos are, based on demographic (age, gender and location)
and social (friendship, interaction and group membership) information of these
users. We use the video access patterns of active users as ground truth (a form
of benchmark). We adopt tag-based user profiling to establish this ground
truth, and justify why it is used instead of video-based methods, or many
latent topic models such as LDA and Collaborative Filtering approaches. We then
show the effectiveness of the different demographic and social features, and
their combinations and derivatives, in predicting user interest similarity,
based on different machine-learning methods for combining multiple features. We
propose a hybrid tree-encoded linear model for combining the features, and show
that it out-performs other linear and treebased models. Our methods can be used
to predict user interest similarity when the ground-truth is not available,
e.g. for new users, or inactive users whose interests may have changed from old
access data, and is useful for video recommendation. Our study is based on a
rich dataset from Tencent, a popular service provider of social networks, video
services, and various other services in China.
| no_new_dataset | 0.944485 |
1603.02211 | Rajesh Kumar | Rajesh Kumar, Vir V Phoha, and Rahul Raina | Authenticating users through their arm movement patterns | null | null | null | null | cs.CV cs.CR | http://creativecommons.org/licenses/by-nc-sa/4.0/ | In this paper, we propose four continuous authentication designs by using the
characteristics of arm movements while individuals walk. The first design uses
acceleration of arms captured by a smartwatch's accelerometer sensor, the
second design uses the rotation of arms captured by a smartwatch's gyroscope
sensor, third uses the fusion of both acceleration and rotation at the
feature-level and fourth uses the fusion at score-level. Each of these designs
is implemented by using four classifiers, namely, k nearest neighbors (k-NN)
with Euclidean distance, Logistic Regression, Multilayer Perceptrons, and
Random Forest resulting in a total of sixteen authentication mechanisms. These
authentication mechanisms are tested under three different environments, namely
an intra-session, inter-session on a dataset of 40 users and an inter-phase on
a dataset of 12 users. The sessions of data collection were separated by at
least ten minutes, whereas the phases of data collection were separated by at
least three months. Under the intra-session environment, all of the twelve
authentication mechanisms achieve a mean dynamic false accept rate (DFAR) of 0%
and dynamic false reject rate (DFRR) of 0%. For the inter-session environment,
feature level fusion-based design with classifier k-NN achieves the best error
rates that are a mean DFAR of 2.2% and DFRR of 4.2%. The DFAR and DFRR
increased from 5.68% and 4.23% to 15.03% and 14.62% respectively when feature
level fusion-based design with classifier k-NN was tested under the inter-phase
environment on a dataset of 12 users.
| [
{
"version": "v1",
"created": "Mon, 7 Mar 2016 19:15:39 GMT"
}
] | 2016-03-08T00:00:00 | [
[
"Kumar",
"Rajesh",
""
],
[
"Phoha",
"Vir V",
""
],
[
"Raina",
"Rahul",
""
]
] | TITLE: Authenticating users through their arm movement patterns
ABSTRACT: In this paper, we propose four continuous authentication designs by using the
characteristics of arm movements while individuals walk. The first design uses
acceleration of arms captured by a smartwatch's accelerometer sensor, the
second design uses the rotation of arms captured by a smartwatch's gyroscope
sensor, third uses the fusion of both acceleration and rotation at the
feature-level and fourth uses the fusion at score-level. Each of these designs
is implemented by using four classifiers, namely, k nearest neighbors (k-NN)
with Euclidean distance, Logistic Regression, Multilayer Perceptrons, and
Random Forest resulting in a total of sixteen authentication mechanisms. These
authentication mechanisms are tested under three different environments, namely
an intra-session, inter-session on a dataset of 40 users and an inter-phase on
a dataset of 12 users. The sessions of data collection were separated by at
least ten minutes, whereas the phases of data collection were separated by at
least three months. Under the intra-session environment, all of the twelve
authentication mechanisms achieve a mean dynamic false accept rate (DFAR) of 0%
and dynamic false reject rate (DFRR) of 0%. For the inter-session environment,
feature level fusion-based design with classifier k-NN achieves the best error
rates that are a mean DFAR of 2.2% and DFRR of 4.2%. The DFAR and DFRR
increased from 5.68% and 4.23% to 15.03% and 14.62% respectively when feature
level fusion-based design with classifier k-NN was tested under the inter-phase
environment on a dataset of 12 users.
| no_new_dataset | 0.947284 |
1504.06103 | Tomas Vojir | Tomas Vojir, Jiri Matas, Jana Noskova | Online Adaptive Hidden Markov Model for Multi-Tracker Fusion | 27 pages, 9 figures, submitted to CVIU journal | null | null | null | cs.CV | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | In this paper, we propose a novel method for visual object tracking called
HMMTxD. The method fuses observations from complementary out-of-the box
trackers and a detector by utilizing a hidden Markov model whose latent states
correspond to a binary vector expressing the failure of individual trackers.
The Markov model is trained in an unsupervised way, relying on an online
learned detector to provide a source of tracker-independent information for a
modified Baum- Welch algorithm that updates the model w.r.t. the partially
annotated data.
We show the effectiveness of the proposed method on combination of two and
three tracking algorithms. The performance of HMMTxD is evaluated on two
standard benchmarks (CVPR2013 and VOT) and on a rich collection of 77 publicly
available sequences. The HMMTxD outperforms the state-of-the-art, often
significantly, on all datasets in almost all criteria.
| [
{
"version": "v1",
"created": "Thu, 23 Apr 2015 09:34:59 GMT"
},
{
"version": "v2",
"created": "Fri, 4 Mar 2016 14:52:18 GMT"
}
] | 2016-03-07T00:00:00 | [
[
"Vojir",
"Tomas",
""
],
[
"Matas",
"Jiri",
""
],
[
"Noskova",
"Jana",
""
]
] | TITLE: Online Adaptive Hidden Markov Model for Multi-Tracker Fusion
ABSTRACT: In this paper, we propose a novel method for visual object tracking called
HMMTxD. The method fuses observations from complementary out-of-the box
trackers and a detector by utilizing a hidden Markov model whose latent states
correspond to a binary vector expressing the failure of individual trackers.
The Markov model is trained in an unsupervised way, relying on an online
learned detector to provide a source of tracker-independent information for a
modified Baum- Welch algorithm that updates the model w.r.t. the partially
annotated data.
We show the effectiveness of the proposed method on combination of two and
three tracking algorithms. The performance of HMMTxD is evaluated on two
standard benchmarks (CVPR2013 and VOT) and on a rich collection of 77 publicly
available sequences. The HMMTxD outperforms the state-of-the-art, often
significantly, on all datasets in almost all criteria.
| no_new_dataset | 0.945248 |
1506.08316 | Jianshu Chao | Jianshu Chao and Eckehard Steinbach | Keypoint Encoding for Improved Feature Extraction from Compressed Video
at Low Bitrates | null | null | null | null | cs.MM cs.CV cs.IR | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | In many mobile visual analysis applications, compressed video is transmitted
over a communication network and analyzed by a server. Typical processing steps
performed at the server include keypoint detection, descriptor calculation, and
feature matching. Video compression has been shown to have an adverse effect on
feature-matching performance. The negative impact of compression can be reduced
by using the keypoints extracted from the uncompressed video to calculate
descriptors from the compressed video. Based on this observation, we propose to
provide these keypoints to the server as side information and to extract only
the descriptors from the compressed video. First, we introduce four different
frame types for keypoint encoding to address different types of changes in
video content. These frame types represent a new scene, the same scene, a
slowly changing scene, or a rapidly moving scene and are determined by
comparing features between successive video frames. Then, we propose Intra,
Skip and Inter modes of encoding the keypoints for different frame types. For
example, keypoints for new scenes are encoded using the Intra mode, and
keypoints for unchanged scenes are skipped. As a result, the bitrate of the
side information related to keypoint encoding is significantly reduced.
Finally, we present pairwise matching and image retrieval experiments conducted
to evaluate the performance of the proposed approach using the Stanford mobile
augmented reality dataset and 720p format videos. The results show that the
proposed approach offers significantly improved feature matching and image
retrieval performance at a given bitrate.
| [
{
"version": "v1",
"created": "Sat, 27 Jun 2015 17:33:34 GMT"
},
{
"version": "v2",
"created": "Fri, 4 Mar 2016 16:43:42 GMT"
}
] | 2016-03-07T00:00:00 | [
[
"Chao",
"Jianshu",
""
],
[
"Steinbach",
"Eckehard",
""
]
] | TITLE: Keypoint Encoding for Improved Feature Extraction from Compressed Video
at Low Bitrates
ABSTRACT: In many mobile visual analysis applications, compressed video is transmitted
over a communication network and analyzed by a server. Typical processing steps
performed at the server include keypoint detection, descriptor calculation, and
feature matching. Video compression has been shown to have an adverse effect on
feature-matching performance. The negative impact of compression can be reduced
by using the keypoints extracted from the uncompressed video to calculate
descriptors from the compressed video. Based on this observation, we propose to
provide these keypoints to the server as side information and to extract only
the descriptors from the compressed video. First, we introduce four different
frame types for keypoint encoding to address different types of changes in
video content. These frame types represent a new scene, the same scene, a
slowly changing scene, or a rapidly moving scene and are determined by
comparing features between successive video frames. Then, we propose Intra,
Skip and Inter modes of encoding the keypoints for different frame types. For
example, keypoints for new scenes are encoded using the Intra mode, and
keypoints for unchanged scenes are skipped. As a result, the bitrate of the
side information related to keypoint encoding is significantly reduced.
Finally, we present pairwise matching and image retrieval experiments conducted
to evaluate the performance of the proposed approach using the Stanford mobile
augmented reality dataset and 720p format videos. The results show that the
proposed approach offers significantly improved feature matching and image
retrieval performance at a given bitrate.
| no_new_dataset | 0.954647 |
1509.02805 | Teng Qiu | Teng Qiu, Yongjie Li | Clustering by Hierarchical Nearest Neighbor Descent (H-NND) | 19 pages, 9 figures | null | null | null | stat.ML cs.CV cs.LG stat.ME | http://creativecommons.org/licenses/by-nc-sa/4.0/ | Previously in 2014, we proposed the Nearest Descent (ND) method, capable of
generating an efficient Graph, called the in-tree (IT). Due to some beautiful
and effective features, this IT structure proves well suited for data
clustering. Although there exist some redundant edges in IT, they usually have
salient features and thus it is not hard to remove them.
Subsequently, in order to prevent the seemingly redundant edges from
occurring, we proposed the Nearest Neighbor Descent (NND) by adding the
"Neighborhood" constraint on ND. Consequently, clusters automatically emerged,
without the additional requirement of removing the redundant edges. However,
NND proved still not perfect, since it brought in a new yet worse problem, the
"over-partitioning" problem.
Now, in this paper, we propose a method, called the Hierarchical Nearest
Neighbor Descent (H-NND), which overcomes the over-partitioning problem of NND
via using the hierarchical strategy. Specifically, H-NND uses ND to effectively
merge the over-segmented sub-graphs or clusters that NND produces. Like ND,
H-NND also generates the IT structure, in which the redundant edges once again
appear. This seemingly comes back to the situation that ND faces. However,
compared with ND, the redundant edges in the IT structure generated by H-NND
generally become more salient, thus being much easier and more reliable to be
identified even by the simplest edge-removing method which takes the edge
length as the only measure. In other words, the IT structure constructed by
H-NND becomes more fitted for data clustering. We prove this on several
clustering datasets of varying shapes, dimensions and attributes. Besides,
compared with ND, H-NND generally takes less computation time to construct the
IT data structure for the input data.
| [
{
"version": "v1",
"created": "Wed, 9 Sep 2015 15:15:44 GMT"
},
{
"version": "v2",
"created": "Mon, 14 Sep 2015 15:43:25 GMT"
},
{
"version": "v3",
"created": "Fri, 4 Mar 2016 15:50:58 GMT"
}
] | 2016-03-07T00:00:00 | [
[
"Qiu",
"Teng",
""
],
[
"Li",
"Yongjie",
""
]
] | TITLE: Clustering by Hierarchical Nearest Neighbor Descent (H-NND)
ABSTRACT: Previously in 2014, we proposed the Nearest Descent (ND) method, capable of
generating an efficient Graph, called the in-tree (IT). Due to some beautiful
and effective features, this IT structure proves well suited for data
clustering. Although there exist some redundant edges in IT, they usually have
salient features and thus it is not hard to remove them.
Subsequently, in order to prevent the seemingly redundant edges from
occurring, we proposed the Nearest Neighbor Descent (NND) by adding the
"Neighborhood" constraint on ND. Consequently, clusters automatically emerged,
without the additional requirement of removing the redundant edges. However,
NND proved still not perfect, since it brought in a new yet worse problem, the
"over-partitioning" problem.
Now, in this paper, we propose a method, called the Hierarchical Nearest
Neighbor Descent (H-NND), which overcomes the over-partitioning problem of NND
via using the hierarchical strategy. Specifically, H-NND uses ND to effectively
merge the over-segmented sub-graphs or clusters that NND produces. Like ND,
H-NND also generates the IT structure, in which the redundant edges once again
appear. This seemingly comes back to the situation that ND faces. However,
compared with ND, the redundant edges in the IT structure generated by H-NND
generally become more salient, thus being much easier and more reliable to be
identified even by the simplest edge-removing method which takes the edge
length as the only measure. In other words, the IT structure constructed by
H-NND becomes more fitted for data clustering. We prove this on several
clustering datasets of varying shapes, dimensions and attributes. Besides,
compared with ND, H-NND generally takes less computation time to construct the
IT data structure for the input data.
| no_new_dataset | 0.948442 |
1511.08198 | John Wieting | John Wieting, Mohit Bansal, Kevin Gimpel, Karen Livescu | Towards Universal Paraphrastic Sentence Embeddings | Published as a conference paper at ICLR 2016 | null | null | null | cs.CL cs.LG | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | We consider the problem of learning general-purpose, paraphrastic sentence
embeddings based on supervision from the Paraphrase Database (Ganitkevitch et
al., 2013). We compare six compositional architectures, evaluating them on
annotated textual similarity datasets drawn both from the same distribution as
the training data and from a wide range of other domains. We find that the most
complex architectures, such as long short-term memory (LSTM) recurrent neural
networks, perform best on the in-domain data. However, in out-of-domain
scenarios, simple architectures such as word averaging vastly outperform LSTMs.
Our simplest averaging model is even competitive with systems tuned for the
particular tasks while also being extremely efficient and easy to use.
In order to better understand how these architectures compare, we conduct
further experiments on three supervised NLP tasks: sentence similarity,
entailment, and sentiment classification. We again find that the word averaging
models perform well for sentence similarity and entailment, outperforming
LSTMs. However, on sentiment classification, we find that the LSTM performs
very strongly-even recording new state-of-the-art performance on the Stanford
Sentiment Treebank.
We then demonstrate how to combine our pretrained sentence embeddings with
these supervised tasks, using them both as a prior and as a black box feature
extractor. This leads to performance rivaling the state of the art on the SICK
similarity and entailment tasks. We release all of our resources to the
research community with the hope that they can serve as the new baseline for
further work on universal sentence embeddings.
| [
{
"version": "v1",
"created": "Wed, 25 Nov 2015 20:52:15 GMT"
},
{
"version": "v2",
"created": "Tue, 12 Jan 2016 20:59:39 GMT"
},
{
"version": "v3",
"created": "Fri, 4 Mar 2016 20:54:30 GMT"
}
] | 2016-03-07T00:00:00 | [
[
"Wieting",
"John",
""
],
[
"Bansal",
"Mohit",
""
],
[
"Gimpel",
"Kevin",
""
],
[
"Livescu",
"Karen",
""
]
] | TITLE: Towards Universal Paraphrastic Sentence Embeddings
ABSTRACT: We consider the problem of learning general-purpose, paraphrastic sentence
embeddings based on supervision from the Paraphrase Database (Ganitkevitch et
al., 2013). We compare six compositional architectures, evaluating them on
annotated textual similarity datasets drawn both from the same distribution as
the training data and from a wide range of other domains. We find that the most
complex architectures, such as long short-term memory (LSTM) recurrent neural
networks, perform best on the in-domain data. However, in out-of-domain
scenarios, simple architectures such as word averaging vastly outperform LSTMs.
Our simplest averaging model is even competitive with systems tuned for the
particular tasks while also being extremely efficient and easy to use.
In order to better understand how these architectures compare, we conduct
further experiments on three supervised NLP tasks: sentence similarity,
entailment, and sentiment classification. We again find that the word averaging
models perform well for sentence similarity and entailment, outperforming
LSTMs. However, on sentiment classification, we find that the LSTM performs
very strongly-even recording new state-of-the-art performance on the Stanford
Sentiment Treebank.
We then demonstrate how to combine our pretrained sentence embeddings with
these supervised tasks, using them both as a prior and as a black box feature
extractor. This leads to performance rivaling the state of the art on the SICK
similarity and entailment tasks. We release all of our resources to the
research community with the hope that they can serve as the new baseline for
further work on universal sentence embeddings.
| no_new_dataset | 0.944587 |
1603.01336 | Sabir Ribas | Sabir Ribas, Alberto Ueda, Rodrygo L. T. Santos, Berthier
Ribeiro-Neto, Nivio Ziviani | Simplified Relative Citation Ratio for Static Paper Ranking: UFMG/LATIN
at WSDM Cup 2016 | WSDM Cup. The 9th ACM International Conference on Web Search and Data
Mining San Francisco, California, USA. February 22-25, 2016 | null | null | null | cs.IR cs.DL | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | Static rankings of papers play a key role in the academic search setting.
Many features are commonly used in the literature to produce such rankings,
some examples are citation-based metrics, distinct applications of PageRank,
among others. More recently, learning to rank techniques have been successfully
applied to combine sets of features producing effective results. In this work,
we propose the metric S-RCR, which is a simplified version of a metric called
Relative Citation Ratio --- both based on the idea of a co-citation network.
When compared to the classical version, our simplification S-RCR leads to
improved efficiency with a reasonable effectiveness. We use S-RCR to rank over
120 million papers in the Microsoft Academic Graph dataset. By using this
single feature, which has no parameters and does not need to be tuned, our team
was able to reach the 3rd position in the first phase of the WSDM Cup 2016.
| [
{
"version": "v1",
"created": "Fri, 4 Mar 2016 03:00:46 GMT"
}
] | 2016-03-07T00:00:00 | [
[
"Ribas",
"Sabir",
""
],
[
"Ueda",
"Alberto",
""
],
[
"Santos",
"Rodrygo L. T.",
""
],
[
"Ribeiro-Neto",
"Berthier",
""
],
[
"Ziviani",
"Nivio",
""
]
] | TITLE: Simplified Relative Citation Ratio for Static Paper Ranking: UFMG/LATIN
at WSDM Cup 2016
ABSTRACT: Static rankings of papers play a key role in the academic search setting.
Many features are commonly used in the literature to produce such rankings,
some examples are citation-based metrics, distinct applications of PageRank,
among others. More recently, learning to rank techniques have been successfully
applied to combine sets of features producing effective results. In this work,
we propose the metric S-RCR, which is a simplified version of a metric called
Relative Citation Ratio --- both based on the idea of a co-citation network.
When compared to the classical version, our simplification S-RCR leads to
improved efficiency with a reasonable effectiveness. We use S-RCR to rank over
120 million papers in the Microsoft Academic Graph dataset. By using this
single feature, which has no parameters and does not need to be tuned, our team
was able to reach the 3rd position in the first phase of the WSDM Cup 2016.
| no_new_dataset | 0.9434 |
1603.01417 | Richard Socher | Caiming Xiong, Stephen Merity, Richard Socher | Dynamic Memory Networks for Visual and Textual Question Answering | null | null | null | null | cs.NE cs.CL cs.CV | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | Neural network architectures with memory and attention mechanisms exhibit
certain reasoning capabilities required for question answering. One such
architecture, the dynamic memory network (DMN), obtained high accuracy on a
variety of language tasks. However, it was not shown whether the architecture
achieves strong results for question answering when supporting facts are not
marked during training or whether it could be applied to other modalities such
as images. Based on an analysis of the DMN, we propose several improvements to
its memory and input modules. Together with these changes we introduce a novel
input module for images in order to be able to answer visual questions. Our new
DMN+ model improves the state of the art on both the Visual Question Answering
dataset and the \babi-10k text question-answering dataset without supporting
fact supervision.
| [
{
"version": "v1",
"created": "Fri, 4 Mar 2016 10:40:28 GMT"
}
] | 2016-03-07T00:00:00 | [
[
"Xiong",
"Caiming",
""
],
[
"Merity",
"Stephen",
""
],
[
"Socher",
"Richard",
""
]
] | TITLE: Dynamic Memory Networks for Visual and Textual Question Answering
ABSTRACT: Neural network architectures with memory and attention mechanisms exhibit
certain reasoning capabilities required for question answering. One such
architecture, the dynamic memory network (DMN), obtained high accuracy on a
variety of language tasks. However, it was not shown whether the architecture
achieves strong results for question answering when supporting facts are not
marked during training or whether it could be applied to other modalities such
as images. Based on an analysis of the DMN, we propose several improvements to
its memory and input modules. Together with these changes we introduce a novel
input module for images in order to be able to answer visual questions. Our new
DMN+ model improves the state of the art on both the Visual Question Answering
dataset and the \babi-10k text question-answering dataset without supporting
fact supervision.
| no_new_dataset | 0.947575 |
1603.00892 | Nikola Mrk\v{s}i\'c | Nikola Mrk\v{s}i\'c and Diarmuid \'O S\'eaghdha and Blaise Thomson and
Milica Ga\v{s}i\'c and Lina Rojas-Barahona and Pei-Hao Su and David Vandyke
and Tsung-Hsien Wen and Steve Young | Counter-fitting Word Vectors to Linguistic Constraints | Paper accepted for the 15th Annual Conference of the North American
Chapter of the Association for Computational Linguistics (NAACL 2016) | null | null | null | cs.CL cs.LG | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | In this work, we present a novel counter-fitting method which injects
antonymy and synonymy constraints into vector space representations in order to
improve the vectors' capability for judging semantic similarity. Applying this
method to publicly available pre-trained word vectors leads to a new state of
the art performance on the SimLex-999 dataset. We also show how the method can
be used to tailor the word vector space for the downstream task of dialogue
state tracking, resulting in robust improvements across different dialogue
domains.
| [
{
"version": "v1",
"created": "Wed, 2 Mar 2016 21:19:36 GMT"
}
] | 2016-03-04T00:00:00 | [
[
"Mrkšić",
"Nikola",
""
],
[
"Séaghdha",
"Diarmuid Ó",
""
],
[
"Thomson",
"Blaise",
""
],
[
"Gašić",
"Milica",
""
],
[
"Rojas-Barahona",
"Lina",
""
],
[
"Su",
"Pei-Hao",
""
],
[
"Vandyke",
"David",
""
],
[
"Wen",
"Tsung-Hsien",
""
],
[
"Young",
"Steve",
""
]
] | TITLE: Counter-fitting Word Vectors to Linguistic Constraints
ABSTRACT: In this work, we present a novel counter-fitting method which injects
antonymy and synonymy constraints into vector space representations in order to
improve the vectors' capability for judging semantic similarity. Applying this
method to publicly available pre-trained word vectors leads to a new state of
the art performance on the SimLex-999 dataset. We also show how the method can
be used to tailor the word vector space for the downstream task of dialogue
state tracking, resulting in robust improvements across different dialogue
domains.
| no_new_dataset | 0.952662 |
1603.01046 | Teemu Helenius | Teemu Helenius and Samuli Siltanen | Photographic dataset: random peppercorns | null | null | null | null | physics.data-an cs.CV | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | This is a photographic dataset collected for testing image processing
algorithms. The idea is to have sets of different but statistically similar
images. In this work the images show randomly distributed peppercorns. The
dataset is made available at www.fips.fi/photographic_dataset.php .
| [
{
"version": "v1",
"created": "Thu, 3 Mar 2016 10:24:07 GMT"
}
] | 2016-03-04T00:00:00 | [
[
"Helenius",
"Teemu",
""
],
[
"Siltanen",
"Samuli",
""
]
] | TITLE: Photographic dataset: random peppercorns
ABSTRACT: This is a photographic dataset collected for testing image processing
algorithms. The idea is to have sets of different but statistically similar
images. In this work the images show randomly distributed peppercorns. The
dataset is made available at www.fips.fi/photographic_dataset.php .
| new_dataset | 0.953579 |
1603.01232 | Tsung-Hsien Wen | Tsung-Hsien Wen, Milica Gasic, Nikola Mrksic, Lina M. Rojas-Barahona,
Pei-Hao Su, David Vandyke, Steve Young | Multi-domain Neural Network Language Generation for Spoken Dialogue
Systems | Accepted as a long paper in NAACL-HLT 2016 | null | null | null | cs.CL | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | Moving from limited-domain natural language generation (NLG) to open domain
is difficult because the number of semantic input combinations grows
exponentially with the number of domains. Therefore, it is important to
leverage existing resources and exploit similarities between domains to
facilitate domain adaptation. In this paper, we propose a procedure to train
multi-domain, Recurrent Neural Network-based (RNN) language generators via
multiple adaptation steps. In this procedure, a model is first trained on
counterfeited data synthesised from an out-of-domain dataset, and then fine
tuned on a small set of in-domain utterances with a discriminative objective
function. Corpus-based evaluation results show that the proposed procedure can
achieve competitive performance in terms of BLEU score and slot error rate
while significantly reducing the data needed to train generators in new, unseen
domains. In subjective testing, human judges confirm that the procedure greatly
improves generator performance when only a small amount of data is available in
the domain.
| [
{
"version": "v1",
"created": "Thu, 3 Mar 2016 19:49:32 GMT"
}
] | 2016-03-04T00:00:00 | [
[
"Wen",
"Tsung-Hsien",
""
],
[
"Gasic",
"Milica",
""
],
[
"Mrksic",
"Nikola",
""
],
[
"Rojas-Barahona",
"Lina M.",
""
],
[
"Su",
"Pei-Hao",
""
],
[
"Vandyke",
"David",
""
],
[
"Young",
"Steve",
""
]
] | TITLE: Multi-domain Neural Network Language Generation for Spoken Dialogue
Systems
ABSTRACT: Moving from limited-domain natural language generation (NLG) to open domain
is difficult because the number of semantic input combinations grows
exponentially with the number of domains. Therefore, it is important to
leverage existing resources and exploit similarities between domains to
facilitate domain adaptation. In this paper, we propose a procedure to train
multi-domain, Recurrent Neural Network-based (RNN) language generators via
multiple adaptation steps. In this procedure, a model is first trained on
counterfeited data synthesised from an out-of-domain dataset, and then fine
tuned on a small set of in-domain utterances with a discriminative objective
function. Corpus-based evaluation results show that the proposed procedure can
achieve competitive performance in terms of BLEU score and slot error rate
while significantly reducing the data needed to train generators in new, unseen
domains. In subjective testing, human judges confirm that the procedure greatly
improves generator performance when only a small amount of data is available in
the domain.
| no_new_dataset | 0.951414 |
1603.01250 | Yani Ioannou | Yani Ioannou, Duncan Robertson, Darko Zikic, Peter Kontschieder, Jamie
Shotton, Matthew Brown, and Antonio Criminisi | Decision Forests, Convolutional Networks and the Models in-Between | Microsoft Research Technical Report | null | null | MSR-TR-2015-58 | cs.CV cs.AI | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | This paper investigates the connections between two state of the art
classifiers: decision forests (DFs, including decision jungles) and
convolutional neural networks (CNNs). Decision forests are computationally
efficient thanks to their conditional computation property (computation is
confined to only a small region of the tree, the nodes along a single branch).
CNNs achieve state of the art accuracy, thanks to their representation learning
capabilities. We present a systematic analysis of how to fuse conditional
computation with representation learning and achieve a continuum of hybrid
models with different ratios of accuracy vs. efficiency. We call this new
family of hybrid models conditional networks. Conditional networks can be
thought of as: i) decision trees augmented with data transformation operators,
or ii) CNNs, with block-diagonal sparse weight matrices, and explicit data
routing functions. Experimental validation is performed on the common task of
image classification on both the CIFAR and Imagenet datasets. Compared to state
of the art CNNs, our hybrid models yield the same accuracy with a fraction of
the compute cost and much smaller number of parameters.
| [
{
"version": "v1",
"created": "Thu, 3 Mar 2016 20:41:47 GMT"
}
] | 2016-03-04T00:00:00 | [
[
"Ioannou",
"Yani",
""
],
[
"Robertson",
"Duncan",
""
],
[
"Zikic",
"Darko",
""
],
[
"Kontschieder",
"Peter",
""
],
[
"Shotton",
"Jamie",
""
],
[
"Brown",
"Matthew",
""
],
[
"Criminisi",
"Antonio",
""
]
] | TITLE: Decision Forests, Convolutional Networks and the Models in-Between
ABSTRACT: This paper investigates the connections between two state of the art
classifiers: decision forests (DFs, including decision jungles) and
convolutional neural networks (CNNs). Decision forests are computationally
efficient thanks to their conditional computation property (computation is
confined to only a small region of the tree, the nodes along a single branch).
CNNs achieve state of the art accuracy, thanks to their representation learning
capabilities. We present a systematic analysis of how to fuse conditional
computation with representation learning and achieve a continuum of hybrid
models with different ratios of accuracy vs. efficiency. We call this new
family of hybrid models conditional networks. Conditional networks can be
thought of as: i) decision trees augmented with data transformation operators,
or ii) CNNs, with block-diagonal sparse weight matrices, and explicit data
routing functions. Experimental validation is performed on the common task of
image classification on both the CIFAR and Imagenet datasets. Compared to state
of the art CNNs, our hybrid models yield the same accuracy with a fraction of
the compute cost and much smaller number of parameters.
| no_new_dataset | 0.952131 |
1506.05011 | Theofanis Karaletsos | Theofanis Karaletsos, Serge Belongie, Gunnar R\"atsch | Bayesian representation learning with oracle constraints | 16 pages, publishes in ICLR 16 | null | null | null | stat.ML cs.CV cs.LG | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | Representation learning systems typically rely on massive amounts of labeled
data in order to be trained to high accuracy. Recently, high-dimensional
parametric models like neural networks have succeeded in building rich
representations using either compressive, reconstructive or supervised
criteria. However, the semantic structure inherent in observations is
oftentimes lost in the process. Human perception excels at understanding
semantics but cannot always be expressed in terms of labels. Thus,
\emph{oracles} or \emph{human-in-the-loop systems}, for example crowdsourcing,
are often employed to generate similarity constraints using an implicit
similarity function encoded in human perception. In this work we propose to
combine \emph{generative unsupervised feature learning} with a
\emph{probabilistic treatment of oracle information like triplets} in order to
transfer implicit privileged oracle knowledge into explicit nonlinear Bayesian
latent factor models of the observations. We use a fast variational algorithm
to learn the joint model and demonstrate applicability to a well-known image
dataset. We show how implicit triplet information can provide rich information
to learn representations that outperform previous metric learning approaches as
well as generative models without this side-information in a variety of
predictive tasks. In addition, we illustrate that the proposed approach
compartmentalizes the latent spaces semantically which allows interpretation of
the latent variables.
| [
{
"version": "v1",
"created": "Tue, 16 Jun 2015 15:54:59 GMT"
},
{
"version": "v2",
"created": "Sat, 21 Nov 2015 05:24:01 GMT"
},
{
"version": "v3",
"created": "Fri, 8 Jan 2016 04:47:21 GMT"
},
{
"version": "v4",
"created": "Tue, 1 Mar 2016 23:36:04 GMT"
}
] | 2016-03-03T00:00:00 | [
[
"Karaletsos",
"Theofanis",
""
],
[
"Belongie",
"Serge",
""
],
[
"Rätsch",
"Gunnar",
""
]
] | TITLE: Bayesian representation learning with oracle constraints
ABSTRACT: Representation learning systems typically rely on massive amounts of labeled
data in order to be trained to high accuracy. Recently, high-dimensional
parametric models like neural networks have succeeded in building rich
representations using either compressive, reconstructive or supervised
criteria. However, the semantic structure inherent in observations is
oftentimes lost in the process. Human perception excels at understanding
semantics but cannot always be expressed in terms of labels. Thus,
\emph{oracles} or \emph{human-in-the-loop systems}, for example crowdsourcing,
are often employed to generate similarity constraints using an implicit
similarity function encoded in human perception. In this work we propose to
combine \emph{generative unsupervised feature learning} with a
\emph{probabilistic treatment of oracle information like triplets} in order to
transfer implicit privileged oracle knowledge into explicit nonlinear Bayesian
latent factor models of the observations. We use a fast variational algorithm
to learn the joint model and demonstrate applicability to a well-known image
dataset. We show how implicit triplet information can provide rich information
to learn representations that outperform previous metric learning approaches as
well as generative models without this side-information in a variety of
predictive tasks. In addition, we illustrate that the proposed approach
compartmentalizes the latent spaces semantically which allows interpretation of
the latent variables.
| no_new_dataset | 0.946646 |
1508.01722 | Jun-Cheng Chen | Jun-Cheng Chen and Vishal M. Patel and Rama Chellappa | Unconstrained Face Verification using Deep CNN Features | null | null | null | null | cs.CV | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | In this paper, we present an algorithm for unconstrained face verification
based on deep convolutional features and evaluate it on the newly released
IARPA Janus Benchmark A (IJB-A) dataset. The IJB-A dataset includes real-world
unconstrained faces from 500 subjects with full pose and illumination
variations which are much harder than the traditional Labeled Face in the Wild
(LFW) and Youtube Face (YTF) datasets. The deep convolutional neural network
(DCNN) is trained using the CASIA-WebFace dataset. Extensive experiments on the
IJB-A dataset are provided.
| [
{
"version": "v1",
"created": "Fri, 7 Aug 2015 15:21:19 GMT"
},
{
"version": "v2",
"created": "Wed, 2 Mar 2016 19:41:42 GMT"
}
] | 2016-03-03T00:00:00 | [
[
"Chen",
"Jun-Cheng",
""
],
[
"Patel",
"Vishal M.",
""
],
[
"Chellappa",
"Rama",
""
]
] | TITLE: Unconstrained Face Verification using Deep CNN Features
ABSTRACT: In this paper, we present an algorithm for unconstrained face verification
based on deep convolutional features and evaluate it on the newly released
IARPA Janus Benchmark A (IJB-A) dataset. The IJB-A dataset includes real-world
unconstrained faces from 500 subjects with full pose and illumination
variations which are much harder than the traditional Labeled Face in the Wild
(LFW) and Youtube Face (YTF) datasets. The deep convolutional neural network
(DCNN) is trained using the CASIA-WebFace dataset. Extensive experiments on the
IJB-A dataset are provided.
| new_dataset | 0.643329 |
1511.05939 | Oren Rippel | Oren Rippel, Manohar Paluri, Piotr Dollar, Lubomir Bourdev | Metric Learning with Adaptive Density Discrimination | ICLR 2016 | null | null | null | stat.ML cs.LG | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | Distance metric learning (DML) approaches learn a transformation to a
representation space where distance is in correspondence with a predefined
notion of similarity. While such models offer a number of compelling benefits,
it has been difficult for these to compete with modern classification
algorithms in performance and even in feature extraction.
In this work, we propose a novel approach explicitly designed to address a
number of subtle yet important issues which have stymied earlier DML
algorithms. It maintains an explicit model of the distributions of the
different classes in representation space. It then employs this knowledge to
adaptively assess similarity, and achieve local discrimination by penalizing
class distribution overlap.
We demonstrate the effectiveness of this idea on several tasks. Our approach
achieves state-of-the-art classification results on a number of fine-grained
visual recognition datasets, surpassing the standard softmax classifier and
outperforming triplet loss by a relative margin of 30-40%. In terms of
computational performance, it alleviates training inefficiencies in the
traditional triplet loss, reaching the same error in 5-30 times fewer
iterations. Beyond classification, we further validate the saliency of the
learnt representations via their attribute concentration and hierarchy recovery
properties, achieving 10-25% relative gains on the softmax classifier and
25-50% on triplet loss in these tasks.
| [
{
"version": "v1",
"created": "Wed, 18 Nov 2015 20:41:05 GMT"
},
{
"version": "v2",
"created": "Wed, 2 Mar 2016 04:52:08 GMT"
}
] | 2016-03-03T00:00:00 | [
[
"Rippel",
"Oren",
""
],
[
"Paluri",
"Manohar",
""
],
[
"Dollar",
"Piotr",
""
],
[
"Bourdev",
"Lubomir",
""
]
] | TITLE: Metric Learning with Adaptive Density Discrimination
ABSTRACT: Distance metric learning (DML) approaches learn a transformation to a
representation space where distance is in correspondence with a predefined
notion of similarity. While such models offer a number of compelling benefits,
it has been difficult for these to compete with modern classification
algorithms in performance and even in feature extraction.
In this work, we propose a novel approach explicitly designed to address a
number of subtle yet important issues which have stymied earlier DML
algorithms. It maintains an explicit model of the distributions of the
different classes in representation space. It then employs this knowledge to
adaptively assess similarity, and achieve local discrimination by penalizing
class distribution overlap.
We demonstrate the effectiveness of this idea on several tasks. Our approach
achieves state-of-the-art classification results on a number of fine-grained
visual recognition datasets, surpassing the standard softmax classifier and
outperforming triplet loss by a relative margin of 30-40%. In terms of
computational performance, it alleviates training inefficiencies in the
traditional triplet loss, reaching the same error in 5-30 times fewer
iterations. Beyond classification, we further validate the saliency of the
learnt representations via their attribute concentration and hierarchy recovery
properties, achieving 10-25% relative gains on the softmax classifier and
25-50% on triplet loss in these tasks.
| no_new_dataset | 0.941654 |
1511.06238 | Miriam Cha | Miriam Cha, Youngjune Gwon, H.T. Kung | Multimodal sparse representation learning and applications | null | null | null | null | cs.LG cs.CV stat.ML | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | Unsupervised methods have proven effective for discriminative tasks in a
single-modality scenario. In this paper, we present a multimodal framework for
learning sparse representations that can capture semantic correlation between
modalities. The framework can model relationships at a higher level by forcing
the shared sparse representation. In particular, we propose the use of joint
dictionary learning technique for sparse coding and formulate the joint
representation for concision, cross-modal representations (in case of a missing
modality), and union of the cross-modal representations. Given the accelerated
growth of multimodal data posted on the Web such as YouTube, Wikipedia, and
Twitter, learning good multimodal features is becoming increasingly important.
We show that the shared representations enabled by our framework substantially
improve the classification performance under both unimodal and multimodal
settings. We further show how deep architectures built on the proposed
framework are effective for the case of highly nonlinear correlations between
modalities. The effectiveness of our approach is demonstrated experimentally in
image denoising, multimedia event detection and retrieval on the TRECVID
dataset (audio-video), category classification on the Wikipedia dataset
(image-text), and sentiment classification on PhotoTweet (image-text).
| [
{
"version": "v1",
"created": "Thu, 19 Nov 2015 16:26:24 GMT"
},
{
"version": "v2",
"created": "Sun, 24 Jan 2016 23:18:09 GMT"
},
{
"version": "v3",
"created": "Wed, 2 Mar 2016 19:22:48 GMT"
}
] | 2016-03-03T00:00:00 | [
[
"Cha",
"Miriam",
""
],
[
"Gwon",
"Youngjune",
""
],
[
"Kung",
"H. T.",
""
]
] | TITLE: Multimodal sparse representation learning and applications
ABSTRACT: Unsupervised methods have proven effective for discriminative tasks in a
single-modality scenario. In this paper, we present a multimodal framework for
learning sparse representations that can capture semantic correlation between
modalities. The framework can model relationships at a higher level by forcing
the shared sparse representation. In particular, we propose the use of joint
dictionary learning technique for sparse coding and formulate the joint
representation for concision, cross-modal representations (in case of a missing
modality), and union of the cross-modal representations. Given the accelerated
growth of multimodal data posted on the Web such as YouTube, Wikipedia, and
Twitter, learning good multimodal features is becoming increasingly important.
We show that the shared representations enabled by our framework substantially
improve the classification performance under both unimodal and multimodal
settings. We further show how deep architectures built on the proposed
framework are effective for the case of highly nonlinear correlations between
modalities. The effectiveness of our approach is demonstrated experimentally in
image denoising, multimedia event detection and retrieval on the TRECVID
dataset (audio-video), category classification on the Wikipedia dataset
(image-text), and sentiment classification on PhotoTweet (image-text).
| no_new_dataset | 0.946001 |
1511.06385 | Chunchuan Lv Mr. | Chunchuan Lyu, Kaizhu Huang, Hai-Ning Liang | A Unified Gradient Regularization Family for Adversarial Examples | The paper has been presented at ICDM 2015 | null | null | null | cs.LG stat.ML | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | Adversarial examples are augmented data points generated by imperceptible
perturbation of input samples. They have recently drawn much attention with the
machine learning and data mining community. Being difficult to distinguish from
real examples, such adversarial examples could change the prediction of many of
the best learning models including the state-of-the-art deep learning models.
Recent attempts have been made to build robust models that take into account
adversarial examples. However, these methods can either lead to performance
drops or lack mathematical motivations. In this paper, we propose a unified
framework to build robust machine learning models against adversarial examples.
More specifically, using the unified framework, we develop a family of gradient
regularization methods that effectively penalize the gradient of loss function
w.r.t. inputs. Our proposed framework is appealing in that it offers a unified
view to deal with adversarial examples. It incorporates another
recently-proposed perturbation based approach as a special case. In addition,
we present some visual effects that reveals semantic meaning in those
perturbations, and thus support our regularization method and provide another
explanation for generalizability of adversarial examples. By applying this
technique to Maxout networks, we conduct a series of experiments and achieve
encouraging results on two benchmark datasets. In particular,we attain the best
accuracy on MNIST data (without data augmentation) and competitive performance
on CIFAR-10 data.
| [
{
"version": "v1",
"created": "Thu, 19 Nov 2015 21:14:43 GMT"
}
] | 2016-03-03T00:00:00 | [
[
"Lyu",
"Chunchuan",
""
],
[
"Huang",
"Kaizhu",
""
],
[
"Liang",
"Hai-Ning",
""
]
] | TITLE: A Unified Gradient Regularization Family for Adversarial Examples
ABSTRACT: Adversarial examples are augmented data points generated by imperceptible
perturbation of input samples. They have recently drawn much attention with the
machine learning and data mining community. Being difficult to distinguish from
real examples, such adversarial examples could change the prediction of many of
the best learning models including the state-of-the-art deep learning models.
Recent attempts have been made to build robust models that take into account
adversarial examples. However, these methods can either lead to performance
drops or lack mathematical motivations. In this paper, we propose a unified
framework to build robust machine learning models against adversarial examples.
More specifically, using the unified framework, we develop a family of gradient
regularization methods that effectively penalize the gradient of loss function
w.r.t. inputs. Our proposed framework is appealing in that it offers a unified
view to deal with adversarial examples. It incorporates another
recently-proposed perturbation based approach as a special case. In addition,
we present some visual effects that reveals semantic meaning in those
perturbations, and thus support our regularization method and provide another
explanation for generalizability of adversarial examples. By applying this
technique to Maxout networks, we conduct a series of experiments and achieve
encouraging results on two benchmark datasets. In particular,we attain the best
accuracy on MNIST data (without data augmentation) and competitive performance
on CIFAR-10 data.
| no_new_dataset | 0.944995 |
1603.00788 | Alp Kucukelbir | Alp Kucukelbir, Dustin Tran, Rajesh Ranganath, Andrew Gelman, David M.
Blei | Automatic Differentiation Variational Inference | null | null | null | null | stat.ML cs.AI cs.LG stat.CO | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | Probabilistic modeling is iterative. A scientist posits a simple model, fits
it to her data, refines it according to her analysis, and repeats. However,
fitting complex models to large data is a bottleneck in this process. Deriving
algorithms for new models can be both mathematically and computationally
challenging, which makes it difficult to efficiently cycle through the steps.
To this end, we develop automatic differentiation variational inference (ADVI).
Using our method, the scientist only provides a probabilistic model and a
dataset, nothing else. ADVI automatically derives an efficient variational
inference algorithm, freeing the scientist to refine and explore many models.
ADVI supports a broad class of models-no conjugacy assumptions are required. We
study ADVI across ten different models and apply it to a dataset with millions
of observations. ADVI is integrated into Stan, a probabilistic programming
system; it is available for immediate use.
| [
{
"version": "v1",
"created": "Wed, 2 Mar 2016 16:43:15 GMT"
}
] | 2016-03-03T00:00:00 | [
[
"Kucukelbir",
"Alp",
""
],
[
"Tran",
"Dustin",
""
],
[
"Ranganath",
"Rajesh",
""
],
[
"Gelman",
"Andrew",
""
],
[
"Blei",
"David M.",
""
]
] | TITLE: Automatic Differentiation Variational Inference
ABSTRACT: Probabilistic modeling is iterative. A scientist posits a simple model, fits
it to her data, refines it according to her analysis, and repeats. However,
fitting complex models to large data is a bottleneck in this process. Deriving
algorithms for new models can be both mathematically and computationally
challenging, which makes it difficult to efficiently cycle through the steps.
To this end, we develop automatic differentiation variational inference (ADVI).
Using our method, the scientist only provides a probabilistic model and a
dataset, nothing else. ADVI automatically derives an efficient variational
inference algorithm, freeing the scientist to refine and explore many models.
ADVI supports a broad class of models-no conjugacy assumptions are required. We
study ADVI across ten different models and apply it to a dataset with millions
of observations. ADVI is integrated into Stan, a probabilistic programming
system; it is available for immediate use.
| no_new_dataset | 0.948489 |
1603.00845 | Xavier Gir\'o-i-Nieto | Junting Pan, Kevin McGuinness, Elisa Sayrol, Noel O'Connor and Xavier
Giro-i-Nieto | Shallow and Deep Convolutional Networks for Saliency Prediction | Preprint of the paper accepted at 2016 IEEE Conference on Computer
Vision and Pattern Recognition (CVPR). Source code and models available at
https://github.com/imatge-upc/saliency-2016-cvpr. Junting Pan and Kevin
McGuinness contributed equally to this work | null | null | null | cs.CV cs.LG | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | The prediction of salient areas in images has been traditionally addressed
with hand-crafted features based on neuroscience principles. This paper,
however, addresses the problem with a completely data-driven approach by
training a convolutional neural network (convnet). The learning process is
formulated as a minimization of a loss function that measures the Euclidean
distance of the predicted saliency map with the provided ground truth. The
recent publication of large datasets of saliency prediction has provided enough
data to train end-to-end architectures that are both fast and accurate. Two
designs are proposed: a shallow convnet trained from scratch, and a another
deeper solution whose first three layers are adapted from another network
trained for classification. To the authors knowledge, these are the first
end-to-end CNNs trained and tested for the purpose of saliency prediction.
| [
{
"version": "v1",
"created": "Wed, 2 Mar 2016 19:54:02 GMT"
}
] | 2016-03-03T00:00:00 | [
[
"Pan",
"Junting",
""
],
[
"McGuinness",
"Kevin",
""
],
[
"Sayrol",
"Elisa",
""
],
[
"O'Connor",
"Noel",
""
],
[
"Giro-i-Nieto",
"Xavier",
""
]
] | TITLE: Shallow and Deep Convolutional Networks for Saliency Prediction
ABSTRACT: The prediction of salient areas in images has been traditionally addressed
with hand-crafted features based on neuroscience principles. This paper,
however, addresses the problem with a completely data-driven approach by
training a convolutional neural network (convnet). The learning process is
formulated as a minimization of a loss function that measures the Euclidean
distance of the predicted saliency map with the provided ground truth. The
recent publication of large datasets of saliency prediction has provided enough
data to train end-to-end architectures that are both fast and accurate. Two
designs are proposed: a shallow convnet trained from scratch, and a another
deeper solution whose first three layers are adapted from another network
trained for classification. To the authors knowledge, these are the first
end-to-end CNNs trained and tested for the purpose of saliency prediction.
| no_new_dataset | 0.950595 |
1503.00848 | Jordi Pont-Tuset | Jordi Pont-Tuset, Pablo Arbelaez, Jonathan T. Barron, Ferran Marques,
Jitendra Malik | Multiscale Combinatorial Grouping for Image Segmentation and Object
Proposal Generation | null | null | 10.1109/TPAMI.2016.2537320 | null | cs.CV | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | We propose a unified approach for bottom-up hierarchical image segmentation
and object proposal generation for recognition, called Multiscale Combinatorial
Grouping (MCG). For this purpose, we first develop a fast normalized cuts
algorithm. We then propose a high-performance hierarchical segmenter that makes
effective use of multiscale information. Finally, we propose a grouping
strategy that combines our multiscale regions into highly-accurate object
proposals by exploring efficiently their combinatorial space. We also present
Single-scale Combinatorial Grouping (SCG), a faster version of MCG that
produces competitive proposals in under five second per image. We conduct an
extensive and comprehensive empirical validation on the BSDS500, SegVOC12, SBD,
and COCO datasets, showing that MCG produces state-of-the-art contours,
hierarchical regions, and object proposals.
| [
{
"version": "v1",
"created": "Tue, 3 Mar 2015 07:58:22 GMT"
},
{
"version": "v2",
"created": "Tue, 17 Mar 2015 12:00:33 GMT"
},
{
"version": "v3",
"created": "Mon, 29 Jun 2015 17:57:06 GMT"
},
{
"version": "v4",
"created": "Tue, 1 Mar 2016 09:00:09 GMT"
}
] | 2016-03-02T00:00:00 | [
[
"Pont-Tuset",
"Jordi",
""
],
[
"Arbelaez",
"Pablo",
""
],
[
"Barron",
"Jonathan T.",
""
],
[
"Marques",
"Ferran",
""
],
[
"Malik",
"Jitendra",
""
]
] | TITLE: Multiscale Combinatorial Grouping for Image Segmentation and Object
Proposal Generation
ABSTRACT: We propose a unified approach for bottom-up hierarchical image segmentation
and object proposal generation for recognition, called Multiscale Combinatorial
Grouping (MCG). For this purpose, we first develop a fast normalized cuts
algorithm. We then propose a high-performance hierarchical segmenter that makes
effective use of multiscale information. Finally, we propose a grouping
strategy that combines our multiscale regions into highly-accurate object
proposals by exploring efficiently their combinatorial space. We also present
Single-scale Combinatorial Grouping (SCG), a faster version of MCG that
produces competitive proposals in under five second per image. We conduct an
extensive and comprehensive empirical validation on the BSDS500, SegVOC12, SBD,
and COCO datasets, showing that MCG produces state-of-the-art contours,
hierarchical regions, and object proposals.
| no_new_dataset | 0.955775 |
1507.04760 | Lex Fridman | Lex Fridman, Philipp Langhans, Joonbum Lee, Bryan Reimer | Driver Gaze Region Estimation Without Using Eye Movement | Accepted for Publication in IEEE Intelligent Systems | null | null | null | cs.CV | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | Automated estimation of the allocation of a driver's visual attention may be
a critical component of future Advanced Driver Assistance Systems. In theory,
vision-based tracking of the eye can provide a good estimate of gaze location.
In practice, eye tracking from video is challenging because of sunglasses,
eyeglass reflections, lighting conditions, occlusions, motion blur, and other
factors. Estimation of head pose, on the other hand, is robust to many of these
effects, but cannot provide as fine-grained of a resolution in localizing the
gaze. However, for the purpose of keeping the driver safe, it is sufficient to
partition gaze into regions. In this effort, we propose a system that extracts
facial features and classifies their spatial configuration into six regions in
real-time. Our proposed method achieves an average accuracy of 91.4% at an
average decision rate of 11 Hz on a dataset of 50 drivers from an on-road
study.
| [
{
"version": "v1",
"created": "Thu, 16 Jul 2015 20:16:20 GMT"
},
{
"version": "v2",
"created": "Tue, 1 Mar 2016 17:21:25 GMT"
}
] | 2016-03-02T00:00:00 | [
[
"Fridman",
"Lex",
""
],
[
"Langhans",
"Philipp",
""
],
[
"Lee",
"Joonbum",
""
],
[
"Reimer",
"Bryan",
""
]
] | TITLE: Driver Gaze Region Estimation Without Using Eye Movement
ABSTRACT: Automated estimation of the allocation of a driver's visual attention may be
a critical component of future Advanced Driver Assistance Systems. In theory,
vision-based tracking of the eye can provide a good estimate of gaze location.
In practice, eye tracking from video is challenging because of sunglasses,
eyeglass reflections, lighting conditions, occlusions, motion blur, and other
factors. Estimation of head pose, on the other hand, is robust to many of these
effects, but cannot provide as fine-grained of a resolution in localizing the
gaze. However, for the purpose of keeping the driver safe, it is sufficient to
partition gaze into regions. In this effort, we propose a system that extracts
facial features and classifies their spatial configuration into six regions in
real-time. Our proposed method achieves an average accuracy of 91.4% at an
average decision rate of 11 Hz on a dataset of 50 drivers from an on-road
study.
| no_new_dataset | 0.940134 |
1509.06664 | Tim Rockt\"aschel | Tim Rockt\"aschel, Edward Grefenstette, Karl Moritz Hermann,
Tom\'a\v{s} Ko\v{c}isk\'y, Phil Blunsom | Reasoning about Entailment with Neural Attention | ICLR 2016 camera-ready, 9 pages, 10 figures (incl. subfigures) | null | null | null | cs.CL cs.AI cs.LG cs.NE | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | While most approaches to automatically recognizing entailment relations have
used classifiers employing hand engineered features derived from complex
natural language processing pipelines, in practice their performance has been
only slightly better than bag-of-word pair classifiers using only lexical
similarity. The only attempt so far to build an end-to-end differentiable
neural network for entailment failed to outperform such a simple similarity
classifier. In this paper, we propose a neural model that reads two sentences
to determine entailment using long short-term memory units. We extend this
model with a word-by-word neural attention mechanism that encourages reasoning
over entailments of pairs of words and phrases. Furthermore, we present a
qualitative analysis of attention weights produced by this model, demonstrating
such reasoning capabilities. On a large entailment dataset this model
outperforms the previous best neural model and a classifier with engineered
features by a substantial margin. It is the first generic end-to-end
differentiable system that achieves state-of-the-art accuracy on a textual
entailment dataset.
| [
{
"version": "v1",
"created": "Tue, 22 Sep 2015 16:08:24 GMT"
},
{
"version": "v2",
"created": "Tue, 10 Nov 2015 22:12:52 GMT"
},
{
"version": "v3",
"created": "Mon, 18 Jan 2016 17:28:30 GMT"
},
{
"version": "v4",
"created": "Tue, 1 Mar 2016 10:32:06 GMT"
}
] | 2016-03-02T00:00:00 | [
[
"Rocktäschel",
"Tim",
""
],
[
"Grefenstette",
"Edward",
""
],
[
"Hermann",
"Karl Moritz",
""
],
[
"Kočiský",
"Tomáš",
""
],
[
"Blunsom",
"Phil",
""
]
] | TITLE: Reasoning about Entailment with Neural Attention
ABSTRACT: While most approaches to automatically recognizing entailment relations have
used classifiers employing hand engineered features derived from complex
natural language processing pipelines, in practice their performance has been
only slightly better than bag-of-word pair classifiers using only lexical
similarity. The only attempt so far to build an end-to-end differentiable
neural network for entailment failed to outperform such a simple similarity
classifier. In this paper, we propose a neural model that reads two sentences
to determine entailment using long short-term memory units. We extend this
model with a word-by-word neural attention mechanism that encourages reasoning
over entailments of pairs of words and phrases. Furthermore, we present a
qualitative analysis of attention weights produced by this model, demonstrating
such reasoning capabilities. On a large entailment dataset this model
outperforms the previous best neural model and a classifier with engineered
features by a substantial margin. It is the first generic end-to-end
differentiable system that achieves state-of-the-art accuracy on a textual
entailment dataset.
| no_new_dataset | 0.946349 |
1511.06432 | Nicolas Ballas | Nicolas Ballas, Li Yao, Chris Pal, Aaron Courville | Delving Deeper into Convolutional Networks for Learning Video
Representations | ICLR 2016 | null | null | null | cs.CV cs.LG cs.NE | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | We propose an approach to learn spatio-temporal features in videos from
intermediate visual representations we call "percepts" using
Gated-Recurrent-Unit Recurrent Networks (GRUs).Our method relies on percepts
that are extracted from all level of a deep convolutional network trained on
the large ImageNet dataset. While high-level percepts contain highly
discriminative information, they tend to have a low-spatial resolution.
Low-level percepts, on the other hand, preserve a higher spatial resolution
from which we can model finer motion patterns. Using low-level percepts can
leads to high-dimensionality video representations. To mitigate this effect and
control the model number of parameters, we introduce a variant of the GRU model
that leverages the convolution operations to enforce sparse connectivity of the
model units and share parameters across the input spatial locations.
We empirically validate our approach on both Human Action Recognition and
Video Captioning tasks. In particular, we achieve results equivalent to
state-of-art on the YouTube2Text dataset using a simpler text-decoder model and
without extra 3D CNN features.
| [
{
"version": "v1",
"created": "Thu, 19 Nov 2015 22:46:13 GMT"
},
{
"version": "v2",
"created": "Mon, 23 Nov 2015 02:46:54 GMT"
},
{
"version": "v3",
"created": "Thu, 7 Jan 2016 19:43:19 GMT"
},
{
"version": "v4",
"created": "Tue, 1 Mar 2016 18:54:11 GMT"
}
] | 2016-03-02T00:00:00 | [
[
"Ballas",
"Nicolas",
""
],
[
"Yao",
"Li",
""
],
[
"Pal",
"Chris",
""
],
[
"Courville",
"Aaron",
""
]
] | TITLE: Delving Deeper into Convolutional Networks for Learning Video
Representations
ABSTRACT: We propose an approach to learn spatio-temporal features in videos from
intermediate visual representations we call "percepts" using
Gated-Recurrent-Unit Recurrent Networks (GRUs).Our method relies on percepts
that are extracted from all level of a deep convolutional network trained on
the large ImageNet dataset. While high-level percepts contain highly
discriminative information, they tend to have a low-spatial resolution.
Low-level percepts, on the other hand, preserve a higher spatial resolution
from which we can model finer motion patterns. Using low-level percepts can
leads to high-dimensionality video representations. To mitigate this effect and
control the model number of parameters, we introduce a variant of the GRU model
that leverages the convolution operations to enforce sparse connectivity of the
model units and share parameters across the input spatial locations.
We empirically validate our approach on both Human Action Recognition and
Video Captioning tasks. In particular, we achieve results equivalent to
state-of-art on the YouTube2Text dataset using a simpler text-decoder model and
without extra 3D CNN features.
| no_new_dataset | 0.9462 |
1511.06455 | Zhenwen Dai | Zhenwen Dai, Andreas Damianou, Javier Gonz\'alez, Neil Lawrence | Variational Auto-encoded Deep Gaussian Processes | null | null | null | null | cs.LG stat.ML | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | We develop a scalable deep non-parametric generative model by augmenting deep
Gaussian processes with a recognition model. Inference is performed in a novel
scalable variational framework where the variational posterior distributions
are reparametrized through a multilayer perceptron. The key aspect of this
reformulation is that it prevents the proliferation of variational parameters
which otherwise grow linearly in proportion to the sample size. We derive a new
formulation of the variational lower bound that allows us to distribute most of
the computation in a way that enables to handle datasets of the size of
mainstream deep learning tasks. We show the efficacy of the method on a variety
of challenges including deep unsupervised learning and deep Bayesian
optimization.
| [
{
"version": "v1",
"created": "Thu, 19 Nov 2015 23:47:34 GMT"
},
{
"version": "v2",
"created": "Mon, 29 Feb 2016 21:34:58 GMT"
}
] | 2016-03-02T00:00:00 | [
[
"Dai",
"Zhenwen",
""
],
[
"Damianou",
"Andreas",
""
],
[
"González",
"Javier",
""
],
[
"Lawrence",
"Neil",
""
]
] | TITLE: Variational Auto-encoded Deep Gaussian Processes
ABSTRACT: We develop a scalable deep non-parametric generative model by augmenting deep
Gaussian processes with a recognition model. Inference is performed in a novel
scalable variational framework where the variational posterior distributions
are reparametrized through a multilayer perceptron. The key aspect of this
reformulation is that it prevents the proliferation of variational parameters
which otherwise grow linearly in proportion to the sample size. We derive a new
formulation of the variational lower bound that allows us to distribute most of
the computation in a way that enables to handle datasets of the size of
mainstream deep learning tasks. We show the efficacy of the method on a variety
of challenges including deep unsupervised learning and deep Bayesian
optimization.
| no_new_dataset | 0.944638 |
1603.00016 | Gerald Eigen | Gerald Eigen, Are Tr{\ae}et, Justas Zalieckas, Jaroslav Cvach, Jiri
Kvasnicka, Ivo Polak | SiPM Gain Stabilization Studies for Adaptive Power Supply | 14 pages, 41 figures, Talk presented at the International Workshop on
Future Linear Colliders (LCWS15), Whistler, Canada, 2-6 November 2015 | null | null | null | physics.ins-det hep-ex | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | We present herein gain stabilization studies of SiPMs using a climate chamber
at CERN. We present results for four detectors not tested before, three from
Hamamatsu and one from KETEK. Two of the Hamamatsu SiPMs are novel sensors with
trenches that reduce cross talk. We use an improved readout system with a
digital oscilloscope controlled with a dedicated LabView program. We improved
and automized the analysis to deal with large datasets. We have measured the
gain-versus-bias-voltage dependence at fixed temperature and
gain-versus-temperature dependence at fixed bias voltage to determine the bias
voltage dependence on temperature $V(T)$ for stable gain. We show that the gain
remains stable to better than $\pm 0.5\%$ in the $20^\circ \rm C - 30^\circ C$
temperature range if the bias voltage is properly adjusted with temperature.
| [
{
"version": "v1",
"created": "Mon, 29 Feb 2016 19:44:24 GMT"
}
] | 2016-03-02T00:00:00 | [
[
"Eigen",
"Gerald",
""
],
[
"Træet",
"Are",
""
],
[
"Zalieckas",
"Justas",
""
],
[
"Cvach",
"Jaroslav",
""
],
[
"Kvasnicka",
"Jiri",
""
],
[
"Polak",
"Ivo",
""
]
] | TITLE: SiPM Gain Stabilization Studies for Adaptive Power Supply
ABSTRACT: We present herein gain stabilization studies of SiPMs using a climate chamber
at CERN. We present results for four detectors not tested before, three from
Hamamatsu and one from KETEK. Two of the Hamamatsu SiPMs are novel sensors with
trenches that reduce cross talk. We use an improved readout system with a
digital oscilloscope controlled with a dedicated LabView program. We improved
and automized the analysis to deal with large datasets. We have measured the
gain-versus-bias-voltage dependence at fixed temperature and
gain-versus-temperature dependence at fixed bias voltage to determine the bias
voltage dependence on temperature $V(T)$ for stable gain. We show that the gain
remains stable to better than $\pm 0.5\%$ in the $20^\circ \rm C - 30^\circ C$
temperature range if the bias voltage is properly adjusted with temperature.
| no_new_dataset | 0.938181 |
1603.00128 | Yanwei Pang | Xiaoheng Jiang, Yanwei Pang, Manli Sun, and Xuelong Li | Cascaded Subpatch Networks for Effective CNNs | null | null | null | null | cs.CV | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | Conventional Convolutional Neural Networks (CNNs) use either a linear or
non-linear filter to extract features from an image patch (region) of spatial
size $ H\times W $ (Typically, $ H $ is small and is equal to $ W$, e.g., $ H $
is 5 or 7). Generally, the size of the filter is equal to the size $ H\times W
$ of the input patch. We argue that the representation ability of equal-size
strategy is not strong enough. To overcome the drawback, we propose to use
subpatch filter whose spatial size $ h\times w $ is smaller than $ H\times W $.
The proposed subpatch filter consists of two subsequent filters. The first one
is a linear filter of spatial size $ h\times w $ and is aimed at extracting
features from spatial domain. The second one is of spatial size $ 1\times 1 $
and is used for strengthening the connection between different input feature
channels and for reducing the number of parameters. The subpatch filter
convolves with the input patch and the resulting network is called a subpatch
network. Taking the output of one subpatch network as input, we further repeat
constructing subpatch networks until the output contains only one neuron in
spatial domain. These subpatch networks form a new network called Cascaded
Subpatch Network (CSNet). The feature layer generated by CSNet is called csconv
layer. For the whole input image, we construct a deep neural network by
stacking a sequence of csconv layers. Experimental results on four benchmark
datasets demonstrate the effectiveness and compactness of the proposed CSNet.
For example, our CSNet reaches a test error of $ 5.68\% $ on the CIFAR10
dataset without model averaging. To the best of our knowledge, this is the best
result ever obtained on the CIFAR10 dataset.
| [
{
"version": "v1",
"created": "Tue, 1 Mar 2016 03:44:49 GMT"
}
] | 2016-03-02T00:00:00 | [
[
"Jiang",
"Xiaoheng",
""
],
[
"Pang",
"Yanwei",
""
],
[
"Sun",
"Manli",
""
],
[
"Li",
"Xuelong",
""
]
] | TITLE: Cascaded Subpatch Networks for Effective CNNs
ABSTRACT: Conventional Convolutional Neural Networks (CNNs) use either a linear or
non-linear filter to extract features from an image patch (region) of spatial
size $ H\times W $ (Typically, $ H $ is small and is equal to $ W$, e.g., $ H $
is 5 or 7). Generally, the size of the filter is equal to the size $ H\times W
$ of the input patch. We argue that the representation ability of equal-size
strategy is not strong enough. To overcome the drawback, we propose to use
subpatch filter whose spatial size $ h\times w $ is smaller than $ H\times W $.
The proposed subpatch filter consists of two subsequent filters. The first one
is a linear filter of spatial size $ h\times w $ and is aimed at extracting
features from spatial domain. The second one is of spatial size $ 1\times 1 $
and is used for strengthening the connection between different input feature
channels and for reducing the number of parameters. The subpatch filter
convolves with the input patch and the resulting network is called a subpatch
network. Taking the output of one subpatch network as input, we further repeat
constructing subpatch networks until the output contains only one neuron in
spatial domain. These subpatch networks form a new network called Cascaded
Subpatch Network (CSNet). The feature layer generated by CSNet is called csconv
layer. For the whole input image, we construct a deep neural network by
stacking a sequence of csconv layers. Experimental results on four benchmark
datasets demonstrate the effectiveness and compactness of the proposed CSNet.
For example, our CSNet reaches a test error of $ 5.68\% $ on the CIFAR10
dataset without model averaging. To the best of our knowledge, this is the best
result ever obtained on the CIFAR10 dataset.
| no_new_dataset | 0.951953 |
1603.00150 | Dylan Campbell | Dylan Campbell and Lars Petersson | GOGMA: Globally-Optimal Gaussian Mixture Alignment | Manuscript in press 2016 IEEE Conference on Computer Vision and
Pattern Recognition | null | null | null | cs.CV cs.RO | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | Gaussian mixture alignment is a family of approaches that are frequently used
for robustly solving the point-set registration problem. However, since they
use local optimisation, they are susceptible to local minima and can only
guarantee local optimality. Consequently, their accuracy is strongly dependent
on the quality of the initialisation. This paper presents the first
globally-optimal solution to the 3D rigid Gaussian mixture alignment problem
under the L2 distance between mixtures. The algorithm, named GOGMA, employs a
branch-and-bound approach to search the space of 3D rigid motions SE(3),
guaranteeing global optimality regardless of the initialisation. The geometry
of SE(3) was used to find novel upper and lower bounds for the objective
function and local optimisation was integrated into the scheme to accelerate
convergence without voiding the optimality guarantee. The evaluation
empirically supported the optimality proof and showed that the method performed
much more robustly on two challenging datasets than an existing
globally-optimal registration solution.
| [
{
"version": "v1",
"created": "Tue, 1 Mar 2016 05:38:50 GMT"
}
] | 2016-03-02T00:00:00 | [
[
"Campbell",
"Dylan",
""
],
[
"Petersson",
"Lars",
""
]
] | TITLE: GOGMA: Globally-Optimal Gaussian Mixture Alignment
ABSTRACT: Gaussian mixture alignment is a family of approaches that are frequently used
for robustly solving the point-set registration problem. However, since they
use local optimisation, they are susceptible to local minima and can only
guarantee local optimality. Consequently, their accuracy is strongly dependent
on the quality of the initialisation. This paper presents the first
globally-optimal solution to the 3D rigid Gaussian mixture alignment problem
under the L2 distance between mixtures. The algorithm, named GOGMA, employs a
branch-and-bound approach to search the space of 3D rigid motions SE(3),
guaranteeing global optimality regardless of the initialisation. The geometry
of SE(3) was used to find novel upper and lower bounds for the objective
function and local optimisation was integrated into the scheme to accelerate
convergence without voiding the optimality guarantee. The evaluation
empirically supported the optimality proof and showed that the method performed
much more robustly on two challenging datasets than an existing
globally-optimal registration solution.
| no_new_dataset | 0.950319 |
1603.00431 | Amirali Sanatinia | Amirali Sanatinia, Guevara Noubir | On GitHub's Programming Languages | null | null | null | null | cs.PL | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | GitHub is the most widely used social, distributed version control system. It
has around 10 million registered users and hosts over 16 million public
repositories. Its user base is also very active as GitHub ranks in the top 100
Alexa most popular websites. In this study, we collect GitHub's state in its
entirety. Doing so, allows us to study new aspects of the ecosystem. Although
GitHub is the home to millions of users and repositories, the analysis of
users' activity time-series reveals that only around 10% of them can be
considered active. The collected dataset allows us to investigate the
popularity of programming languages and existence of pattens in the relations
between users, repositories, and programming languages.
By, applying a k-means clustering method to the users-repositories commits
matrix, we find that two clear clusters of programming languages separate from
the remaining. One cluster forms for "web programming" languages (Java Script,
Ruby, PHP, CSS), and a second for "system oriented programming" languages (C,
C++, Python). Further classification, allow us to build a phylogenetic tree of
the use of programming languages in GitHub. Additionally, we study the main and
the auxiliary programming languages of the top 1000 repositories in more
detail. We provide a ranking of these auxiliary programming languages using
various metrics, such as percentage of lines of code, and PageRank.
| [
{
"version": "v1",
"created": "Tue, 1 Mar 2016 20:03:44 GMT"
}
] | 2016-03-02T00:00:00 | [
[
"Sanatinia",
"Amirali",
""
],
[
"Noubir",
"Guevara",
""
]
] | TITLE: On GitHub's Programming Languages
ABSTRACT: GitHub is the most widely used social, distributed version control system. It
has around 10 million registered users and hosts over 16 million public
repositories. Its user base is also very active as GitHub ranks in the top 100
Alexa most popular websites. In this study, we collect GitHub's state in its
entirety. Doing so, allows us to study new aspects of the ecosystem. Although
GitHub is the home to millions of users and repositories, the analysis of
users' activity time-series reveals that only around 10% of them can be
considered active. The collected dataset allows us to investigate the
popularity of programming languages and existence of pattens in the relations
between users, repositories, and programming languages.
By, applying a k-means clustering method to the users-repositories commits
matrix, we find that two clear clusters of programming languages separate from
the remaining. One cluster forms for "web programming" languages (Java Script,
Ruby, PHP, CSS), and a second for "system oriented programming" languages (C,
C++, Python). Further classification, allow us to build a phylogenetic tree of
the use of programming languages in GitHub. Additionally, we study the main and
the auxiliary programming languages of the top 1000 repositories in more
detail. We provide a ranking of these auxiliary programming languages using
various metrics, such as percentage of lines of code, and PageRank.
| new_dataset | 0.92912 |
1603.00438 | Team Lear | Mattis Paulin (LEAR), Julien Mairal (LEAR), Matthijs Douze (LEAR),
Zaid Harchaoui (NYU), Florent Perronnin, Cordelia Schmid (LEAR) | Convolutional Patch Representations for Image Retrieval: an Unsupervised
Approach | null | null | null | null | cs.CV | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | Convolutional neural networks (CNNs) have recently received a lot of
attention due to their ability to model local stationary structures in natural
images in a multi-scale fashion, when learning all model parameters with
supervision. While excellent performance was achieved for image classification
when large amounts of labeled visual data are available, their success for
un-supervised tasks such as image retrieval has been moderate so far. Our paper
focuses on this latter setting and explores several methods for learning patch
descriptors without supervision with application to matching and instance-level
retrieval. To that effect, we propose a new family of convolutional descriptors
for patch representation , based on the recently introduced convolutional
kernel networks. We show that our descriptor, named Patch-CKN, performs better
than SIFT as well as other convolutional networks learned by artificially
introducing supervision and is significantly faster to train. To demonstrate
its effectiveness, we perform an extensive evaluation on standard benchmarks
for patch and image retrieval where we obtain state-of-the-art results. We also
introduce a new dataset called RomePatches, which allows to simultaneously
study descriptor performance for patch and image retrieval.
| [
{
"version": "v1",
"created": "Tue, 1 Mar 2016 20:13:07 GMT"
}
] | 2016-03-02T00:00:00 | [
[
"Paulin",
"Mattis",
"",
"LEAR"
],
[
"Mairal",
"Julien",
"",
"LEAR"
],
[
"Douze",
"Matthijs",
"",
"LEAR"
],
[
"Harchaoui",
"Zaid",
"",
"NYU"
],
[
"Perronnin",
"Florent",
"",
"LEAR"
],
[
"Schmid",
"Cordelia",
"",
"LEAR"
]
] | TITLE: Convolutional Patch Representations for Image Retrieval: an Unsupervised
Approach
ABSTRACT: Convolutional neural networks (CNNs) have recently received a lot of
attention due to their ability to model local stationary structures in natural
images in a multi-scale fashion, when learning all model parameters with
supervision. While excellent performance was achieved for image classification
when large amounts of labeled visual data are available, their success for
un-supervised tasks such as image retrieval has been moderate so far. Our paper
focuses on this latter setting and explores several methods for learning patch
descriptors without supervision with application to matching and instance-level
retrieval. To that effect, we propose a new family of convolutional descriptors
for patch representation , based on the recently introduced convolutional
kernel networks. We show that our descriptor, named Patch-CKN, performs better
than SIFT as well as other convolutional networks learned by artificially
introducing supervision and is significantly faster to train. To demonstrate
its effectiveness, we perform an extensive evaluation on standard benchmarks
for patch and image retrieval where we obtain state-of-the-art results. We also
introduce a new dataset called RomePatches, which allows to simultaneously
study descriptor performance for patch and image retrieval.
| new_dataset | 0.954393 |
1407.3422 | Igor Melnyk | Igor Melnyk and Arindam Banerjee | A Spectral Algorithm for Inference in Hidden Semi-Markov Models | null | null | null | null | stat.ML cs.LG | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | Hidden semi-Markov models (HSMMs) are latent variable models which allow
latent state persistence and can be viewed as a generalization of the popular
hidden Markov models (HMMs). In this paper, we introduce a novel spectral
algorithm to perform inference in HSMMs. Unlike expectation maximization (EM),
our approach correctly estimates the probability of given observation sequence
based on a set of training sequences. Our approach is based on estimating
moments from the sample, whose number of dimensions depends only
logarithmically on the maximum length of the hidden state persistence.
Moreover, the algorithm requires only a few matrix inversions and is therefore
computationally efficient. Empirical evaluations on synthetic and real data
demonstrate the advantage of the algorithm over EM in terms of speed and
accuracy, especially for large datasets.
| [
{
"version": "v1",
"created": "Sat, 12 Jul 2014 23:57:07 GMT"
},
{
"version": "v2",
"created": "Sun, 21 Feb 2016 19:29:03 GMT"
},
{
"version": "v3",
"created": "Mon, 29 Feb 2016 00:29:23 GMT"
}
] | 2016-03-01T00:00:00 | [
[
"Melnyk",
"Igor",
""
],
[
"Banerjee",
"Arindam",
""
]
] | TITLE: A Spectral Algorithm for Inference in Hidden Semi-Markov Models
ABSTRACT: Hidden semi-Markov models (HSMMs) are latent variable models which allow
latent state persistence and can be viewed as a generalization of the popular
hidden Markov models (HMMs). In this paper, we introduce a novel spectral
algorithm to perform inference in HSMMs. Unlike expectation maximization (EM),
our approach correctly estimates the probability of given observation sequence
based on a set of training sequences. Our approach is based on estimating
moments from the sample, whose number of dimensions depends only
logarithmically on the maximum length of the hidden state persistence.
Moreover, the algorithm requires only a few matrix inversions and is therefore
computationally efficient. Empirical evaluations on synthetic and real data
demonstrate the advantage of the algorithm over EM in terms of speed and
accuracy, especially for large datasets.
| no_new_dataset | 0.949716 |
1506.05907 | Bastien Mussard | Bastien Mussard (LCT, ICS), Peter Reinhardt (LCT), Janos Angyan (UL),
Julien Toulouse (LCT) | Spin-unrestricted random-phase approximation with range separation:
Benchmark on atomization energies and reaction barrier heights | null | Journal of Chemical Physics, American Institute of Physics, 2015,
pp.00 | 10.1063/1.4918710 | null | physics.chem-ph | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | We consider several spin-unrestricted random-phase approximation (RPA)
variants for calculating correlation energies, with and without range
separation, and test them on datasets of atomization energies and reaction
barrier heights. We show that range separation greatly improves the accuracy of
all RPA variants for these properties. Moreover, we show that a RPA variant
with exchange, hereafter referred to as RPAx-SO2, first proposed by Sz-abo and
Ostlund [A. Szabo and N. S. Ostlund, J. Chem. Phys. 67, 4351 (1977)] in a
spin-restricted closed-shell formalism, and extended here to a
spin-unrestricted formalism , provides on average the most accurate
range-separated RPA variant for atomization energies and reaction barrier
heights. Since this range-separated RPAx-SO2 method had already been shown to
be among the most accurate range-separated RPA variants for weak intermolecular
interactions [J. Toulouse, W. Zhu, A. Savin, G. Jansen, and J. G.
{\'A}ngy{\'a}n, J. Chem. Phys. 135, 084119 (2011)], this works confirms
range-separated RPAx-SO2 as a promising method for general chemical
applications.
| [
{
"version": "v1",
"created": "Fri, 19 Jun 2015 08:27:25 GMT"
},
{
"version": "v2",
"created": "Fri, 26 Feb 2016 14:07:17 GMT"
},
{
"version": "v3",
"created": "Mon, 29 Feb 2016 19:42:36 GMT"
}
] | 2016-03-01T00:00:00 | [
[
"Mussard",
"Bastien",
"",
"LCT, ICS"
],
[
"Reinhardt",
"Peter",
"",
"LCT"
],
[
"Angyan",
"Janos",
"",
"UL"
],
[
"Toulouse",
"Julien",
"",
"LCT"
]
] | TITLE: Spin-unrestricted random-phase approximation with range separation:
Benchmark on atomization energies and reaction barrier heights
ABSTRACT: We consider several spin-unrestricted random-phase approximation (RPA)
variants for calculating correlation energies, with and without range
separation, and test them on datasets of atomization energies and reaction
barrier heights. We show that range separation greatly improves the accuracy of
all RPA variants for these properties. Moreover, we show that a RPA variant
with exchange, hereafter referred to as RPAx-SO2, first proposed by Sz-abo and
Ostlund [A. Szabo and N. S. Ostlund, J. Chem. Phys. 67, 4351 (1977)] in a
spin-restricted closed-shell formalism, and extended here to a
spin-unrestricted formalism , provides on average the most accurate
range-separated RPA variant for atomization energies and reaction barrier
heights. Since this range-separated RPAx-SO2 method had already been shown to
be among the most accurate range-separated RPA variants for weak intermolecular
interactions [J. Toulouse, W. Zhu, A. Savin, G. Jansen, and J. G.
{\'A}ngy{\'a}n, J. Chem. Phys. 135, 084119 (2011)], this works confirms
range-separated RPAx-SO2 as a promising method for general chemical
applications.
| no_new_dataset | 0.951142 |
1511.02793 | Elman Mansimov | Elman Mansimov, Emilio Parisotto, Jimmy Lei Ba, Ruslan Salakhutdinov | Generating Images from Captions with Attention | Published as a conference paper at ICLR 2016 | null | null | null | cs.LG cs.CV | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | Motivated by the recent progress in generative models, we introduce a model
that generates images from natural language descriptions. The proposed model
iteratively draws patches on a canvas, while attending to the relevant words in
the description. After training on Microsoft COCO, we compare our model with
several baseline generative models on image generation and retrieval tasks. We
demonstrate that our model produces higher quality samples than other
approaches and generates images with novel scene compositions corresponding to
previously unseen captions in the dataset.
| [
{
"version": "v1",
"created": "Mon, 9 Nov 2015 18:18:53 GMT"
},
{
"version": "v2",
"created": "Mon, 29 Feb 2016 17:56:29 GMT"
}
] | 2016-03-01T00:00:00 | [
[
"Mansimov",
"Elman",
""
],
[
"Parisotto",
"Emilio",
""
],
[
"Ba",
"Jimmy Lei",
""
],
[
"Salakhutdinov",
"Ruslan",
""
]
] | TITLE: Generating Images from Captions with Attention
ABSTRACT: Motivated by the recent progress in generative models, we introduce a model
that generates images from natural language descriptions. The proposed model
iteratively draws patches on a canvas, while attending to the relevant words in
the description. After training on Microsoft COCO, we compare our model with
several baseline generative models on image generation and retrieval tasks. We
demonstrate that our model produces higher quality samples than other
approaches and generates images with novel scene compositions corresponding to
previously unseen captions in the dataset.
| no_new_dataset | 0.960398 |
1511.04773 | Weiran Wang | Weiran Wang, Karen Livescu | Large-Scale Approximate Kernel Canonical Correlation Analysis | Published as a conference paper at International Conference on
Learning Representations (ICLR) 2016 | null | null | null | cs.LG | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | Kernel canonical correlation analysis (KCCA) is a nonlinear multi-view
representation learning technique with broad applicability in statistics and
machine learning. Although there is a closed-form solution for the KCCA
objective, it involves solving an $N\times N$ eigenvalue system where $N$ is
the training set size, making its computational requirements in both memory and
time prohibitive for large-scale problems. Various approximation techniques
have been developed for KCCA. A commonly used approach is to first transform
the original inputs to an $M$-dimensional random feature space so that inner
products in the feature space approximate kernel evaluations, and then apply
linear CCA to the transformed inputs. In many applications, however, the
dimensionality $M$ of the random feature space may need to be very large in
order to obtain a sufficiently good approximation; it then becomes challenging
to perform the linear CCA step on the resulting very high-dimensional data
matrices. We show how to use a stochastic optimization algorithm, recently
proposed for linear CCA and its neural-network extension, to further alleviate
the computation requirements of approximate KCCA. This approach allows us to
run approximate KCCA on a speech dataset with $1.4$ million training samples
and a random feature space of dimensionality $M=100000$ on a typical
workstation.
| [
{
"version": "v1",
"created": "Sun, 15 Nov 2015 22:20:02 GMT"
},
{
"version": "v2",
"created": "Tue, 17 Nov 2015 16:31:14 GMT"
},
{
"version": "v3",
"created": "Thu, 7 Jan 2016 00:27:20 GMT"
},
{
"version": "v4",
"created": "Mon, 29 Feb 2016 16:04:46 GMT"
}
] | 2016-03-01T00:00:00 | [
[
"Wang",
"Weiran",
""
],
[
"Livescu",
"Karen",
""
]
] | TITLE: Large-Scale Approximate Kernel Canonical Correlation Analysis
ABSTRACT: Kernel canonical correlation analysis (KCCA) is a nonlinear multi-view
representation learning technique with broad applicability in statistics and
machine learning. Although there is a closed-form solution for the KCCA
objective, it involves solving an $N\times N$ eigenvalue system where $N$ is
the training set size, making its computational requirements in both memory and
time prohibitive for large-scale problems. Various approximation techniques
have been developed for KCCA. A commonly used approach is to first transform
the original inputs to an $M$-dimensional random feature space so that inner
products in the feature space approximate kernel evaluations, and then apply
linear CCA to the transformed inputs. In many applications, however, the
dimensionality $M$ of the random feature space may need to be very large in
order to obtain a sufficiently good approximation; it then becomes challenging
to perform the linear CCA step on the resulting very high-dimensional data
matrices. We show how to use a stochastic optimization algorithm, recently
proposed for linear CCA and its neural-network extension, to further alleviate
the computation requirements of approximate KCCA. This approach allows us to
run approximate KCCA on a speech dataset with $1.4$ million training samples
and a random feature space of dimensionality $M=100000$ on a typical
workstation.
| no_new_dataset | 0.941815 |
1511.05440 | Michael Mathieu | Michael Mathieu, Camille Couprie and Yann LeCun | Deep multi-scale video prediction beyond mean square error | null | null | null | null | cs.LG cs.CV stat.ML | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | Learning to predict future images from a video sequence involves the
construction of an internal representation that models the image evolution
accurately, and therefore, to some degree, its content and dynamics. This is
why pixel-space video prediction may be viewed as a promising avenue for
unsupervised feature learning. In addition, while optical flow has been a very
studied problem in computer vision for a long time, future frame prediction is
rarely approached. Still, many vision applications could benefit from the
knowledge of the next frames of videos, that does not require the complexity of
tracking every pixel trajectories. In this work, we train a convolutional
network to generate future frames given an input sequence. To deal with the
inherently blurry predictions obtained from the standard Mean Squared Error
(MSE) loss function, we propose three different and complementary feature
learning strategies: a multi-scale architecture, an adversarial training
method, and an image gradient difference loss function. We compare our
predictions to different published results based on recurrent neural networks
on the UCF101 dataset
| [
{
"version": "v1",
"created": "Tue, 17 Nov 2015 15:36:32 GMT"
},
{
"version": "v2",
"created": "Thu, 19 Nov 2015 23:21:22 GMT"
},
{
"version": "v3",
"created": "Mon, 23 Nov 2015 04:58:24 GMT"
},
{
"version": "v4",
"created": "Thu, 7 Jan 2016 21:52:53 GMT"
},
{
"version": "v5",
"created": "Fri, 15 Jan 2016 02:09:16 GMT"
},
{
"version": "v6",
"created": "Fri, 26 Feb 2016 22:10:30 GMT"
}
] | 2016-03-01T00:00:00 | [
[
"Mathieu",
"Michael",
""
],
[
"Couprie",
"Camille",
""
],
[
"LeCun",
"Yann",
""
]
] | TITLE: Deep multi-scale video prediction beyond mean square error
ABSTRACT: Learning to predict future images from a video sequence involves the
construction of an internal representation that models the image evolution
accurately, and therefore, to some degree, its content and dynamics. This is
why pixel-space video prediction may be viewed as a promising avenue for
unsupervised feature learning. In addition, while optical flow has been a very
studied problem in computer vision for a long time, future frame prediction is
rarely approached. Still, many vision applications could benefit from the
knowledge of the next frames of videos, that does not require the complexity of
tracking every pixel trajectories. In this work, we train a convolutional
network to generate future frames given an input sequence. To deal with the
inherently blurry predictions obtained from the standard Mean Squared Error
(MSE) loss function, we propose three different and complementary feature
learning strategies: a multi-scale architecture, an adversarial training
method, and an image gradient difference loss function. We compare our
predictions to different published results based on recurrent neural networks
on the UCF101 dataset
| no_new_dataset | 0.949342 |
1511.06051 | Robert Nishihara | Philipp Moritz, Robert Nishihara, Ion Stoica, Michael I. Jordan | SparkNet: Training Deep Networks in Spark | 12 pages, 7 figures | null | null | null | stat.ML cs.DC cs.LG cs.NE math.OC | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | Training deep networks is a time-consuming process, with networks for object
recognition often requiring multiple days to train. For this reason, leveraging
the resources of a cluster to speed up training is an important area of work.
However, widely-popular batch-processing computational frameworks like
MapReduce and Spark were not designed to support the asynchronous and
communication-intensive workloads of existing distributed deep learning
systems. We introduce SparkNet, a framework for training deep networks in
Spark. Our implementation includes a convenient interface for reading data from
Spark RDDs, a Scala interface to the Caffe deep learning framework, and a
lightweight multi-dimensional tensor library. Using a simple parallelization
scheme for stochastic gradient descent, SparkNet scales well with the cluster
size and tolerates very high-latency communication. Furthermore, it is easy to
deploy and use with no parameter tuning, and it is compatible with existing
Caffe models. We quantify the dependence of the speedup obtained by SparkNet on
the number of machines, the communication frequency, and the cluster's
communication overhead, and we benchmark our system's performance on the
ImageNet dataset.
| [
{
"version": "v1",
"created": "Thu, 19 Nov 2015 03:29:56 GMT"
},
{
"version": "v2",
"created": "Thu, 26 Nov 2015 10:35:40 GMT"
},
{
"version": "v3",
"created": "Wed, 6 Jan 2016 07:48:06 GMT"
},
{
"version": "v4",
"created": "Sun, 28 Feb 2016 23:43:36 GMT"
}
] | 2016-03-01T00:00:00 | [
[
"Moritz",
"Philipp",
""
],
[
"Nishihara",
"Robert",
""
],
[
"Stoica",
"Ion",
""
],
[
"Jordan",
"Michael I.",
""
]
] | TITLE: SparkNet: Training Deep Networks in Spark
ABSTRACT: Training deep networks is a time-consuming process, with networks for object
recognition often requiring multiple days to train. For this reason, leveraging
the resources of a cluster to speed up training is an important area of work.
However, widely-popular batch-processing computational frameworks like
MapReduce and Spark were not designed to support the asynchronous and
communication-intensive workloads of existing distributed deep learning
systems. We introduce SparkNet, a framework for training deep networks in
Spark. Our implementation includes a convenient interface for reading data from
Spark RDDs, a Scala interface to the Caffe deep learning framework, and a
lightweight multi-dimensional tensor library. Using a simple parallelization
scheme for stochastic gradient descent, SparkNet scales well with the cluster
size and tolerates very high-latency communication. Furthermore, it is easy to
deploy and use with no parameter tuning, and it is compatible with existing
Caffe models. We quantify the dependence of the speedup obtained by SparkNet on
the number of machines, the communication frequency, and the cluster's
communication overhead, and we benchmark our system's performance on the
ImageNet dataset.
| no_new_dataset | 0.937669 |
1602.06245 | Paul Bendich | Paul Bendich, Ellen Gasparovic, Christopher J. Tralie, John Harer | Scaffoldings and Spines: Organizing High-Dimensional Data Using Cover
Trees, Local Principal Component Analysis, and Persistent Homology | 14 pages | null | null | null | cs.CG | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | We propose a flexible and multi-scale method for organizing, visualizing, and
understanding datasets sampled from or near stratified spaces. The first part
of the algorithm produces a cover tree using adaptive thresholds based on a
combination of multi-scale local principal component analysis and topological
data analysis. The resulting cover tree nodes consist of points within or near
the same stratum of the stratified space. They are then connected to form a
\emph{scaffolding} graph, which is then simplified and collapsed down into a
\emph{spine} graph. From this latter graph the stratified structure becomes
apparent. We demonstrate our technique on several synthetic point cloud
examples and we use it to understand song structure in musical audio data.
| [
{
"version": "v1",
"created": "Fri, 19 Feb 2016 18:19:05 GMT"
},
{
"version": "v2",
"created": "Sat, 27 Feb 2016 05:21:51 GMT"
}
] | 2016-03-01T00:00:00 | [
[
"Bendich",
"Paul",
""
],
[
"Gasparovic",
"Ellen",
""
],
[
"Tralie",
"Christopher J.",
""
],
[
"Harer",
"John",
""
]
] | TITLE: Scaffoldings and Spines: Organizing High-Dimensional Data Using Cover
Trees, Local Principal Component Analysis, and Persistent Homology
ABSTRACT: We propose a flexible and multi-scale method for organizing, visualizing, and
understanding datasets sampled from or near stratified spaces. The first part
of the algorithm produces a cover tree using adaptive thresholds based on a
combination of multi-scale local principal component analysis and topological
data analysis. The resulting cover tree nodes consist of points within or near
the same stratum of the stratified space. They are then connected to form a
\emph{scaffolding} graph, which is then simplified and collapsed down into a
\emph{spine} graph. From this latter graph the stratified structure becomes
apparent. We demonstrate our technique on several synthetic point cloud
examples and we use it to understand song structure in musical audio data.
| no_new_dataset | 0.956675 |
1602.06550 | Igor Melnyk | Igor Melnyk, Arindam Banerjee, Bryan Matthews, and Nikunj Oza | Semi-Markov Switching Vector Autoregressive Model-based Anomaly
Detection in Aviation Systems | null | null | null | null | cs.LG stat.AP stat.ML | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | In this work we consider the problem of anomaly detection in heterogeneous,
multivariate, variable-length time series datasets. Our focus is on the
aviation safety domain, where data objects are flights and time series are
sensor readings and pilot switches. In this context the goal is to detect
anomalous flight segments, due to mechanical, environmental, or human factors
in order to identifying operationally significant events and provide insights
into the flight operations and highlight otherwise unavailable potential safety
risks and precursors to accidents. For this purpose, we propose a framework
which represents each flight using a semi-Markov switching vector
autoregressive (SMS-VAR) model. Detection of anomalies is then based on
measuring dissimilarities between the model's prediction and data observation.
The framework is scalable, due to the inherent parallel nature of most
computations, and can be used to perform online anomaly detection. Extensive
experimental results on simulated and real datasets illustrate that the
framework can detect various types of anomalies along with the key parameters
involved.
| [
{
"version": "v1",
"created": "Sun, 21 Feb 2016 16:55:36 GMT"
},
{
"version": "v2",
"created": "Sun, 28 Feb 2016 23:12:31 GMT"
}
] | 2016-03-01T00:00:00 | [
[
"Melnyk",
"Igor",
""
],
[
"Banerjee",
"Arindam",
""
],
[
"Matthews",
"Bryan",
""
],
[
"Oza",
"Nikunj",
""
]
] | TITLE: Semi-Markov Switching Vector Autoregressive Model-based Anomaly
Detection in Aviation Systems
ABSTRACT: In this work we consider the problem of anomaly detection in heterogeneous,
multivariate, variable-length time series datasets. Our focus is on the
aviation safety domain, where data objects are flights and time series are
sensor readings and pilot switches. In this context the goal is to detect
anomalous flight segments, due to mechanical, environmental, or human factors
in order to identifying operationally significant events and provide insights
into the flight operations and highlight otherwise unavailable potential safety
risks and precursors to accidents. For this purpose, we propose a framework
which represents each flight using a semi-Markov switching vector
autoregressive (SMS-VAR) model. Detection of anomalies is then based on
measuring dissimilarities between the model's prediction and data observation.
The framework is scalable, due to the inherent parallel nature of most
computations, and can be used to perform online anomaly detection. Extensive
experimental results on simulated and real datasets illustrate that the
framework can detect various types of anomalies along with the key parameters
involved.
| no_new_dataset | 0.943764 |
1602.08721 | Oren Kalinsky | Oren Kalinsky, Yoav Etsion, Benny Kimelfeld | Flexible Caching in Trie Joins | null | null | null | null | cs.DB | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | Traditional algorithms for multiway join computation are based on rewriting
the order of joins and combining results of intermediate subqueries. Recently,
several approaches have been proposed for algorithms that are "worst-case
optimal" wherein all relations are scanned simultaneously. An example is
Veldhuizen's Leapfrog Trie Join (LFTJ). An important advantage of LFTJ is its
small memory footprint, due to the fact that intermediate results are full
tuples that can be dumped immediately. However, since the algorithm does not
store intermediate results, recurring joins must be reconstructed from the
source relations, resulting in excessive memory traffic. In this paper, we
address this problem by incorporating caches into LFTJ. We do so by adopting
recent developments on join optimization, tying variable ordering to tree
decomposition. While the traditional usage of tree decomposition computes the
result for each bag in advance, our proposed approach incorporates caching
directly into LFTJ and can dynamically adjust the size of the cache.
Consequently, our solution balances memory usage and repeated computation, as
confirmed by our experiments over SNAP datasets.
| [
{
"version": "v1",
"created": "Sun, 28 Feb 2016 14:26:08 GMT"
}
] | 2016-03-01T00:00:00 | [
[
"Kalinsky",
"Oren",
""
],
[
"Etsion",
"Yoav",
""
],
[
"Kimelfeld",
"Benny",
""
]
] | TITLE: Flexible Caching in Trie Joins
ABSTRACT: Traditional algorithms for multiway join computation are based on rewriting
the order of joins and combining results of intermediate subqueries. Recently,
several approaches have been proposed for algorithms that are "worst-case
optimal" wherein all relations are scanned simultaneously. An example is
Veldhuizen's Leapfrog Trie Join (LFTJ). An important advantage of LFTJ is its
small memory footprint, due to the fact that intermediate results are full
tuples that can be dumped immediately. However, since the algorithm does not
store intermediate results, recurring joins must be reconstructed from the
source relations, resulting in excessive memory traffic. In this paper, we
address this problem by incorporating caches into LFTJ. We do so by adopting
recent developments on join optimization, tying variable ordering to tree
decomposition. While the traditional usage of tree decomposition computes the
result for each bag in advance, our proposed approach incorporates caching
directly into LFTJ and can dynamically adjust the size of the cache.
Consequently, our solution balances memory usage and repeated computation, as
confirmed by our experiments over SNAP datasets.
| no_new_dataset | 0.940298 |
1602.08791 | Vijay Gadepally | Vijay Gadepally, Jennie Duggan, Aaron Elmore, Jeremy Kepner, Samuel
Madden, Tim Mattson, Michael Stonebraker | The BigDAWG Architecture | null | null | null | null | cs.DB | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | BigDAWG is a polystore system designed to work on complex problems that
naturally span across different processing or storage engines. BigDAWG provides
an architecture that supports diverse database systems working with different
data models, support for the competing notions of location transparency and
semantic completeness via islands of information and a middleware that provides
a uniform multi-island interface. In this article, we describe the current
architecture of BigDAWG, its application on the MIMIC II medical dataset, and
our plans for the mechanics of cross-system queries. During the presentation,
we will also deliver a brief demonstration of the current version of BigDAWG.
| [
{
"version": "v1",
"created": "Mon, 29 Feb 2016 00:49:11 GMT"
}
] | 2016-03-01T00:00:00 | [
[
"Gadepally",
"Vijay",
""
],
[
"Duggan",
"Jennie",
""
],
[
"Elmore",
"Aaron",
""
],
[
"Kepner",
"Jeremy",
""
],
[
"Madden",
"Samuel",
""
],
[
"Mattson",
"Tim",
""
],
[
"Stonebraker",
"Michael",
""
]
] | TITLE: The BigDAWG Architecture
ABSTRACT: BigDAWG is a polystore system designed to work on complex problems that
naturally span across different processing or storage engines. BigDAWG provides
an architecture that supports diverse database systems working with different
data models, support for the competing notions of location transparency and
semantic completeness via islands of information and a middleware that provides
a uniform multi-island interface. In this article, we describe the current
architecture of BigDAWG, its application on the MIMIC II medical dataset, and
our plans for the mechanics of cross-system queries. During the presentation,
we will also deliver a brief demonstration of the current version of BigDAWG.
| no_new_dataset | 0.950088 |
1602.08855 | Corneliu Florea | Corneliu Florea, Razvan Condorovici, Constantin Vertan, Raluca Boia,
Laura Florea, Ruxandra Vranceanu | Pandora: Description of a Painting Database for Art Movement Recognition
with Baselines and Perspectives | 11 pages, 1 figure, 6 tables | null | null | null | cs.CV | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | To facilitate computer analysis of visual art, in the form of paintings, we
introduce Pandora (Paintings Dataset for Recognizing the Art movement)
database, a collection of digitized paintings labelled with respect to the
artistic movement. Noting that the set of databases available as benchmarks for
evaluation is highly reduced and most existing ones are limited in variability
and number of images, we propose a novel large scale dataset of digital
paintings. The database consists of more than 7700 images from 12 art
movements. Each genre is illustrated by a number of images varying from 250 to
nearly 1000. We investigate how local and global features and classification
systems are able to recognize the art movement. Our experimental results
suggest that accurate recognition is achievable by a combination of various
categories.To facilitate computer analysis of visual art, in the form of
paintings, we introduce Pandora (Paintings Dataset for Recognizing the Art
movement) database, a collection of digitized paintings labelled with respect
to the artistic movement. Noting that the set of databases available as
benchmarks for evaluation is highly reduced and most existing ones are limited
in variability and number of images, we propose a novel large scale dataset of
digital paintings. The database consists of more than 7700 images from 12 art
movements. Each genre is illustrated by a number of images varying from 250 to
nearly 1000. We investigate how local and global features and classification
systems are able to recognize the art movement. Our experimental results
suggest that accurate recognition is achievable by a combination of various
categories.
| [
{
"version": "v1",
"created": "Mon, 29 Feb 2016 08:24:01 GMT"
}
] | 2016-03-01T00:00:00 | [
[
"Florea",
"Corneliu",
""
],
[
"Condorovici",
"Razvan",
""
],
[
"Vertan",
"Constantin",
""
],
[
"Boia",
"Raluca",
""
],
[
"Florea",
"Laura",
""
],
[
"Vranceanu",
"Ruxandra",
""
]
] | TITLE: Pandora: Description of a Painting Database for Art Movement Recognition
with Baselines and Perspectives
ABSTRACT: To facilitate computer analysis of visual art, in the form of paintings, we
introduce Pandora (Paintings Dataset for Recognizing the Art movement)
database, a collection of digitized paintings labelled with respect to the
artistic movement. Noting that the set of databases available as benchmarks for
evaluation is highly reduced and most existing ones are limited in variability
and number of images, we propose a novel large scale dataset of digital
paintings. The database consists of more than 7700 images from 12 art
movements. Each genre is illustrated by a number of images varying from 250 to
nearly 1000. We investigate how local and global features and classification
systems are able to recognize the art movement. Our experimental results
suggest that accurate recognition is achievable by a combination of various
categories.To facilitate computer analysis of visual art, in the form of
paintings, we introduce Pandora (Paintings Dataset for Recognizing the Art
movement) database, a collection of digitized paintings labelled with respect
to the artistic movement. Noting that the set of databases available as
benchmarks for evaluation is highly reduced and most existing ones are limited
in variability and number of images, we propose a novel large scale dataset of
digital paintings. The database consists of more than 7700 images from 12 art
movements. Each genre is illustrated by a number of images varying from 250 to
nearly 1000. We investigate how local and global features and classification
systems are able to recognize the art movement. Our experimental results
suggest that accurate recognition is achievable by a combination of various
categories.
| new_dataset | 0.964589 |
1602.08978 | Yoshiharu Maeno | Yoshiharu Maeno | Epidemiological geographic profiling for a meta-population network | null | null | null | null | cs.SI physics.soc-ph | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | Epidemiological geographic profiling is a statistical method for making
inferences about likely areas of a source from the geographical distribution of
patients. Epidemiological geographic profiling algorithms are developed to
locate a source from the dataset on the number of new cases for a
meta-population network model. It is found from the WHO dataset on the SARS
outbreak that Hong Kong remains the most likely source throughout the period of
observation. This reasoning is pertinent under the restricted circumstance that
the number of reported probable cases in China was missing, unreliable, and
incomprehensive. It may also imply that globally connected Hong Kong was more
influential as a spreader than China. Singapore, Taiwan, Canada, and the United
States follow Hong Kong in the likeliness ranking list.
| [
{
"version": "v1",
"created": "Tue, 17 Nov 2015 07:29:56 GMT"
}
] | 2016-03-01T00:00:00 | [
[
"Maeno",
"Yoshiharu",
""
]
] | TITLE: Epidemiological geographic profiling for a meta-population network
ABSTRACT: Epidemiological geographic profiling is a statistical method for making
inferences about likely areas of a source from the geographical distribution of
patients. Epidemiological geographic profiling algorithms are developed to
locate a source from the dataset on the number of new cases for a
meta-population network model. It is found from the WHO dataset on the SARS
outbreak that Hong Kong remains the most likely source throughout the period of
observation. This reasoning is pertinent under the restricted circumstance that
the number of reported probable cases in China was missing, unreliable, and
incomprehensive. It may also imply that globally connected Hong Kong was more
influential as a spreader than China. Singapore, Taiwan, Canada, and the United
States follow Hong Kong in the likeliness ranking list.
| no_new_dataset | 0.95096 |
1602.08986 | G\'eraud Le Falher | G\'eraud Le Falher and Fabio Vitale | Even Trolls Are Useful: Efficient Link Classification in Signed Networks | 17 pages, 3 figures | null | null | null | cs.LG cs.SI physics.soc-ph | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | We address the problem of classifying the links of signed social networks
given their full structural topology. Motivated by a binary user behaviour
assumption, which is supported by decades of research in psychology, we develop
an efficient and surprisingly simple approach to solve this classification
problem. Our methods operate both within the active and batch settings. We
demonstrate that the algorithms we developed are extremely fast in both
theoretical and practical terms. Within the active setting, we provide a new
complexity measure and a rigorous analysis of our methods that hold for
arbitrary signed networks. We validate our theoretical claims carrying out a
set of experiments on three well known real-world datasets, showing that our
methods outperform the competitors while being much faster.
| [
{
"version": "v1",
"created": "Mon, 29 Feb 2016 14:45:18 GMT"
}
] | 2016-03-01T00:00:00 | [
[
"Falher",
"Géraud Le",
""
],
[
"Vitale",
"Fabio",
""
]
] | TITLE: Even Trolls Are Useful: Efficient Link Classification in Signed Networks
ABSTRACT: We address the problem of classifying the links of signed social networks
given their full structural topology. Motivated by a binary user behaviour
assumption, which is supported by decades of research in psychology, we develop
an efficient and surprisingly simple approach to solve this classification
problem. Our methods operate both within the active and batch settings. We
demonstrate that the algorithms we developed are extremely fast in both
theoretical and practical terms. Within the active setting, we provide a new
complexity measure and a rigorous analysis of our methods that hold for
arbitrary signed networks. We validate our theoretical claims carrying out a
set of experiments on three well known real-world datasets, showing that our
methods outperform the competitors while being much faster.
| no_new_dataset | 0.944485 |
1510.03009 | Zhouhan Lin | Zhouhan Lin, Matthieu Courbariaux, Roland Memisevic, Yoshua Bengio | Neural Networks with Few Multiplications | Published as a conference paper at ICLR 2016. 9 pages, 3 figures | null | null | null | cs.LG cs.NE | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | For most deep learning algorithms training is notoriously time consuming.
Since most of the computation in training neural networks is typically spent on
floating point multiplications, we investigate an approach to training that
eliminates the need for most of these. Our method consists of two parts: First
we stochastically binarize weights to convert multiplications involved in
computing hidden states to sign changes. Second, while back-propagating error
derivatives, in addition to binarizing the weights, we quantize the
representations at each layer to convert the remaining multiplications into
binary shifts. Experimental results across 3 popular datasets (MNIST, CIFAR10,
SVHN) show that this approach not only does not hurt classification performance
but can result in even better performance than standard stochastic gradient
descent training, paving the way to fast, hardware-friendly training of neural
networks.
| [
{
"version": "v1",
"created": "Sun, 11 Oct 2015 04:32:39 GMT"
},
{
"version": "v2",
"created": "Mon, 9 Nov 2015 20:16:10 GMT"
},
{
"version": "v3",
"created": "Fri, 26 Feb 2016 05:24:30 GMT"
}
] | 2016-02-29T00:00:00 | [
[
"Lin",
"Zhouhan",
""
],
[
"Courbariaux",
"Matthieu",
""
],
[
"Memisevic",
"Roland",
""
],
[
"Bengio",
"Yoshua",
""
]
] | TITLE: Neural Networks with Few Multiplications
ABSTRACT: For most deep learning algorithms training is notoriously time consuming.
Since most of the computation in training neural networks is typically spent on
floating point multiplications, we investigate an approach to training that
eliminates the need for most of these. Our method consists of two parts: First
we stochastically binarize weights to convert multiplications involved in
computing hidden states to sign changes. Second, while back-propagating error
derivatives, in addition to binarizing the weights, we quantize the
representations at each layer to convert the remaining multiplications into
binary shifts. Experimental results across 3 popular datasets (MNIST, CIFAR10,
SVHN) show that this approach not only does not hurt classification performance
but can result in even better performance than standard stochastic gradient
descent training, paving the way to fast, hardware-friendly training of neural
networks.
| no_new_dataset | 0.949153 |
1511.06426 | Moontae Lee | Moontae Lee, Xiaodong He, Wen-tau Yih, Jianfeng Gao, Li Deng, Paul
Smolensky | Reasoning in Vector Space: An Exploratory Study of Question Answering | null | null | null | null | cs.CL | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | Question answering tasks have shown remarkable progress with distributed
vector representation. In this paper, we investigate the recently proposed
Facebook bAbI tasks which consist of twenty different categories of questions
that require complex reasoning. Because the previous work on bAbI are all
end-to-end models, errors could come from either an imperfect understanding of
semantics or in certain steps of the reasoning. For clearer analysis, we
propose two vector space models inspired by Tensor Product Representation (TPR)
to perform knowledge encoding and logical reasoning based on common-sense
inference. They together achieve near-perfect accuracy on all categories
including positional reasoning and path finding that have proved difficult for
most of the previous approaches. We hypothesize that the difficulties in these
categories are due to the multi-relations in contrast to uni-relational
characteristic of other categories. Our exploration sheds light on designing
more sophisticated dataset and moving one step toward integrating transparent
and interpretable formalism of TPR into existing learning paradigms.
| [
{
"version": "v1",
"created": "Thu, 19 Nov 2015 22:30:10 GMT"
},
{
"version": "v2",
"created": "Thu, 7 Jan 2016 22:30:01 GMT"
},
{
"version": "v3",
"created": "Tue, 19 Jan 2016 11:16:46 GMT"
},
{
"version": "v4",
"created": "Fri, 26 Feb 2016 18:49:34 GMT"
}
] | 2016-02-29T00:00:00 | [
[
"Lee",
"Moontae",
""
],
[
"He",
"Xiaodong",
""
],
[
"Yih",
"Wen-tau",
""
],
[
"Gao",
"Jianfeng",
""
],
[
"Deng",
"Li",
""
],
[
"Smolensky",
"Paul",
""
]
] | TITLE: Reasoning in Vector Space: An Exploratory Study of Question Answering
ABSTRACT: Question answering tasks have shown remarkable progress with distributed
vector representation. In this paper, we investigate the recently proposed
Facebook bAbI tasks which consist of twenty different categories of questions
that require complex reasoning. Because the previous work on bAbI are all
end-to-end models, errors could come from either an imperfect understanding of
semantics or in certain steps of the reasoning. For clearer analysis, we
propose two vector space models inspired by Tensor Product Representation (TPR)
to perform knowledge encoding and logical reasoning based on common-sense
inference. They together achieve near-perfect accuracy on all categories
including positional reasoning and path finding that have proved difficult for
most of the previous approaches. We hypothesize that the difficulties in these
categories are due to the multi-relations in contrast to uni-relational
characteristic of other categories. Our exploration sheds light on designing
more sophisticated dataset and moving one step toward integrating transparent
and interpretable formalism of TPR into existing learning paradigms.
| new_dataset | 0.968768 |
1602.08141 | Thomas Castelli | Thomas Castelli, Aidean Sharghi, Don Harper, Alain Tremeau and Mubarak
Shah | Autonomous navigation for low-altitude UAVs in urban areas | null | null | null | null | cs.RO cs.CV | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | In recent years, consumer Unmanned Aerial Vehicles have become very popular,
everyone can buy and fly a drone without previous experience, which raises
concern in regards to regulations and public safety. In this paper, we present
a novel approach towards enabling safe operation of such vehicles in urban
areas. Our method uses geodetically accurate dataset images with Geographical
Information System (GIS) data of road networks and buildings provided by Google
Maps, to compute a weighted A* shortest path from start to end locations of a
mission. Weights represent the potential risk of injuries for individuals in
all categories of land-use, i.e. flying over buildings is considered safer than
above roads. We enable safe UAV operation in regards to 1- land-use by
computing a static global path dependent on environmental structures, and 2-
avoiding flying over moving objects such as cars and pedestrians by dynamically
optimizing the path locally during the flight. As all input sources are first
geo-registered, pixels and GPS coordinates are equivalent, it therefore allows
us to generate an automated and user-friendly mission with GPS waypoints
readable by consumer drones' autopilots. We simulated 54 missions and show
significant improvement in maximizing UAV's standoff distance to moving objects
with a quantified safety parameter over 40 times better than the naive straight
line navigation.
| [
{
"version": "v1",
"created": "Thu, 25 Feb 2016 22:43:14 GMT"
}
] | 2016-02-29T00:00:00 | [
[
"Castelli",
"Thomas",
""
],
[
"Sharghi",
"Aidean",
""
],
[
"Harper",
"Don",
""
],
[
"Tremeau",
"Alain",
""
],
[
"Shah",
"Mubarak",
""
]
] | TITLE: Autonomous navigation for low-altitude UAVs in urban areas
ABSTRACT: In recent years, consumer Unmanned Aerial Vehicles have become very popular,
everyone can buy and fly a drone without previous experience, which raises
concern in regards to regulations and public safety. In this paper, we present
a novel approach towards enabling safe operation of such vehicles in urban
areas. Our method uses geodetically accurate dataset images with Geographical
Information System (GIS) data of road networks and buildings provided by Google
Maps, to compute a weighted A* shortest path from start to end locations of a
mission. Weights represent the potential risk of injuries for individuals in
all categories of land-use, i.e. flying over buildings is considered safer than
above roads. We enable safe UAV operation in regards to 1- land-use by
computing a static global path dependent on environmental structures, and 2-
avoiding flying over moving objects such as cars and pedestrians by dynamically
optimizing the path locally during the flight. As all input sources are first
geo-registered, pixels and GPS coordinates are equivalent, it therefore allows
us to generate an automated and user-friendly mission with GPS waypoints
readable by consumer drones' autopilots. We simulated 54 missions and show
significant improvement in maximizing UAV's standoff distance to moving objects
with a quantified safety parameter over 40 times better than the naive straight
line navigation.
| no_new_dataset | 0.947284 |
1602.08225 | Wei Liu | Wei Liu, Wei-Long Zheng, Bao-Liang Lu | Multimodal Emotion Recognition Using Multimodal Deep Learning | null | null | null | null | cs.HC cs.CV cs.LG | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | To enhance the performance of affective models and reduce the cost of
acquiring physiological signals for real-world applications, we adopt
multimodal deep learning approach to construct affective models from multiple
physiological signals. For unimodal enhancement task, we indicate that the best
recognition accuracy of 82.11% on SEED dataset is achieved with shared
representations generated by Deep AutoEncoder (DAE) model. For multimodal
facilitation tasks, we demonstrate that the Bimodal Deep AutoEncoder (BDAE)
achieves the mean accuracies of 91.01% and 83.25% on SEED and DEAP datasets,
respectively, which are much superior to the state-of-the-art approaches. For
cross-modal learning task, our experimental results demonstrate that the mean
accuracy of 66.34% is achieved on SEED dataset through shared representations
generated by EEG-based DAE as training samples and shared representations
generated by eye-based DAE as testing sample, and vice versa.
| [
{
"version": "v1",
"created": "Fri, 26 Feb 2016 07:43:14 GMT"
}
] | 2016-02-29T00:00:00 | [
[
"Liu",
"Wei",
""
],
[
"Zheng",
"Wei-Long",
""
],
[
"Lu",
"Bao-Liang",
""
]
] | TITLE: Multimodal Emotion Recognition Using Multimodal Deep Learning
ABSTRACT: To enhance the performance of affective models and reduce the cost of
acquiring physiological signals for real-world applications, we adopt
multimodal deep learning approach to construct affective models from multiple
physiological signals. For unimodal enhancement task, we indicate that the best
recognition accuracy of 82.11% on SEED dataset is achieved with shared
representations generated by Deep AutoEncoder (DAE) model. For multimodal
facilitation tasks, we demonstrate that the Bimodal Deep AutoEncoder (BDAE)
achieves the mean accuracies of 91.01% and 83.25% on SEED and DEAP datasets,
respectively, which are much superior to the state-of-the-art approaches. For
cross-modal learning task, our experimental results demonstrate that the mean
accuracy of 66.34% is achieved on SEED dataset through shared representations
generated by EEG-based DAE as training samples and shared representations
generated by eye-based DAE as testing sample, and vice versa.
| no_new_dataset | 0.951504 |
1602.08393 | Anshumali Shrivastava | Anshumali Shrivastava | Exact Weighted Minwise Hashing in Constant Time | null | null | null | null | cs.DS cs.IR | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | Weighted minwise hashing (WMH) is one of the fundamental subroutine, required
by many celebrated approximation algorithms, commonly adopted in industrial
practice for large scale-search and learning. The resource bottleneck of the
algorithms is the computation of multiple (typically a few hundreds to
thousands) independent hashes of the data. The fastest hashing algorithm is by
Ioffe \cite{Proc:Ioffe_ICDM10}, which requires one pass over the entire data
vector, $O(d)$ ($d$ is the number of non-zeros), for computing one hash.
However, the requirement of multiple hashes demands hundreds or thousands
passes over the data. This is very costly for modern massive dataset.
In this work, we break this expensive barrier and show an expected constant
amortized time algorithm which computes $k$ independent and unbiased WMH in
time $O(k)$ instead of $O(dk)$ required by Ioffe's method. Moreover, our
proposal only needs a few bits (5 - 9 bits) of storage per hash value compared
to around $64$ bits required by the state-of-art-methodologies. Experimental
evaluations, on real datasets, show that for computing 500 WMH, our proposal
can be 60000x faster than the Ioffe's method without losing any accuracy. Our
method is also around 100x faster than approximate heuristics capitalizing on
the efficient "densified" one permutation hashing schemes
\cite{Proc:OneHashLSH_ICML14}. Given the simplicity of our approach and its
significant advantages, we hope that it will replace existing implementations
in practice.
| [
{
"version": "v1",
"created": "Fri, 26 Feb 2016 16:55:08 GMT"
}
] | 2016-02-29T00:00:00 | [
[
"Shrivastava",
"Anshumali",
""
]
] | TITLE: Exact Weighted Minwise Hashing in Constant Time
ABSTRACT: Weighted minwise hashing (WMH) is one of the fundamental subroutine, required
by many celebrated approximation algorithms, commonly adopted in industrial
practice for large scale-search and learning. The resource bottleneck of the
algorithms is the computation of multiple (typically a few hundreds to
thousands) independent hashes of the data. The fastest hashing algorithm is by
Ioffe \cite{Proc:Ioffe_ICDM10}, which requires one pass over the entire data
vector, $O(d)$ ($d$ is the number of non-zeros), for computing one hash.
However, the requirement of multiple hashes demands hundreds or thousands
passes over the data. This is very costly for modern massive dataset.
In this work, we break this expensive barrier and show an expected constant
amortized time algorithm which computes $k$ independent and unbiased WMH in
time $O(k)$ instead of $O(dk)$ required by Ioffe's method. Moreover, our
proposal only needs a few bits (5 - 9 bits) of storage per hash value compared
to around $64$ bits required by the state-of-art-methodologies. Experimental
evaluations, on real datasets, show that for computing 500 WMH, our proposal
can be 60000x faster than the Ioffe's method without losing any accuracy. Our
method is also around 100x faster than approximate heuristics capitalizing on
the efficient "densified" one permutation hashing schemes
\cite{Proc:OneHashLSH_ICML14}. Given the simplicity of our approach and its
significant advantages, we hope that it will replace existing implementations
in practice.
| no_new_dataset | 0.944177 |
1602.08447 | Le Hoang Son | Mumtaz Ali, Nguyen Van Minh, Le Hoang Son | A Neutrosophic Recommender System for Medical Diagnosis Based on
Algebraic Neutrosophic Measures | Keywords: Medical diagnosis, neutrosophic set, neutrosophic
recommender system, non-linear regression model | null | null | null | cs.AI | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | Neutrosophic set has the ability to handle uncertain, incomplete,
inconsistent, indeterminate information in a more accurate way. In this paper,
we proposed a neutrosophic recommender system to predict the diseases based on
neutrosophic set which includes single-criterion neutrosophic recommender
system (SC-NRS) and multi-criterion neutrosophic recommender system (MC-NRS).
Further, we investigated some algebraic operations of neutrosophic recommender
system such as union, complement, intersection, probabilistic sum, bold sum,
bold intersection, bounded difference, symmetric difference, convex linear sum
of min and max operators, Cartesian product, associativity, commutativity and
distributive. Based on these operations, we studied the algebraic structures
such as lattices, Kleen algebra, de Morgan algebra, Brouwerian algebra, BCK
algebra, Stone algebra and MV algebra. In addition, we introduced several types
of similarity measures based on these algebraic operations and studied some of
their theoretic properties. Moreover, we accomplished a prediction formula
using the proposed algebraic similarity measure. We also proposed a new
algorithm for medical diagnosis based on neutrosophic recommender system.
Finally to check the validity of the proposed methodology, we made experiments
on the datasets Heart, RHC, Breast cancer, Diabetes and DMD. At the end, we
presented the MSE and computational time by comparing the proposed algorithm
with the relevant ones such as ICSM, DSM, CARE, CFMD, as well as other variants
namely Variant 67, Variant 69, and Varian 71 both in tabular and graphical form
to analyze the efficiency and accuracy. Finally we analyzed the strength of all
8 algorithms by ANOVA statistical tool.
| [
{
"version": "v1",
"created": "Thu, 25 Feb 2016 03:20:00 GMT"
}
] | 2016-02-29T00:00:00 | [
[
"Ali",
"Mumtaz",
""
],
[
"Van Minh",
"Nguyen",
""
],
[
"Son",
"Le Hoang",
""
]
] | TITLE: A Neutrosophic Recommender System for Medical Diagnosis Based on
Algebraic Neutrosophic Measures
ABSTRACT: Neutrosophic set has the ability to handle uncertain, incomplete,
inconsistent, indeterminate information in a more accurate way. In this paper,
we proposed a neutrosophic recommender system to predict the diseases based on
neutrosophic set which includes single-criterion neutrosophic recommender
system (SC-NRS) and multi-criterion neutrosophic recommender system (MC-NRS).
Further, we investigated some algebraic operations of neutrosophic recommender
system such as union, complement, intersection, probabilistic sum, bold sum,
bold intersection, bounded difference, symmetric difference, convex linear sum
of min and max operators, Cartesian product, associativity, commutativity and
distributive. Based on these operations, we studied the algebraic structures
such as lattices, Kleen algebra, de Morgan algebra, Brouwerian algebra, BCK
algebra, Stone algebra and MV algebra. In addition, we introduced several types
of similarity measures based on these algebraic operations and studied some of
their theoretic properties. Moreover, we accomplished a prediction formula
using the proposed algebraic similarity measure. We also proposed a new
algorithm for medical diagnosis based on neutrosophic recommender system.
Finally to check the validity of the proposed methodology, we made experiments
on the datasets Heart, RHC, Breast cancer, Diabetes and DMD. At the end, we
presented the MSE and computational time by comparing the proposed algorithm
with the relevant ones such as ICSM, DSM, CARE, CFMD, as well as other variants
namely Variant 67, Variant 69, and Varian 71 both in tabular and graphical form
to analyze the efficiency and accuracy. Finally we analyzed the strength of all
8 algorithms by ANOVA statistical tool.
| no_new_dataset | 0.945551 |
1508.06163 | Guangliang Cheng | Guangliang Cheng, Feiyun Zhu, Shiming Xiang and Chunhong Pan | Accurate Urban Road Centerline Extraction from VHR Imagery via
Multiscale Segmentation and Tensor Voting | 25 pages, 11 figures | null | null | null | cs.CV | http://creativecommons.org/licenses/by-nc-sa/4.0/ | It is very useful and increasingly popular to extract accurate road
centerlines from very-high-resolution (VHR) re- mote sensing imagery for
various applications, such as road map generation and updating etc. There are
three shortcomings of current methods: (a) Due to the noise and occlusions
(owing to vehicles and trees), most road extraction methods bring in
heterogeneous classification results; (b) Morphological thinning algorithm is
widely used to extract road centerlines, while it pro- duces small spurs around
the centerlines; (c) Many methods are ineffective to extract centerlines around
the road intersections. To address the above three issues, we propose a novel
method to ex- tract smooth and complete road centerlines via three techniques:
the multiscale joint collaborative representation (MJCR) & graph cuts (GC),
tensor voting (TV) & non-maximum suppression (NMS) and fitting based connection
algorithm. Specifically, a MJCR-GC based road area segmentation method is
proposed by incorporating mutiscale features and spatial information. In this
way, a homogenous road segmentation result is achieved. Then, to obtain a
smooth and correct road centerline network, a TV-NMS based centerline
extraction method is introduced. This method not only extracts smooth road
centerlines, but also connects the discontinuous road centerlines. Finally, to
overcome the ineffectiveness of current methods in the road intersection, a
fitting based road centerline connection algorithm is proposed. As a result, we
can get a complete road centerline network. Extensive experiments on two
datasets demonstrate that our method achieves higher quantitative results, as
well as more satisfactory visual performances by comparing with state-of-the-
art methods.
| [
{
"version": "v1",
"created": "Tue, 25 Aug 2015 14:18:34 GMT"
},
{
"version": "v2",
"created": "Thu, 25 Feb 2016 15:29:25 GMT"
}
] | 2016-02-26T00:00:00 | [
[
"Cheng",
"Guangliang",
""
],
[
"Zhu",
"Feiyun",
""
],
[
"Xiang",
"Shiming",
""
],
[
"Pan",
"Chunhong",
""
]
] | TITLE: Accurate Urban Road Centerline Extraction from VHR Imagery via
Multiscale Segmentation and Tensor Voting
ABSTRACT: It is very useful and increasingly popular to extract accurate road
centerlines from very-high-resolution (VHR) re- mote sensing imagery for
various applications, such as road map generation and updating etc. There are
three shortcomings of current methods: (a) Due to the noise and occlusions
(owing to vehicles and trees), most road extraction methods bring in
heterogeneous classification results; (b) Morphological thinning algorithm is
widely used to extract road centerlines, while it pro- duces small spurs around
the centerlines; (c) Many methods are ineffective to extract centerlines around
the road intersections. To address the above three issues, we propose a novel
method to ex- tract smooth and complete road centerlines via three techniques:
the multiscale joint collaborative representation (MJCR) & graph cuts (GC),
tensor voting (TV) & non-maximum suppression (NMS) and fitting based connection
algorithm. Specifically, a MJCR-GC based road area segmentation method is
proposed by incorporating mutiscale features and spatial information. In this
way, a homogenous road segmentation result is achieved. Then, to obtain a
smooth and correct road centerline network, a TV-NMS based centerline
extraction method is introduced. This method not only extracts smooth road
centerlines, but also connects the discontinuous road centerlines. Finally, to
overcome the ineffectiveness of current methods in the road intersection, a
fitting based road centerline connection algorithm is proposed. As a result, we
can get a complete road centerline network. Extensive experiments on two
datasets demonstrate that our method achieves higher quantitative results, as
well as more satisfactory visual performances by comparing with state-of-the-
art methods.
| no_new_dataset | 0.954774 |
1602.06844 | Hao Wu | Hao Wu, Yue Ning, Prithwish Chakraborty, Jilles Vreeken, Nikolaj Tatti
and Naren Ramakrishnan | Generating Realistic Synthetic Population Datasets | The conference version of the paper is submitted for publication | null | null | null | cs.DB | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | Modern studies of societal phenomena rely on the availability of large
datasets capturing attributes and activities of synthetic, city-level,
populations. For instance, in epidemiology, synthetic population datasets are
necessary to study disease propagation and intervention measures before
implementation. In social science, synthetic population datasets are needed to
understand how policy decisions might affect preferences and behaviors of
individuals. In public health, synthetic population datasets are necessary to
capture diagnostic and procedural characteristics of patient records without
violating confidentialities of individuals. To generate such datasets over a
large set of categorical variables, we propose the use of the maximum entropy
principle to formalize a generative model such that in a statistically
well-founded way we can optimally utilize given prior information about the
data, and are unbiased otherwise. An efficient inference algorithm is designed
to estimate the maximum entropy model, and we demonstrate how our approach is
adept at estimating underlying data distributions. We evaluate this approach
against both simulated data and on US census datasets, and demonstrate its
feasibility using an epidemic simulation application.
| [
{
"version": "v1",
"created": "Mon, 22 Feb 2016 16:28:17 GMT"
},
{
"version": "v2",
"created": "Tue, 23 Feb 2016 16:50:26 GMT"
},
{
"version": "v3",
"created": "Thu, 25 Feb 2016 17:20:42 GMT"
}
] | 2016-02-26T00:00:00 | [
[
"Wu",
"Hao",
""
],
[
"Ning",
"Yue",
""
],
[
"Chakraborty",
"Prithwish",
""
],
[
"Vreeken",
"Jilles",
""
],
[
"Tatti",
"Nikolaj",
""
],
[
"Ramakrishnan",
"Naren",
""
]
] | TITLE: Generating Realistic Synthetic Population Datasets
ABSTRACT: Modern studies of societal phenomena rely on the availability of large
datasets capturing attributes and activities of synthetic, city-level,
populations. For instance, in epidemiology, synthetic population datasets are
necessary to study disease propagation and intervention measures before
implementation. In social science, synthetic population datasets are needed to
understand how policy decisions might affect preferences and behaviors of
individuals. In public health, synthetic population datasets are necessary to
capture diagnostic and procedural characteristics of patient records without
violating confidentialities of individuals. To generate such datasets over a
large set of categorical variables, we propose the use of the maximum entropy
principle to formalize a generative model such that in a statistically
well-founded way we can optimally utilize given prior information about the
data, and are unbiased otherwise. An efficient inference algorithm is designed
to estimate the maximum entropy model, and we demonstrate how our approach is
adept at estimating underlying data distributions. We evaluate this approach
against both simulated data and on US census datasets, and demonstrate its
feasibility using an epidemic simulation application.
| no_new_dataset | 0.868994 |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.