id
stringlengths 9
16
| submitter
stringlengths 3
64
⌀ | authors
stringlengths 5
6.63k
| title
stringlengths 7
245
| comments
stringlengths 1
482
⌀ | journal-ref
stringlengths 4
382
⌀ | doi
stringlengths 9
151
⌀ | report-no
stringclasses 984
values | categories
stringlengths 5
108
| license
stringclasses 9
values | abstract
stringlengths 83
3.41k
| versions
listlengths 1
20
| update_date
timestamp[s]date 2007-05-23 00:00:00
2025-04-11 00:00:00
| authors_parsed
sequencelengths 1
427
| prompt
stringlengths 166
3.49k
| label
stringclasses 2
values | prob
float64 0.5
0.98
|
---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
1509.04777 | Daniele Spiga | D. Spiga, L. Raimondi | X-ray optical systems: from metrology to Point Spread Function | null | Proceedings of the SPIE, Advances in Computational Methods for
X-Ray Optics III, Vol. 9209, 92090E (2014) | 10.1117/12.2061657 | null | physics.optics astro-ph.IM | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | One of the problems often encountered in X-ray mirror manufacturing is
setting proper manufacturing tolerances to guarantee an angular resolution -
often expressed in terms of Point Spread Function (PSF) - as needed by the
specific science goal. To do this, we need an accurate metrological apparatus,
covering a very broad range of spatial frequencies, and an affordable method to
compute the PSF from the metrology dataset. [...] However, the separation
between these spectral ranges is difficult do define exactly, and it is also
unclear how to affordably combine the PSFs, computed with different methods in
different spectral ranges, into a PSF expectation at a given X-ray energy. For
this reason, we have proposed a method entirely based on the Huygens-Fresnel
principle to compute the diffracted field of real Wolter-I optics, including
measured defects over a wide range of spatial frequencies. Owing to the shallow
angles at play, the computation can be simplified limiting the computation to
the longitudinal profiles, neglecting completely the effect of roundness
errors. Other authors had already proposed similar approaches in the past, but
only in far-field approximation, therefore they could not be applied to the
case of Wolter-I optics, in which two reflections occur in sequence within a
short range. The method we suggest is versatile, as it can be applied to
multiple reflection systems, at any X-ray energy, and regardless of the nominal
shape of the mirrors in the optical system. The method has been implemented in
the WISE code, successfully used to explain the measured PSFs of
multilayer-coated optics for astronomic use, and of a K-B optical system in use
at the FERMI free electron laser.
| [
{
"version": "v1",
"created": "Wed, 16 Sep 2015 00:29:46 GMT"
}
] | 2015-09-17T00:00:00 | [
[
"Spiga",
"D.",
""
],
[
"Raimondi",
"L.",
""
]
] | TITLE: X-ray optical systems: from metrology to Point Spread Function
ABSTRACT: One of the problems often encountered in X-ray mirror manufacturing is
setting proper manufacturing tolerances to guarantee an angular resolution -
often expressed in terms of Point Spread Function (PSF) - as needed by the
specific science goal. To do this, we need an accurate metrological apparatus,
covering a very broad range of spatial frequencies, and an affordable method to
compute the PSF from the metrology dataset. [...] However, the separation
between these spectral ranges is difficult do define exactly, and it is also
unclear how to affordably combine the PSFs, computed with different methods in
different spectral ranges, into a PSF expectation at a given X-ray energy. For
this reason, we have proposed a method entirely based on the Huygens-Fresnel
principle to compute the diffracted field of real Wolter-I optics, including
measured defects over a wide range of spatial frequencies. Owing to the shallow
angles at play, the computation can be simplified limiting the computation to
the longitudinal profiles, neglecting completely the effect of roundness
errors. Other authors had already proposed similar approaches in the past, but
only in far-field approximation, therefore they could not be applied to the
case of Wolter-I optics, in which two reflections occur in sequence within a
short range. The method we suggest is versatile, as it can be applied to
multiple reflection systems, at any X-ray energy, and regardless of the nominal
shape of the mirrors in the optical system. The method has been implemented in
the WISE code, successfully used to explain the measured PSFs of
multilayer-coated optics for astronomic use, and of a K-B optical system in use
at the FERMI free electron laser.
| no_new_dataset | 0.944689 |
1509.04783 | Ziming Zhang | Ziming Zhang, Yuting Chen, and Venkatesh Saligrama | Group Membership Prediction | accepted for ICCV 2015 | null | null | null | cs.CV stat.ML | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | The group membership prediction (GMP) problem involves predicting whether or
not a collection of instances share a certain semantic property. For instance,
in kinship verification given a collection of images, the goal is to predict
whether or not they share a {\it familial} relationship. In this context we
propose a novel probability model and introduce latent {\em view-specific} and
{\em view-shared} random variables to jointly account for the view-specific
appearance and cross-view similarities among data instances. Our model posits
that data from each view is independent conditioned on the shared variables.
This postulate leads to a parametric probability model that decomposes group
membership likelihood into a tensor product of data-independent parameters and
data-dependent factors. We propose learning the data-independent parameters in
a discriminative way with bilinear classifiers, and test our prediction
algorithm on challenging visual recognition tasks such as multi-camera person
re-identification and kinship verification. On most benchmark datasets, our
method can significantly outperform the current state-of-the-art.
| [
{
"version": "v1",
"created": "Wed, 16 Sep 2015 01:22:40 GMT"
}
] | 2015-09-17T00:00:00 | [
[
"Zhang",
"Ziming",
""
],
[
"Chen",
"Yuting",
""
],
[
"Saligrama",
"Venkatesh",
""
]
] | TITLE: Group Membership Prediction
ABSTRACT: The group membership prediction (GMP) problem involves predicting whether or
not a collection of instances share a certain semantic property. For instance,
in kinship verification given a collection of images, the goal is to predict
whether or not they share a {\it familial} relationship. In this context we
propose a novel probability model and introduce latent {\em view-specific} and
{\em view-shared} random variables to jointly account for the view-specific
appearance and cross-view similarities among data instances. Our model posits
that data from each view is independent conditioned on the shared variables.
This postulate leads to a parametric probability model that decomposes group
membership likelihood into a tensor product of data-independent parameters and
data-dependent factors. We propose learning the data-independent parameters in
a discriminative way with bilinear classifiers, and test our prediction
algorithm on challenging visual recognition tasks such as multi-camera person
re-identification and kinship verification. On most benchmark datasets, our
method can significantly outperform the current state-of-the-art.
| no_new_dataset | 0.947866 |
1509.04904 | Tshilidzi Marwala | Pramod Kumar Parida, Tshilidzi Marwala and Snehashish Chakraverty | Causal Model Analysis using Collider v-structure with Negative
Percentage Mapping | null | null | null | null | cs.AI | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | A major problem of causal inference is the arrangement of dependent nodes in
a directed acyclic graph (DAG) with path coefficients and observed confounders.
Path coefficients do not provide the units to measure the strength of
information flowing from one node to the other. Here we proposed the method of
causal structure learning using collider v-structures (CVS) with Negative
Percentage Mapping (NPM) to get selective thresholds of information strength,
to direct the edges and subjective confounders in a DAG. The NPM is used to
scale the strength of information passed through nodes in units of percentage
from interval from 0 to 1. The causal structures are constructed by bottom up
approach using path coefficients, causal directions and confounders, derived
implementing collider v-structure and NPM. The method is self-sufficient to
observe all the latent confounders present in the causal model and capable of
detecting every responsible causal direction. The results are tested for
simulated datasets of non-Gaussian distributions and compared with DirectLiNGAM
and ICA-LiNGAM to check efficiency of the proposed method.
| [
{
"version": "v1",
"created": "Wed, 16 Sep 2015 12:37:30 GMT"
}
] | 2015-09-17T00:00:00 | [
[
"Parida",
"Pramod Kumar",
""
],
[
"Marwala",
"Tshilidzi",
""
],
[
"Chakraverty",
"Snehashish",
""
]
] | TITLE: Causal Model Analysis using Collider v-structure with Negative
Percentage Mapping
ABSTRACT: A major problem of causal inference is the arrangement of dependent nodes in
a directed acyclic graph (DAG) with path coefficients and observed confounders.
Path coefficients do not provide the units to measure the strength of
information flowing from one node to the other. Here we proposed the method of
causal structure learning using collider v-structures (CVS) with Negative
Percentage Mapping (NPM) to get selective thresholds of information strength,
to direct the edges and subjective confounders in a DAG. The NPM is used to
scale the strength of information passed through nodes in units of percentage
from interval from 0 to 1. The causal structures are constructed by bottom up
approach using path coefficients, causal directions and confounders, derived
implementing collider v-structure and NPM. The method is self-sufficient to
observe all the latent confounders present in the causal model and capable of
detecting every responsible causal direction. The results are tested for
simulated datasets of non-Gaussian distributions and compared with DirectLiNGAM
and ICA-LiNGAM to check efficiency of the proposed method.
| no_new_dataset | 0.94743 |
1509.04916 | Mengyang Yu | Li Liu, Mengyang Yu, Ling Shao | Projection Bank: From High-dimensional Data to Medium-length Binary
Codes | null | null | null | null | cs.CV | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | Recently, very high-dimensional feature representations, e.g., Fisher Vector,
have achieved excellent performance for visual recognition and retrieval.
However, these lengthy representations always cause extremely heavy
computational and storage costs and even become unfeasible in some large-scale
applications. A few existing techniques can transfer very high-dimensional data
into binary codes, but they still require the reduced code length to be
relatively long to maintain acceptable accuracies. To target a better balance
between computational efficiency and accuracies, in this paper, we propose a
novel embedding method called Binary Projection Bank (BPB), which can
effectively reduce the very high-dimensional representations to
medium-dimensional binary codes without sacrificing accuracies. Instead of
using conventional single linear or bilinear projections, the proposed method
learns a bank of small projections via the max-margin constraint to optimally
preserve the intrinsic data similarity. We have systematically evaluated the
proposed method on three datasets: Flickr 1M, ILSVR2010 and UCF101, showing
competitive retrieval and recognition accuracies compared with state-of-the-art
approaches, but with a significantly smaller memory footprint and lower coding
complexity.
| [
{
"version": "v1",
"created": "Wed, 16 Sep 2015 13:42:42 GMT"
}
] | 2015-09-17T00:00:00 | [
[
"Liu",
"Li",
""
],
[
"Yu",
"Mengyang",
""
],
[
"Shao",
"Ling",
""
]
] | TITLE: Projection Bank: From High-dimensional Data to Medium-length Binary
Codes
ABSTRACT: Recently, very high-dimensional feature representations, e.g., Fisher Vector,
have achieved excellent performance for visual recognition and retrieval.
However, these lengthy representations always cause extremely heavy
computational and storage costs and even become unfeasible in some large-scale
applications. A few existing techniques can transfer very high-dimensional data
into binary codes, but they still require the reduced code length to be
relatively long to maintain acceptable accuracies. To target a better balance
between computational efficiency and accuracies, in this paper, we propose a
novel embedding method called Binary Projection Bank (BPB), which can
effectively reduce the very high-dimensional representations to
medium-dimensional binary codes without sacrificing accuracies. Instead of
using conventional single linear or bilinear projections, the proposed method
learns a bank of small projections via the max-margin constraint to optimally
preserve the intrinsic data similarity. We have systematically evaluated the
proposed method on three datasets: Flickr 1M, ILSVR2010 and UCF101, showing
competitive retrieval and recognition accuracies compared with state-of-the-art
approaches, but with a significantly smaller memory footprint and lower coding
complexity.
| no_new_dataset | 0.948917 |
1509.04942 | Xu Jia | Xu Jia and Efstratios Gavves and Basura Fernando and Tinne Tuytelaars | Guiding Long-Short Term Memory for Image Caption Generation | accepted by ICCV 2015 | null | null | null | cs.CV | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | In this work we focus on the problem of image caption generation. We propose
an extension of the long short term memory (LSTM) model, which we coin gLSTM
for short. In particular, we add semantic information extracted from the image
as extra input to each unit of the LSTM block, with the aim of guiding the
model towards solutions that are more tightly coupled to the image content.
Additionally, we explore different length normalization strategies for beam
search in order to prevent from favoring short sentences. On various benchmark
datasets such as Flickr8K, Flickr30K and MS COCO, we obtain results that are on
par with or even outperform the current state-of-the-art.
| [
{
"version": "v1",
"created": "Wed, 16 Sep 2015 15:02:30 GMT"
}
] | 2015-09-17T00:00:00 | [
[
"Jia",
"Xu",
""
],
[
"Gavves",
"Efstratios",
""
],
[
"Fernando",
"Basura",
""
],
[
"Tuytelaars",
"Tinne",
""
]
] | TITLE: Guiding Long-Short Term Memory for Image Caption Generation
ABSTRACT: In this work we focus on the problem of image caption generation. We propose
an extension of the long short term memory (LSTM) model, which we coin gLSTM
for short. In particular, we add semantic information extracted from the image
as extra input to each unit of the LSTM block, with the aim of guiding the
model towards solutions that are more tightly coupled to the image content.
Additionally, we explore different length normalization strategies for beam
search in order to prevent from favoring short sentences. On various benchmark
datasets such as Flickr8K, Flickr30K and MS COCO, we obtain results that are on
par with or even outperform the current state-of-the-art.
| no_new_dataset | 0.951774 |
1509.04954 | Heng Yang | Heng Yang and Renqiao Zhang and Peter Robinson | Human and Sheep Facial Landmarks Localisation by Triplet Interpolated
Features | submitted to WACV2016 | null | null | null | cs.CV | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | In this paper we present a method for localisation of facial landmarks on
human and sheep. We introduce a new feature extraction scheme called
triplet-interpolated feature used at each iteration of the cascaded shape
regression framework. It is able to extract features from similar semantic
location given an estimated shape, even when head pose variations are large and
the facial landmarks are very sparsely distributed. Furthermore, we study the
impact of training data imbalance on model performance and propose a training
sample augmentation scheme that produces more initialisations for training
samples from the minority. More specifically, the augmentation number for a
training sample is made to be negatively correlated to the value of the fitted
probability density function at the sample's position. We evaluate the proposed
scheme on both human and sheep facial landmarks localisation. On the benchmark
300w human face dataset, we demonstrate the benefits of our proposed methods
and show very competitive performance when comparing to other methods. On a
newly created sheep face dataset, we get very good performance despite the fact
that we only have a limited number of training samples and a set of sparse
landmarks are annotated.
| [
{
"version": "v1",
"created": "Wed, 16 Sep 2015 15:50:01 GMT"
}
] | 2015-09-17T00:00:00 | [
[
"Yang",
"Heng",
""
],
[
"Zhang",
"Renqiao",
""
],
[
"Robinson",
"Peter",
""
]
] | TITLE: Human and Sheep Facial Landmarks Localisation by Triplet Interpolated
Features
ABSTRACT: In this paper we present a method for localisation of facial landmarks on
human and sheep. We introduce a new feature extraction scheme called
triplet-interpolated feature used at each iteration of the cascaded shape
regression framework. It is able to extract features from similar semantic
location given an estimated shape, even when head pose variations are large and
the facial landmarks are very sparsely distributed. Furthermore, we study the
impact of training data imbalance on model performance and propose a training
sample augmentation scheme that produces more initialisations for training
samples from the minority. More specifically, the augmentation number for a
training sample is made to be negatively correlated to the value of the fitted
probability density function at the sample's position. We evaluate the proposed
scheme on both human and sheep facial landmarks localisation. On the benchmark
300w human face dataset, we demonstrate the benefits of our proposed methods
and show very competitive performance when comparing to other methods. On a
newly created sheep face dataset, we get very good performance despite the fact
that we only have a limited number of training samples and a set of sparse
landmarks are annotated.
| new_dataset | 0.959269 |
1509.04397 | Suriya Gunasekar | Suriya Gunasekar, Pradeep Ravikumar, Joydeep Ghosh | Exponential Family Matrix Completion under Structural Constraints | 20 pages, 9 figures | Gunasekar, Suriya, Pradeep Ravikumar, and Joydeep Ghosh.
"Exponential family matrix completion under structural constraints".
Proceedings of The 31st International Conference on Machine Learning, pp.
1917-1925, 2014 | null | null | stat.ML cs.LG | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | We consider the matrix completion problem of recovering a structured matrix
from noisy and partial measurements. Recent works have proposed tractable
estimators with strong statistical guarantees for the case where the underlying
matrix is low--rank, and the measurements consist of a subset, either of the
exact individual entries, or of the entries perturbed by additive Gaussian
noise, which is thus implicitly suited for thin--tailed continuous data.
Arguably, common applications of matrix completion require estimators for (a)
heterogeneous data--types, such as skewed--continuous, count, binary, etc., (b)
for heterogeneous noise models (beyond Gaussian), which capture varied
uncertainty in the measurements, and (c) heterogeneous structural constraints
beyond low--rank, such as block--sparsity, or a superposition structure of
low--rank plus elementwise sparseness, among others. In this paper, we provide
a vastly unified framework for generalized matrix completion by considering a
matrix completion setting wherein the matrix entries are sampled from any
member of the rich family of exponential family distributions; and impose
general structural constraints on the underlying matrix, as captured by a
general regularizer $\mathcal{R}(.)$. We propose a simple convex regularized
$M$--estimator for the generalized framework, and provide a unified and novel
statistical analysis for this general class of estimators. We finally
corroborate our theoretical results on simulated datasets.
| [
{
"version": "v1",
"created": "Tue, 15 Sep 2015 04:49:57 GMT"
}
] | 2015-09-16T00:00:00 | [
[
"Gunasekar",
"Suriya",
""
],
[
"Ravikumar",
"Pradeep",
""
],
[
"Ghosh",
"Joydeep",
""
]
] | TITLE: Exponential Family Matrix Completion under Structural Constraints
ABSTRACT: We consider the matrix completion problem of recovering a structured matrix
from noisy and partial measurements. Recent works have proposed tractable
estimators with strong statistical guarantees for the case where the underlying
matrix is low--rank, and the measurements consist of a subset, either of the
exact individual entries, or of the entries perturbed by additive Gaussian
noise, which is thus implicitly suited for thin--tailed continuous data.
Arguably, common applications of matrix completion require estimators for (a)
heterogeneous data--types, such as skewed--continuous, count, binary, etc., (b)
for heterogeneous noise models (beyond Gaussian), which capture varied
uncertainty in the measurements, and (c) heterogeneous structural constraints
beyond low--rank, such as block--sparsity, or a superposition structure of
low--rank plus elementwise sparseness, among others. In this paper, we provide
a vastly unified framework for generalized matrix completion by considering a
matrix completion setting wherein the matrix entries are sampled from any
member of the rich family of exponential family distributions; and impose
general structural constraints on the underlying matrix, as captured by a
general regularizer $\mathcal{R}(.)$. We propose a simple convex regularized
$M$--estimator for the generalized framework, and provide a unified and novel
statistical analysis for this general class of estimators. We finally
corroborate our theoretical results on simulated datasets.
| no_new_dataset | 0.946001 |
1509.04581 | Zhen Liu | Zhen Liu | Kernelized Deep Convolutional Neural Network for Describing Complex
Images | 9 pages | null | null | null | cs.CV cs.AI cs.IR cs.MM | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | With the impressive capability to capture visual content, deep convolutional
neural networks (CNN) have demon- strated promising performance in various
vision-based ap- plications, such as classification, recognition, and objec- t
detection. However, due to the intrinsic structure design of CNN, for images
with complex content, it achieves lim- ited capability on invariance to
translation, rotation, and re-sizing changes, which is strongly emphasized in
the s- cenario of content-based image retrieval. In this paper, to address this
problem, we proposed a new kernelized deep convolutional neural network. We
first discuss our motiva- tion by an experimental study to demonstrate the
sensitivi- ty of the global CNN feature to the basic geometric trans-
formations. Then, we propose to represent visual content with approximate
invariance to the above geometric trans- formations from a kernelized
perspective. We extract CNN features on the detected object-like patches and
aggregate these patch-level CNN features to form a vectorial repre- sentation
with the Fisher vector model. The effectiveness of our proposed algorithm is
demonstrated on image search application with three benchmark datasets.
| [
{
"version": "v1",
"created": "Tue, 15 Sep 2015 14:35:11 GMT"
}
] | 2015-09-16T00:00:00 | [
[
"Liu",
"Zhen",
""
]
] | TITLE: Kernelized Deep Convolutional Neural Network for Describing Complex
Images
ABSTRACT: With the impressive capability to capture visual content, deep convolutional
neural networks (CNN) have demon- strated promising performance in various
vision-based ap- plications, such as classification, recognition, and objec- t
detection. However, due to the intrinsic structure design of CNN, for images
with complex content, it achieves lim- ited capability on invariance to
translation, rotation, and re-sizing changes, which is strongly emphasized in
the s- cenario of content-based image retrieval. In this paper, to address this
problem, we proposed a new kernelized deep convolutional neural network. We
first discuss our motiva- tion by an experimental study to demonstrate the
sensitivi- ty of the global CNN feature to the basic geometric trans-
formations. Then, we propose to represent visual content with approximate
invariance to the above geometric trans- formations from a kernelized
perspective. We extract CNN features on the detected object-like patches and
aggregate these patch-level CNN features to form a vectorial repre- sentation
with the Fisher vector model. The effectiveness of our proposed algorithm is
demonstrated on image search application with three benchmark datasets.
| no_new_dataset | 0.948775 |
1308.0271 | Qiang Qiu | Qiang Qiu, Rama Chellappa | Compositional Dictionaries for Domain Adaptive Face Recognition | Transactions on Image Processing, 2015 | null | null | null | cs.CV | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | We present a dictionary learning approach to compensate for the
transformation of faces due to changes in view point, illumination, resolution,
etc. The key idea of our approach is to force domain-invariant sparse coding,
i.e., design a consistent sparse representation of the same face in different
domains. In this way, classifiers trained on the sparse codes in the source
domain consisting of frontal faces for example can be applied to the target
domain (consisting of faces in different poses, illumination conditions, etc)
without much loss in recognition accuracy. The approach is to first learn a
domain base dictionary, and then describe each domain shift (identity, pose,
illumination) using a sparse representation over the base dictionary. The
dictionary adapted to each domain is expressed as sparse linear combinations of
the base dictionary. In the context of face recognition, with the proposed
compositional dictionary approach, a face image can be decomposed into sparse
representations for a given subject, pose and illumination respectively. This
approach has three advantages: first, the extracted sparse representation for a
subject is consistent across domains and enables pose and illumination
insensitive face recognition. Second, sparse representations for pose and
illumination can subsequently be used to estimate the pose and illumination
condition of a face image. Finally, by composing sparse representations for
subject and the different domains, we can also perform pose alignment and
illumination normalization. Extensive experiments using two public face
datasets are presented to demonstrate the effectiveness of our approach for
face recognition.
| [
{
"version": "v1",
"created": "Thu, 1 Aug 2013 17:27:31 GMT"
},
{
"version": "v2",
"created": "Sat, 12 Sep 2015 20:55:51 GMT"
}
] | 2015-09-15T00:00:00 | [
[
"Qiu",
"Qiang",
""
],
[
"Chellappa",
"Rama",
""
]
] | TITLE: Compositional Dictionaries for Domain Adaptive Face Recognition
ABSTRACT: We present a dictionary learning approach to compensate for the
transformation of faces due to changes in view point, illumination, resolution,
etc. The key idea of our approach is to force domain-invariant sparse coding,
i.e., design a consistent sparse representation of the same face in different
domains. In this way, classifiers trained on the sparse codes in the source
domain consisting of frontal faces for example can be applied to the target
domain (consisting of faces in different poses, illumination conditions, etc)
without much loss in recognition accuracy. The approach is to first learn a
domain base dictionary, and then describe each domain shift (identity, pose,
illumination) using a sparse representation over the base dictionary. The
dictionary adapted to each domain is expressed as sparse linear combinations of
the base dictionary. In the context of face recognition, with the proposed
compositional dictionary approach, a face image can be decomposed into sparse
representations for a given subject, pose and illumination respectively. This
approach has three advantages: first, the extracted sparse representation for a
subject is consistent across domains and enables pose and illumination
insensitive face recognition. Second, sparse representations for pose and
illumination can subsequently be used to estimate the pose and illumination
condition of a face image. Finally, by composing sparse representations for
subject and the different domains, we can also perform pose alignment and
illumination normalization. Extensive experiments using two public face
datasets are presented to demonstrate the effectiveness of our approach for
face recognition.
| no_new_dataset | 0.94887 |
1409.4988 | Filippo Maria Bianchi | Filippo Maria Bianchi, Enrico Maiorino, Lorenzo Livi, Antonello Rizzi
and Alireza Sadeghian | An Agent-Based Algorithm exploiting Multiple Local Dissimilarities for
Clusters Mining and Knowledge Discovery | null | null | 10.1007/s00500-015-1876-1 | null | cs.LG cs.DC cs.MA | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | We propose a multi-agent algorithm able to automatically discover relevant
regularities in a given dataset, determining at the same time the set of
configurations of the adopted parametric dissimilarity measure yielding compact
and separated clusters. Each agent operates independently by performing a
Markovian random walk on a suitable weighted graph representation of the input
dataset. Such a weighted graph representation is induced by the specific
parameter configuration of the dissimilarity measure adopted by the agent,
which searches and takes decisions autonomously for one cluster at a time.
Results show that the algorithm is able to discover parameter configurations
that yield a consistent and interpretable collection of clusters. Moreover, we
demonstrate that our algorithm shows comparable performances with other similar
state-of-the-art algorithms when facing specific clustering problems.
| [
{
"version": "v1",
"created": "Wed, 17 Sep 2014 14:39:37 GMT"
}
] | 2015-09-15T00:00:00 | [
[
"Bianchi",
"Filippo Maria",
""
],
[
"Maiorino",
"Enrico",
""
],
[
"Livi",
"Lorenzo",
""
],
[
"Rizzi",
"Antonello",
""
],
[
"Sadeghian",
"Alireza",
""
]
] | TITLE: An Agent-Based Algorithm exploiting Multiple Local Dissimilarities for
Clusters Mining and Knowledge Discovery
ABSTRACT: We propose a multi-agent algorithm able to automatically discover relevant
regularities in a given dataset, determining at the same time the set of
configurations of the adopted parametric dissimilarity measure yielding compact
and separated clusters. Each agent operates independently by performing a
Markovian random walk on a suitable weighted graph representation of the input
dataset. Such a weighted graph representation is induced by the specific
parameter configuration of the dissimilarity measure adopted by the agent,
which searches and takes decisions autonomously for one cluster at a time.
Results show that the algorithm is able to discover parameter configurations
that yield a consistent and interpretable collection of clusters. Moreover, we
demonstrate that our algorithm shows comparable performances with other similar
state-of-the-art algorithms when facing specific clustering problems.
| no_new_dataset | 0.950319 |
1509.02587 | Bardia Yousefi | Bardia Yousefi, C.K. Loo | A Dual Fast and Slow Feature Interaction in Biologically Inspired Visual
Recognition of Human Action | This paper has been withdrawn by the author due to a mistake in file | null | null | null | cs.CV | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | Computational neuroscience studies that have examined human visual system
through functional magnetic resonance imaging (fMRI) have identified a model
where the mammalian brain pursues two distinct pathways (for recognition of
biological movement tasks). In the brain, dorsal stream analyzes the
information of motion (optical flow), which is the fast features, and ventral
stream (form pathway) analyzes form information (through active basis model
based incremental slow feature analysis ) as slow features. The proposed
approach suggests the motion perception of the human visual system composes of
fast and slow feature interactions that identifies biological movements. Form
features in the visual system biologically follows the application of active
basis model with incremental slow feature analysis for the extraction of the
slowest form features of human objects movements in the ventral stream.
Applying incremental slow feature analysis provides an opportunity to use the
action prototypes. To extract the slowest features episodic observation is
required but the fast features updates the processing of motion information in
every frames. Experimental results have shown promising accuracy for the
proposed model and good performance with two datasets (KTH and Weizmann).
| [
{
"version": "v1",
"created": "Wed, 9 Sep 2015 00:31:53 GMT"
},
{
"version": "v2",
"created": "Sat, 12 Sep 2015 23:44:42 GMT"
}
] | 2015-09-15T00:00:00 | [
[
"Yousefi",
"Bardia",
""
],
[
"Loo",
"C. K.",
""
]
] | TITLE: A Dual Fast and Slow Feature Interaction in Biologically Inspired Visual
Recognition of Human Action
ABSTRACT: Computational neuroscience studies that have examined human visual system
through functional magnetic resonance imaging (fMRI) have identified a model
where the mammalian brain pursues two distinct pathways (for recognition of
biological movement tasks). In the brain, dorsal stream analyzes the
information of motion (optical flow), which is the fast features, and ventral
stream (form pathway) analyzes form information (through active basis model
based incremental slow feature analysis ) as slow features. The proposed
approach suggests the motion perception of the human visual system composes of
fast and slow feature interactions that identifies biological movements. Form
features in the visual system biologically follows the application of active
basis model with incremental slow feature analysis for the extraction of the
slowest form features of human objects movements in the ventral stream.
Applying incremental slow feature analysis provides an opportunity to use the
action prototypes. To extract the slowest features episodic observation is
required but the fast features updates the processing of motion information in
every frames. Experimental results have shown promising accuracy for the
proposed model and good performance with two datasets (KTH and Weizmann).
| no_new_dataset | 0.954223 |
1509.02730 | Rangeet Mitra | Rangeet Mitra and Vimal Bhatia | Finite Dictionary Variants of the Diffusion KLMS Algorithm | null | null | null | null | cs.SY cs.DC cs.IT cs.LG math.IT | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | The diffusion based distributed learning approaches have been found to be a
viable solution for learning over linearly separable datasets over a network.
However, approaches till date are suitable for linearly separable datasets and
need to be extended to scenarios in which we need to learn a non-linearity. In
such scenarios, the recently proposed diffusion kernel least mean squares
(KLMS) has been found to be performing better than diffusion least mean squares
(LMS). The drawback of diffusion KLMS is that it requires infinite storage for
observations (also called dictionary). This paper formulates the diffusion KLMS
in a fixed budget setting such that the storage requirement is curtailed while
maintaining appreciable performance in terms of convergence. Simulations have
been carried out to validate the two newly proposed algorithms named as
quantised diffusion KLMS (QDKLMS) and fixed budget diffusion KLMS (FBDKLMS)
against KLMS, which indicate that both the proposed algorithms deliver better
performance as compared to the KLMS while reducing the dictionary size storage
requirement.
| [
{
"version": "v1",
"created": "Wed, 9 Sep 2015 11:38:01 GMT"
}
] | 2015-09-15T00:00:00 | [
[
"Mitra",
"Rangeet",
""
],
[
"Bhatia",
"Vimal",
""
]
] | TITLE: Finite Dictionary Variants of the Diffusion KLMS Algorithm
ABSTRACT: The diffusion based distributed learning approaches have been found to be a
viable solution for learning over linearly separable datasets over a network.
However, approaches till date are suitable for linearly separable datasets and
need to be extended to scenarios in which we need to learn a non-linearity. In
such scenarios, the recently proposed diffusion kernel least mean squares
(KLMS) has been found to be performing better than diffusion least mean squares
(LMS). The drawback of diffusion KLMS is that it requires infinite storage for
observations (also called dictionary). This paper formulates the diffusion KLMS
in a fixed budget setting such that the storage requirement is curtailed while
maintaining appreciable performance in terms of convergence. Simulations have
been carried out to validate the two newly proposed algorithms named as
quantised diffusion KLMS (QDKLMS) and fixed budget diffusion KLMS (FBDKLMS)
against KLMS, which indicate that both the proposed algorithms deliver better
performance as compared to the KLMS while reducing the dictionary size storage
requirement.
| no_new_dataset | 0.949012 |
1509.03302 | Matt Barnes | Matt Barnes, Kyle Miller, Artur Dubrawski | Performance Bounds for Pairwise Entity Resolution | null | null | null | null | stat.ML cs.CY cs.DB cs.LG | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | One significant challenge to scaling entity resolution algorithms to massive
datasets is understanding how performance changes after moving beyond the realm
of small, manually labeled reference datasets. Unlike traditional machine
learning tasks, when an entity resolution algorithm performs well on small
hold-out datasets, there is no guarantee this performance holds on larger
hold-out datasets. We prove simple bounding properties between the performance
of a match function on a small validation set and the performance of a pairwise
entity resolution algorithm on arbitrarily sized datasets. Thus, our approach
enables optimization of pairwise entity resolution algorithms for large
datasets, using a small set of labeled data.
| [
{
"version": "v1",
"created": "Thu, 10 Sep 2015 19:58:44 GMT"
}
] | 2015-09-15T00:00:00 | [
[
"Barnes",
"Matt",
""
],
[
"Miller",
"Kyle",
""
],
[
"Dubrawski",
"Artur",
""
]
] | TITLE: Performance Bounds for Pairwise Entity Resolution
ABSTRACT: One significant challenge to scaling entity resolution algorithms to massive
datasets is understanding how performance changes after moving beyond the realm
of small, manually labeled reference datasets. Unlike traditional machine
learning tasks, when an entity resolution algorithm performs well on small
hold-out datasets, there is no guarantee this performance holds on larger
hold-out datasets. We prove simple bounding properties between the performance
of a match function on a small validation set and the performance of a pairwise
entity resolution algorithm on arbitrarily sized datasets. Thus, our approach
enables optimization of pairwise entity resolution algorithms for large
datasets, using a small set of labeled data.
| no_new_dataset | 0.949435 |
1509.03844 | Yiannis Andreopoulos | Alhabib Abbas, Nikos Deligiannis and Yiannis Andreopoulos | Vectors of Locally Aggregated Centers for Compact Video Representation | Proc. IEEE International Conference on Multimedia and Expo, ICME
2015, Torino, Italy | null | 10.1109/ICME.2015.7177501 | null | cs.MM cs.CV cs.IR | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | We propose a novel vector aggregation technique for compact video
representation, with application in accurate similarity detection within large
video datasets. The current state-of-the-art in visual search is formed by the
vector of locally aggregated descriptors (VLAD) of Jegou et. al. VLAD generates
compact video representations based on scale-invariant feature transform (SIFT)
vectors (extracted per frame) and local feature centers computed over a
training set. With the aim to increase robustness to visual distortions, we
propose a new approach that operates at a coarser level in the feature
representation. We create vectors of locally aggregated centers (VLAC) by first
clustering SIFT features to obtain local feature centers (LFCs) and then
encoding the latter with respect to given centers of local feature centers
(CLFCs), extracted from a training set. The sum-of-differences between the LFCs
and the CLFCs are aggregated to generate an extremely-compact video description
used for accurate video segment similarity detection. Experimentation using a
video dataset, comprising more than 1000 minutes of content from the Open Video
Project, shows that VLAC obtains substantial gains in terms of mean Average
Precision (mAP) against VLAD and the hyper-pooling method of Douze et. al.,
under the same compaction factor and the same set of distortions.
| [
{
"version": "v1",
"created": "Sun, 13 Sep 2015 13:06:36 GMT"
}
] | 2015-09-15T00:00:00 | [
[
"Abbas",
"Alhabib",
""
],
[
"Deligiannis",
"Nikos",
""
],
[
"Andreopoulos",
"Yiannis",
""
]
] | TITLE: Vectors of Locally Aggregated Centers for Compact Video Representation
ABSTRACT: We propose a novel vector aggregation technique for compact video
representation, with application in accurate similarity detection within large
video datasets. The current state-of-the-art in visual search is formed by the
vector of locally aggregated descriptors (VLAD) of Jegou et. al. VLAD generates
compact video representations based on scale-invariant feature transform (SIFT)
vectors (extracted per frame) and local feature centers computed over a
training set. With the aim to increase robustness to visual distortions, we
propose a new approach that operates at a coarser level in the feature
representation. We create vectors of locally aggregated centers (VLAC) by first
clustering SIFT features to obtain local feature centers (LFCs) and then
encoding the latter with respect to given centers of local feature centers
(CLFCs), extracted from a training set. The sum-of-differences between the LFCs
and the CLFCs are aggregated to generate an extremely-compact video description
used for accurate video segment similarity detection. Experimentation using a
video dataset, comprising more than 1000 minutes of content from the Open Video
Project, shows that VLAC obtains substantial gains in terms of mean Average
Precision (mAP) against VLAD and the hyper-pooling method of Douze et. al.,
under the same compaction factor and the same set of distortions.
| no_new_dataset | 0.947721 |
1509.03936 | Zhanpeng Zhang | Zhanpeng Zhang, Ping Luo, Chen Change Loy, Xiaoou Tang | Learning Social Relation Traits from Face Images | To appear in International Conference on Computer Vision (ICCV) 2015 | null | null | null | cs.CV cs.CY | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | Social relation defines the association, e.g, warm, friendliness, and
dominance, between two or more people. Motivated by psychological studies, we
investigate if such fine-grained and high-level relation traits can be
characterised and quantified from face images in the wild. To address this
challenging problem we propose a deep model that learns a rich face
representation to capture gender, expression, head pose, and age-related
attributes, and then performs pairwise-face reasoning for relation prediction.
To learn from heterogeneous attribute sources, we formulate a new network
architecture with a bridging layer to leverage the inherent correspondences
among these datasets. It can also cope with missing target attribute labels.
Extensive experiments show that our approach is effective for fine-grained
social relation learning in images and videos.
| [
{
"version": "v1",
"created": "Mon, 14 Sep 2015 03:02:36 GMT"
}
] | 2015-09-15T00:00:00 | [
[
"Zhang",
"Zhanpeng",
""
],
[
"Luo",
"Ping",
""
],
[
"Loy",
"Chen Change",
""
],
[
"Tang",
"Xiaoou",
""
]
] | TITLE: Learning Social Relation Traits from Face Images
ABSTRACT: Social relation defines the association, e.g, warm, friendliness, and
dominance, between two or more people. Motivated by psychological studies, we
investigate if such fine-grained and high-level relation traits can be
characterised and quantified from face images in the wild. To address this
challenging problem we propose a deep model that learns a rich face
representation to capture gender, expression, head pose, and age-related
attributes, and then performs pairwise-face reasoning for relation prediction.
To learn from heterogeneous attribute sources, we formulate a new network
architecture with a bridging layer to leverage the inherent correspondences
among these datasets. It can also cope with missing target attribute labels.
Extensive experiments show that our approach is effective for fine-grained
social relation learning in images and videos.
| no_new_dataset | 0.949716 |
1509.03956 | Francesco Solera | Francesco Solera, Simone Calderara and Rita Cucchiara | Learning to Divide and Conquer for Online Multi-Target Tracking | null | null | null | null | cs.CV | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | Online Multiple Target Tracking (MTT) is often addressed within the
tracking-by-detection paradigm. Detections are previously extracted
independently in each frame and then objects trajectories are built by
maximizing specifically designed coherence functions. Nevertheless, ambiguities
arise in presence of occlusions or detection errors. In this paper we claim
that the ambiguities in tracking could be solved by a selective use of the
features, by working with more reliable features if possible and exploiting a
deeper representation of the target only if necessary. To this end, we propose
an online divide and conquer tracker for static camera scenes, which partitions
the assignment problem in local subproblems and solves them by selectively
choosing and combining the best features. The complete framework is cast as a
structural learning task that unifies these phases and learns tracker
parameters from examples. Experiments on two different datasets highlights a
significant improvement of tracking performances (MOTA +10%) over the state of
the art.
| [
{
"version": "v1",
"created": "Mon, 14 Sep 2015 05:25:52 GMT"
}
] | 2015-09-15T00:00:00 | [
[
"Solera",
"Francesco",
""
],
[
"Calderara",
"Simone",
""
],
[
"Cucchiara",
"Rita",
""
]
] | TITLE: Learning to Divide and Conquer for Online Multi-Target Tracking
ABSTRACT: Online Multiple Target Tracking (MTT) is often addressed within the
tracking-by-detection paradigm. Detections are previously extracted
independently in each frame and then objects trajectories are built by
maximizing specifically designed coherence functions. Nevertheless, ambiguities
arise in presence of occlusions or detection errors. In this paper we claim
that the ambiguities in tracking could be solved by a selective use of the
features, by working with more reliable features if possible and exploiting a
deeper representation of the target only if necessary. To this end, we propose
an online divide and conquer tracker for static camera scenes, which partitions
the assignment problem in local subproblems and solves them by selectively
choosing and combining the best features. The complete framework is cast as a
structural learning task that unifies these phases and learns tracker
parameters from examples. Experiments on two different datasets highlights a
significant improvement of tracking performances (MOTA +10%) over the state of
the art.
| no_new_dataset | 0.94545 |
1509.03413 | Saikat Basu | Saikat Basu, Manohar Karki, Sangram Ganguly, Robert DiBiano, Supratik
Mukhopadhyay and Ramakrishna Nemani | Learning Sparse Feature Representations using Probabilistic Quadtrees
and Deep Belief Nets | Published in the European Symposium on Artificial Neural Networks,
ESANN 2015 | null | null | null | cs.CV | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | Learning sparse feature representations is a useful instrument for solving an
unsupervised learning problem. In this paper, we present three labeled
handwritten digit datasets, collectively called n-MNIST. Then, we propose a
novel framework for the classification of handwritten digits that learns sparse
representations using probabilistic quadtrees and Deep Belief Nets. On the
MNIST and n-MNIST datasets, our framework shows promising results and
significantly outperforms traditional Deep Belief Networks.
| [
{
"version": "v1",
"created": "Fri, 11 Sep 2015 08:13:35 GMT"
}
] | 2015-09-14T00:00:00 | [
[
"Basu",
"Saikat",
""
],
[
"Karki",
"Manohar",
""
],
[
"Ganguly",
"Sangram",
""
],
[
"DiBiano",
"Robert",
""
],
[
"Mukhopadhyay",
"Supratik",
""
],
[
"Nemani",
"Ramakrishna",
""
]
] | TITLE: Learning Sparse Feature Representations using Probabilistic Quadtrees
and Deep Belief Nets
ABSTRACT: Learning sparse feature representations is a useful instrument for solving an
unsupervised learning problem. In this paper, we present three labeled
handwritten digit datasets, collectively called n-MNIST. Then, we propose a
novel framework for the classification of handwritten digits that learns sparse
representations using probabilistic quadtrees and Deep Belief Nets. On the
MNIST and n-MNIST datasets, our framework shows promising results and
significantly outperforms traditional Deep Belief Networks.
| new_dataset | 0.944689 |
1509.03456 | Abdeslam El Harraj | Abdeslam El Harraj, Naoufal Raissouni | OCR accuracy improvement on document images through a novel
pre-processing approach | null | Signal & Image Processing : An International Journal (SIPIJ)
Vol.6, No.4, August 2015 | 10.5121/sipij.2015.6401 | null | cs.CV | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | Digital camera and mobile document image acquisition are new trends arising
in the world of Optical Character Recognition and text detection. In some
cases, such process integrates many distortions and produces poorly scanned
text or text-photo images and natural images, leading to an unreliable OCR
digitization. In this paper, we present a novel nonparametric and unsupervised
method to compensate for undesirable document image distortions aiming to
optimally improve OCR accuracy. Our approach relies on a very efficient stack
of document image enhancing techniques to recover deformation of the entire
document image. First, we propose a local brightness and contrast adjustment
method to effectively handle lighting variations and the irregular distribution
of image illumination. Second, we use an optimized greyscale conversion
algorithm to transform our document image to greyscale level. Third, we sharpen
the useful information in the resulting greyscale image using Un-sharp Masking
method. Finally, an optimal global binarization approach is used to prepare the
final document image to OCR recognition. The proposed approach can
significantly improve text detection rate and optical character recognition
accuracy. To demonstrate the efficiency of our approach, an exhaustive
experimentation on a standard dataset is presented.
| [
{
"version": "v1",
"created": "Fri, 11 Sep 2015 10:52:52 GMT"
}
] | 2015-09-14T00:00:00 | [
[
"Harraj",
"Abdeslam El",
""
],
[
"Raissouni",
"Naoufal",
""
]
] | TITLE: OCR accuracy improvement on document images through a novel
pre-processing approach
ABSTRACT: Digital camera and mobile document image acquisition are new trends arising
in the world of Optical Character Recognition and text detection. In some
cases, such process integrates many distortions and produces poorly scanned
text or text-photo images and natural images, leading to an unreliable OCR
digitization. In this paper, we present a novel nonparametric and unsupervised
method to compensate for undesirable document image distortions aiming to
optimally improve OCR accuracy. Our approach relies on a very efficient stack
of document image enhancing techniques to recover deformation of the entire
document image. First, we propose a local brightness and contrast adjustment
method to effectively handle lighting variations and the irregular distribution
of image illumination. Second, we use an optimized greyscale conversion
algorithm to transform our document image to greyscale level. Third, we sharpen
the useful information in the resulting greyscale image using Un-sharp Masking
method. Finally, an optimal global binarization approach is used to prepare the
final document image to OCR recognition. The proposed approach can
significantly improve text detection rate and optical character recognition
accuracy. To demonstrate the efficiency of our approach, an exhaustive
experimentation on a standard dataset is presented.
| new_dataset | 0.966851 |
1509.03602 | Saikat Basu | Saikat Basu, Sangram Ganguly, Supratik Mukhopadhyay, Robert DiBiano,
Manohar Karki and Ramakrishna Nemani | DeepSat - A Learning framework for Satellite Imagery | Paper was accepted at ACM SIGSPATIAL 2015 | null | null | null | cs.CV | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | Satellite image classification is a challenging problem that lies at the
crossroads of remote sensing, computer vision, and machine learning. Due to the
high variability inherent in satellite data, most of the current object
classification approaches are not suitable for handling satellite datasets. The
progress of satellite image analytics has also been inhibited by the lack of a
single labeled high-resolution dataset with multiple class labels. The
contributions of this paper are twofold - (1) first, we present two new
satellite datasets called SAT-4 and SAT-6, and (2) then, we propose a
classification framework that extracts features from an input image, normalizes
them and feeds the normalized feature vectors to a Deep Belief Network for
classification. On the SAT-4 dataset, our best network produces a
classification accuracy of 97.95% and outperforms three state-of-the-art object
recognition algorithms, namely - Deep Belief Networks, Convolutional Neural
Networks and Stacked Denoising Autoencoders by ~11%. On SAT-6, it produces a
classification accuracy of 93.9% and outperforms the other algorithms by ~15%.
Comparative studies with a Random Forest classifier show the advantage of an
unsupervised learning approach over traditional supervised learning techniques.
A statistical analysis based on Distribution Separability Criterion and
Intrinsic Dimensionality Estimation substantiates the effectiveness of our
approach in learning better representations for satellite imagery.
| [
{
"version": "v1",
"created": "Fri, 11 Sep 2015 18:32:51 GMT"
}
] | 2015-09-14T00:00:00 | [
[
"Basu",
"Saikat",
""
],
[
"Ganguly",
"Sangram",
""
],
[
"Mukhopadhyay",
"Supratik",
""
],
[
"DiBiano",
"Robert",
""
],
[
"Karki",
"Manohar",
""
],
[
"Nemani",
"Ramakrishna",
""
]
] | TITLE: DeepSat - A Learning framework for Satellite Imagery
ABSTRACT: Satellite image classification is a challenging problem that lies at the
crossroads of remote sensing, computer vision, and machine learning. Due to the
high variability inherent in satellite data, most of the current object
classification approaches are not suitable for handling satellite datasets. The
progress of satellite image analytics has also been inhibited by the lack of a
single labeled high-resolution dataset with multiple class labels. The
contributions of this paper are twofold - (1) first, we present two new
satellite datasets called SAT-4 and SAT-6, and (2) then, we propose a
classification framework that extracts features from an input image, normalizes
them and feeds the normalized feature vectors to a Deep Belief Network for
classification. On the SAT-4 dataset, our best network produces a
classification accuracy of 97.95% and outperforms three state-of-the-art object
recognition algorithms, namely - Deep Belief Networks, Convolutional Neural
Networks and Stacked Denoising Autoencoders by ~11%. On SAT-6, it produces a
classification accuracy of 93.9% and outperforms the other algorithms by ~15%.
Comparative studies with a Random Forest classifier show the advantage of an
unsupervised learning approach over traditional supervised learning techniques.
A statistical analysis based on Distribution Separability Criterion and
Intrinsic Dimensionality Estimation substantiates the effectiveness of our
approach in learning better representations for satellite imagery.
| new_dataset | 0.950549 |
1407.5599 | Bo Dai | Bo Dai, Bo Xie, Niao He, Yingyu Liang, Anant Raj, Maria-Florina
Balcan, Le Song | Scalable Kernel Methods via Doubly Stochastic Gradients | 32 pages, 22 figures | null | null | null | cs.LG stat.ML | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | The general perception is that kernel methods are not scalable, and neural
nets are the methods of choice for nonlinear learning problems. Or have we
simply not tried hard enough for kernel methods? Here we propose an approach
that scales up kernel methods using a novel concept called "doubly stochastic
functional gradients". Our approach relies on the fact that many kernel methods
can be expressed as convex optimization problems, and we solve the problems by
making two unbiased stochastic approximations to the functional gradient, one
using random training points and another using random functions associated with
the kernel, and then descending using this noisy functional gradient. We show
that a function produced by this procedure after $t$ iterations converges to
the optimal function in the reproducing kernel Hilbert space in rate $O(1/t)$,
and achieves a generalization performance of $O(1/\sqrt{t})$. This doubly
stochasticity also allows us to avoid keeping the support vectors and to
implement the algorithm in a small memory footprint, which is linear in number
of iterations and independent of data dimension. Our approach can readily scale
kernel methods up to the regimes which are dominated by neural nets. We show
that our method can achieve competitive performance to neural nets in datasets
such as 8 million handwritten digits from MNIST, 2.3 million energy materials
from MolecularSpace, and 1 million photos from ImageNet.
| [
{
"version": "v1",
"created": "Mon, 21 Jul 2014 19:05:47 GMT"
},
{
"version": "v2",
"created": "Tue, 5 Aug 2014 17:58:57 GMT"
},
{
"version": "v3",
"created": "Tue, 23 Sep 2014 15:39:03 GMT"
},
{
"version": "v4",
"created": "Thu, 10 Sep 2015 16:40:45 GMT"
}
] | 2015-09-11T00:00:00 | [
[
"Dai",
"Bo",
""
],
[
"Xie",
"Bo",
""
],
[
"He",
"Niao",
""
],
[
"Liang",
"Yingyu",
""
],
[
"Raj",
"Anant",
""
],
[
"Balcan",
"Maria-Florina",
""
],
[
"Song",
"Le",
""
]
] | TITLE: Scalable Kernel Methods via Doubly Stochastic Gradients
ABSTRACT: The general perception is that kernel methods are not scalable, and neural
nets are the methods of choice for nonlinear learning problems. Or have we
simply not tried hard enough for kernel methods? Here we propose an approach
that scales up kernel methods using a novel concept called "doubly stochastic
functional gradients". Our approach relies on the fact that many kernel methods
can be expressed as convex optimization problems, and we solve the problems by
making two unbiased stochastic approximations to the functional gradient, one
using random training points and another using random functions associated with
the kernel, and then descending using this noisy functional gradient. We show
that a function produced by this procedure after $t$ iterations converges to
the optimal function in the reproducing kernel Hilbert space in rate $O(1/t)$,
and achieves a generalization performance of $O(1/\sqrt{t})$. This doubly
stochasticity also allows us to avoid keeping the support vectors and to
implement the algorithm in a small memory footprint, which is linear in number
of iterations and independent of data dimension. Our approach can readily scale
kernel methods up to the regimes which are dominated by neural nets. We show
that our method can achieve competitive performance to neural nets in datasets
such as 8 million handwritten digits from MNIST, 2.3 million energy materials
from MolecularSpace, and 1 million photos from ImageNet.
| no_new_dataset | 0.946498 |
1505.05253 | Jun Feng | Jun Feng, Mantong Zhou, Yu Hao, Minlie Huang and Xiaoyan Zhu | Knowlege Graph Embedding by Flexible Translation | This paper has been withdraw by the author due to an error in sec3.1 | null | null | null | cs.CL | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | Knowledge graph embedding refers to projecting entities and relations in
knowledge graph into continuous vector spaces. State-of-the-art methods, such
as TransE, TransH, and TransR build embeddings by treating relation as
translation from head entity to tail entity. However, previous models can not
deal with reflexive/one-to-many/many-to-one/many-to-many relations properly, or
lack of scalability and efficiency. Thus, we propose a novel method, flexible
translation, named TransF, to address the above issues. TransF regards relation
as translation between head entity vector and tail entity vector with flexible
magnitude. To evaluate the proposed model, we conduct link prediction and
triple classification on benchmark datasets. Experimental results show that our
method remarkably improve the performance compared with several
state-of-the-art baselines.
| [
{
"version": "v1",
"created": "Wed, 20 May 2015 05:57:32 GMT"
},
{
"version": "v2",
"created": "Thu, 10 Sep 2015 03:48:55 GMT"
}
] | 2015-09-11T00:00:00 | [
[
"Feng",
"Jun",
""
],
[
"Zhou",
"Mantong",
""
],
[
"Hao",
"Yu",
""
],
[
"Huang",
"Minlie",
""
],
[
"Zhu",
"Xiaoyan",
""
]
] | TITLE: Knowlege Graph Embedding by Flexible Translation
ABSTRACT: Knowledge graph embedding refers to projecting entities and relations in
knowledge graph into continuous vector spaces. State-of-the-art methods, such
as TransE, TransH, and TransR build embeddings by treating relation as
translation from head entity to tail entity. However, previous models can not
deal with reflexive/one-to-many/many-to-one/many-to-many relations properly, or
lack of scalability and efficiency. Thus, we propose a novel method, flexible
translation, named TransF, to address the above issues. TransF regards relation
as translation between head entity vector and tail entity vector with flexible
magnitude. To evaluate the proposed model, we conduct link prediction and
triple classification on benchmark datasets. Experimental results show that our
method remarkably improve the performance compared with several
state-of-the-art baselines.
| no_new_dataset | 0.947672 |
1509.02954 | Joseph Wang | Joseph Wang, Kirill Trapeznikov, Venkatesh Saligrama | Sensor Selection by Linear Programming | null | null | null | null | stat.ML cs.LG | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | We learn sensor trees from training data to minimize sensor acquisition costs
during test time. Our system adaptively selects sensors at each stage if
necessary to make a confident classification. We pose the problem as empirical
risk minimization over the choice of trees and node decision rules. We
decompose the problem, which is known to be intractable, into combinatorial
(tree structures) and continuous parts (node decision rules) and propose to
solve them separately. Using training data we greedily solve for the
combinatorial tree structures and for the continuous part, which is a
non-convex multilinear objective function, we derive convex surrogate loss
functions that are piecewise linear. The resulting problem can be cast as a
linear program and has the advantage of guaranteed convergence, global
optimality, repeatability and computational efficiency. We show that our
proposed approach outperforms the state-of-art on a number of benchmark
datasets.
| [
{
"version": "v1",
"created": "Wed, 9 Sep 2015 21:15:32 GMT"
}
] | 2015-09-11T00:00:00 | [
[
"Wang",
"Joseph",
""
],
[
"Trapeznikov",
"Kirill",
""
],
[
"Saligrama",
"Venkatesh",
""
]
] | TITLE: Sensor Selection by Linear Programming
ABSTRACT: We learn sensor trees from training data to minimize sensor acquisition costs
during test time. Our system adaptively selects sensors at each stage if
necessary to make a confident classification. We pose the problem as empirical
risk minimization over the choice of trees and node decision rules. We
decompose the problem, which is known to be intractable, into combinatorial
(tree structures) and continuous parts (node decision rules) and propose to
solve them separately. Using training data we greedily solve for the
combinatorial tree structures and for the continuous part, which is a
non-convex multilinear objective function, we derive convex surrogate loss
functions that are piecewise linear. The resulting problem can be cast as a
linear program and has the advantage of guaranteed convergence, global
optimality, repeatability and computational efficiency. We show that our
proposed approach outperforms the state-of-art on a number of benchmark
datasets.
| no_new_dataset | 0.948251 |
1509.03005 | David Balduzzi | David Balduzzi, Muhammad Ghifary | Compatible Value Gradients for Reinforcement Learning of Continuous Deep
Policies | 27 pages | null | null | null | cs.LG cs.AI cs.NE stat.ML | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | This paper proposes GProp, a deep reinforcement learning algorithm for
continuous policies with compatible function approximation. The algorithm is
based on two innovations. Firstly, we present a temporal-difference based
method for learning the gradient of the value-function. Secondly, we present
the deviator-actor-critic (DAC) model, which comprises three neural networks
that estimate the value function, its gradient, and determine the actor's
policy respectively. We evaluate GProp on two challenging tasks: a contextual
bandit problem constructed from nonparametric regression datasets that is
designed to probe the ability of reinforcement learning algorithms to
accurately estimate gradients; and the octopus arm, a challenging reinforcement
learning benchmark. GProp is competitive with fully supervised methods on the
bandit task and achieves the best performance to date on the octopus arm.
| [
{
"version": "v1",
"created": "Thu, 10 Sep 2015 04:14:54 GMT"
}
] | 2015-09-11T00:00:00 | [
[
"Balduzzi",
"David",
""
],
[
"Ghifary",
"Muhammad",
""
]
] | TITLE: Compatible Value Gradients for Reinforcement Learning of Continuous Deep
Policies
ABSTRACT: This paper proposes GProp, a deep reinforcement learning algorithm for
continuous policies with compatible function approximation. The algorithm is
based on two innovations. Firstly, we present a temporal-difference based
method for learning the gradient of the value-function. Secondly, we present
the deviator-actor-critic (DAC) model, which comprises three neural networks
that estimate the value function, its gradient, and determine the actor's
policy respectively. We evaluate GProp on two challenging tasks: a contextual
bandit problem constructed from nonparametric regression datasets that is
designed to probe the ability of reinforcement learning algorithms to
accurately estimate gradients; and the octopus arm, a challenging reinforcement
learning benchmark. GProp is competitive with fully supervised methods on the
bandit task and achieves the best performance to date on the octopus arm.
| no_new_dataset | 0.944074 |
1509.03247 | Arindam Chaudhuri AC | Arindam Chaudhuri | An Epsilon Hierarchical Fuzzy Twin Support Vector Regression | Research work at Samsung Research and Development Institute Delhi | null | null | null | cs.AI | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | The research presents epsilon hierarchical fuzzy twin support vector
regression based on epsilon fuzzy twin support vector regression and epsilon
twin support vector regression. Epsilon FTSVR is achieved by incorporating
trapezoidal fuzzy numbers to epsilon TSVR which takes care of uncertainty
existing in forecasting problems. Epsilon FTSVR determines a pair of epsilon
insensitive proximal functions by solving two related quadratic programming
problems. The structural risk minimization principle is implemented by
introducing regularization term in primal problems of epsilon FTSVR. This
yields dual stable positive definite problems which improves regression
performance. Epsilon FTSVR is then reformulated as epsilon HFTSVR consisting of
a set of hierarchical layers each containing epsilon FTSVR. Experimental
results on both synthetic and real datasets reveal that epsilon HFTSVR has
remarkable generalization performance with minimum training time.
| [
{
"version": "v1",
"created": "Thu, 10 Sep 2015 17:37:20 GMT"
}
] | 2015-09-11T00:00:00 | [
[
"Chaudhuri",
"Arindam",
""
]
] | TITLE: An Epsilon Hierarchical Fuzzy Twin Support Vector Regression
ABSTRACT: The research presents epsilon hierarchical fuzzy twin support vector
regression based on epsilon fuzzy twin support vector regression and epsilon
twin support vector regression. Epsilon FTSVR is achieved by incorporating
trapezoidal fuzzy numbers to epsilon TSVR which takes care of uncertainty
existing in forecasting problems. Epsilon FTSVR determines a pair of epsilon
insensitive proximal functions by solving two related quadratic programming
problems. The structural risk minimization principle is implemented by
introducing regularization term in primal problems of epsilon FTSVR. This
yields dual stable positive definite problems which improves regression
performance. Epsilon FTSVR is then reformulated as epsilon HFTSVR consisting of
a set of hierarchical layers each containing epsilon FTSVR. Experimental
results on both synthetic and real datasets reveal that epsilon HFTSVR has
remarkable generalization performance with minimum training time.
| no_new_dataset | 0.948346 |
1509.03248 | George Trigeorgis | George Trigeorgis, Konstantinos Bousmalis, Stefanos Zafeiriou, Bjoern
W.Schuller | A deep matrix factorization method for learning attribute
representations | Submitted to TPAMI (16-Mar-2015) | null | null | null | cs.CV cs.LG stat.ML | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | Semi-Non-negative Matrix Factorization is a technique that learns a
low-dimensional representation of a dataset that lends itself to a clustering
interpretation. It is possible that the mapping between this new representation
and our original data matrix contains rather complex hierarchical information
with implicit lower-level hidden attributes, that classical one level
clustering methodologies can not interpret. In this work we propose a novel
model, Deep Semi-NMF, that is able to learn such hidden representations that
allow themselves to an interpretation of clustering according to different,
unknown attributes of a given dataset. We also present a semi-supervised
version of the algorithm, named Deep WSF, that allows the use of (partial)
prior information for each of the known attributes of a dataset, that allows
the model to be used on datasets with mixed attribute knowledge. Finally, we
show that our models are able to learn low-dimensional representations that are
better suited for clustering, but also classification, outperforming
Semi-Non-negative Matrix Factorization, but also other state-of-the-art
methodologies variants.
| [
{
"version": "v1",
"created": "Thu, 10 Sep 2015 17:57:03 GMT"
}
] | 2015-09-11T00:00:00 | [
[
"Trigeorgis",
"George",
""
],
[
"Bousmalis",
"Konstantinos",
""
],
[
"Zafeiriou",
"Stefanos",
""
],
[
"Schuller",
"Bjoern W.",
""
]
] | TITLE: A deep matrix factorization method for learning attribute
representations
ABSTRACT: Semi-Non-negative Matrix Factorization is a technique that learns a
low-dimensional representation of a dataset that lends itself to a clustering
interpretation. It is possible that the mapping between this new representation
and our original data matrix contains rather complex hierarchical information
with implicit lower-level hidden attributes, that classical one level
clustering methodologies can not interpret. In this work we propose a novel
model, Deep Semi-NMF, that is able to learn such hidden representations that
allow themselves to an interpretation of clustering according to different,
unknown attributes of a given dataset. We also present a semi-supervised
version of the algorithm, named Deep WSF, that allows the use of (partial)
prior information for each of the known attributes of a dataset, that allows
the model to be used on datasets with mixed attribute knowledge. Finally, we
show that our models are able to learn low-dimensional representations that are
better suited for clustering, but also classification, outperforming
Semi-Non-negative Matrix Factorization, but also other state-of-the-art
methodologies variants.
| no_new_dataset | 0.947235 |
1509.02533 | Michael Mathioudakis | Charalampos Mavroforakis, Michael Mathioudakis and Aristides Gionis | Absorbing random-walk centrality: Theory and algorithms | 11 pages, 11 figures, short paper to appear at ICDM 2015 | null | null | null | cs.SI cs.DS | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | We study a new notion of graph centrality based on absorbing random walks.
Given a graph $G=(V,E)$ and a set of query nodes $Q\subseteq V$, we aim to
identify the $k$ most central nodes in $G$ with respect to $Q$. Specifically,
we consider central nodes to be absorbing for random walks that start at the
query nodes $Q$. The goal is to find the set of $k$ central nodes that
minimizes the expected length of a random walk until absorption. The proposed
measure, which we call $k$ absorbing random-walk centrality, favors diverse
sets, as it is beneficial to place the $k$ absorbing nodes in different parts
of the graph so as to "intercept" random walks that start from different query
nodes.
Although similar problem definitions have been considered in the literature,
e.g., in information-retrieval settings where the goal is to diversify
web-search results, in this paper we study the problem formally and prove some
of its properties. We show that the problem is NP-hard, while the objective
function is monotone and supermodular, implying that a greedy algorithm
provides solutions with an approximation guarantee. On the other hand, the
greedy algorithm involves expensive matrix operations that make it prohibitive
to employ on large datasets. To confront this challenge, we develop more
efficient algorithms based on spectral clustering and on personalized PageRank.
| [
{
"version": "v1",
"created": "Tue, 8 Sep 2015 20:10:04 GMT"
}
] | 2015-09-10T00:00:00 | [
[
"Mavroforakis",
"Charalampos",
""
],
[
"Mathioudakis",
"Michael",
""
],
[
"Gionis",
"Aristides",
""
]
] | TITLE: Absorbing random-walk centrality: Theory and algorithms
ABSTRACT: We study a new notion of graph centrality based on absorbing random walks.
Given a graph $G=(V,E)$ and a set of query nodes $Q\subseteq V$, we aim to
identify the $k$ most central nodes in $G$ with respect to $Q$. Specifically,
we consider central nodes to be absorbing for random walks that start at the
query nodes $Q$. The goal is to find the set of $k$ central nodes that
minimizes the expected length of a random walk until absorption. The proposed
measure, which we call $k$ absorbing random-walk centrality, favors diverse
sets, as it is beneficial to place the $k$ absorbing nodes in different parts
of the graph so as to "intercept" random walks that start from different query
nodes.
Although similar problem definitions have been considered in the literature,
e.g., in information-retrieval settings where the goal is to diversify
web-search results, in this paper we study the problem formally and prove some
of its properties. We show that the problem is NP-hard, while the objective
function is monotone and supermodular, implying that a greedy algorithm
provides solutions with an approximation guarantee. On the other hand, the
greedy algorithm involves expensive matrix operations that make it prohibitive
to employ on large datasets. To confront this challenge, we develop more
efficient algorithms based on spectral clustering and on personalized PageRank.
| no_new_dataset | 0.945096 |
1407.0623 | Lamberto Ballan | Lamberto Ballan, Marco Bertini, Giuseppe Serra, Alberto Del Bimbo | A Data-Driven Approach for Tag Refinement and Localization in Web Videos | Preprint submitted to Computer Vision and Image Understanding (CVIU) | null | 10.1016/j.cviu.2015.05.009 | null | cs.CV cs.IR cs.MM | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | Tagging of visual content is becoming more and more widespread as web-based
services and social networks have popularized tagging functionalities among
their users. These user-generated tags are used to ease browsing and
exploration of media collections, e.g. using tag clouds, or to retrieve
multimedia content. However, not all media are equally tagged by users. Using
the current systems is easy to tag a single photo, and even tagging a part of a
photo, like a face, has become common in sites like Flickr and Facebook. On the
other hand, tagging a video sequence is more complicated and time consuming, so
that users just tag the overall content of a video. In this paper we present a
method for automatic video annotation that increases the number of tags
originally provided by users, and localizes them temporally, associating tags
to keyframes. Our approach exploits collective knowledge embedded in
user-generated tags and web sources, and visual similarity of keyframes and
images uploaded to social sites like YouTube and Flickr, as well as web sources
like Google and Bing. Given a keyframe, our method is able to select on the fly
from these visual sources the training exemplars that should be the most
relevant for this test sample, and proceeds to transfer labels across similar
images. Compared to existing video tagging approaches that require training
classifiers for each tag, our system has few parameters, is easy to implement
and can deal with an open vocabulary scenario. We demonstrate the approach on
tag refinement and localization on DUT-WEBV, a large dataset of web videos, and
show state-of-the-art results.
| [
{
"version": "v1",
"created": "Wed, 2 Jul 2014 15:48:37 GMT"
},
{
"version": "v2",
"created": "Sat, 11 Apr 2015 18:12:36 GMT"
},
{
"version": "v3",
"created": "Thu, 28 May 2015 17:12:54 GMT"
}
] | 2015-09-09T00:00:00 | [
[
"Ballan",
"Lamberto",
""
],
[
"Bertini",
"Marco",
""
],
[
"Serra",
"Giuseppe",
""
],
[
"Del Bimbo",
"Alberto",
""
]
] | TITLE: A Data-Driven Approach for Tag Refinement and Localization in Web Videos
ABSTRACT: Tagging of visual content is becoming more and more widespread as web-based
services and social networks have popularized tagging functionalities among
their users. These user-generated tags are used to ease browsing and
exploration of media collections, e.g. using tag clouds, or to retrieve
multimedia content. However, not all media are equally tagged by users. Using
the current systems is easy to tag a single photo, and even tagging a part of a
photo, like a face, has become common in sites like Flickr and Facebook. On the
other hand, tagging a video sequence is more complicated and time consuming, so
that users just tag the overall content of a video. In this paper we present a
method for automatic video annotation that increases the number of tags
originally provided by users, and localizes them temporally, associating tags
to keyframes. Our approach exploits collective knowledge embedded in
user-generated tags and web sources, and visual similarity of keyframes and
images uploaded to social sites like YouTube and Flickr, as well as web sources
like Google and Bing. Given a keyframe, our method is able to select on the fly
from these visual sources the training exemplars that should be the most
relevant for this test sample, and proceeds to transfer labels across similar
images. Compared to existing video tagging approaches that require training
classifiers for each tag, our system has few parameters, is easy to implement
and can deal with an open vocabulary scenario. We demonstrate the approach on
tag refinement and localization on DUT-WEBV, a large dataset of web videos, and
show state-of-the-art results.
| no_new_dataset | 0.948728 |
1412.5732 | Changsheng Li | Changsheng Li and Fan Wei and Weishan Dong and Qingshan Liu and
Xiangfeng Wang and Xin Zhang | Dynamic Structure Embedded Online Multiple-Output Regression for Stream
Data | null | null | null | null | cs.LG | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | Online multiple-output regression is an important machine learning technique
for modeling, predicting, and compressing multi-dimensional correlated data
streams. In this paper, we propose a novel online multiple-output regression
method, called MORES, for stream data. MORES can \emph{dynamically} learn the
structure of the coefficients change in each update step to facilitate the
model's continuous refinement. We observe that limited expressive ability of
the regression model, especially in the preliminary stage of online update,
often leads to the variables in the residual errors being dependent. In light
of this point, MORES intends to \emph{dynamically} learn and leverage the
structure of the residual errors to improve the prediction accuracy. Moreover,
we define three statistical variables to \emph{exactly} represent all the seen
samples for \emph{incrementally} calculating prediction loss in each online
update round, which can avoid loading all the training data into memory for
updating model, and also effectively prevent drastic fluctuation of the model
in the presence of noise. Furthermore, we introduce a forgetting factor to set
different weights on samples so as to track the data streams' evolving
characteristics quickly from the latest samples. Experiments on one synthetic
dataset and three real-world datasets validate the effectiveness of the
proposed method. In addition, the update speed of MORES is at least 2000
samples processed per second on the three real-world datasets, more than 15
times faster than the state-of-the-art online learning algorithm.
| [
{
"version": "v1",
"created": "Thu, 18 Dec 2014 06:37:50 GMT"
},
{
"version": "v2",
"created": "Tue, 8 Sep 2015 03:00:55 GMT"
}
] | 2015-09-09T00:00:00 | [
[
"Li",
"Changsheng",
""
],
[
"Wei",
"Fan",
""
],
[
"Dong",
"Weishan",
""
],
[
"Liu",
"Qingshan",
""
],
[
"Wang",
"Xiangfeng",
""
],
[
"Zhang",
"Xin",
""
]
] | TITLE: Dynamic Structure Embedded Online Multiple-Output Regression for Stream
Data
ABSTRACT: Online multiple-output regression is an important machine learning technique
for modeling, predicting, and compressing multi-dimensional correlated data
streams. In this paper, we propose a novel online multiple-output regression
method, called MORES, for stream data. MORES can \emph{dynamically} learn the
structure of the coefficients change in each update step to facilitate the
model's continuous refinement. We observe that limited expressive ability of
the regression model, especially in the preliminary stage of online update,
often leads to the variables in the residual errors being dependent. In light
of this point, MORES intends to \emph{dynamically} learn and leverage the
structure of the residual errors to improve the prediction accuracy. Moreover,
we define three statistical variables to \emph{exactly} represent all the seen
samples for \emph{incrementally} calculating prediction loss in each online
update round, which can avoid loading all the training data into memory for
updating model, and also effectively prevent drastic fluctuation of the model
in the presence of noise. Furthermore, we introduce a forgetting factor to set
different weights on samples so as to track the data streams' evolving
characteristics quickly from the latest samples. Experiments on one synthetic
dataset and three real-world datasets validate the effectiveness of the
proposed method. In addition, the update speed of MORES is at least 2000
samples processed per second on the three real-world datasets, more than 15
times faster than the state-of-the-art online learning algorithm.
| no_new_dataset | 0.949902 |
1501.07492 | Huaizu Jiang | Huaizu Jiang | Weakly Supervised Learning for Salient Object Detection | technical report | null | null | null | cs.CV | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | Recent advances in supervised salient object detection has resulted in
significant performance on benchmark datasets. Training such models, however,
requires expensive pixel-wise annotations of salient objects. Moreover, many
existing salient object detection models assume that at least one salient
object exists in the input image. Such an assumption often leads to less
appealing saliency maps on the background images, which contain no salient
object at all. To avoid the requirement of expensive pixel-wise salient region
annotations, in this paper, we study weakly supervised learning approaches for
salient object detection. Given a set of background images and salient object
images, we propose a solution toward jointly addressing the salient object
existence and detection tasks. We adopt the latent SVM framework and formulate
the two problems together in a single integrated objective function: saliency
labels of superpixels are modeled as hidden variables and involved in a
classification term conditioned to the salient object existence variable, which
in turn depends on both global image and regional saliency features and
saliency label assignment. Experimental results on benchmark datasets validate
the effectiveness of our proposed approach.
| [
{
"version": "v1",
"created": "Thu, 29 Jan 2015 15:57:52 GMT"
},
{
"version": "v2",
"created": "Tue, 8 Sep 2015 13:34:24 GMT"
}
] | 2015-09-09T00:00:00 | [
[
"Jiang",
"Huaizu",
""
]
] | TITLE: Weakly Supervised Learning for Salient Object Detection
ABSTRACT: Recent advances in supervised salient object detection has resulted in
significant performance on benchmark datasets. Training such models, however,
requires expensive pixel-wise annotations of salient objects. Moreover, many
existing salient object detection models assume that at least one salient
object exists in the input image. Such an assumption often leads to less
appealing saliency maps on the background images, which contain no salient
object at all. To avoid the requirement of expensive pixel-wise salient region
annotations, in this paper, we study weakly supervised learning approaches for
salient object detection. Given a set of background images and salient object
images, we propose a solution toward jointly addressing the salient object
existence and detection tasks. We adopt the latent SVM framework and formulate
the two problems together in a single integrated objective function: saliency
labels of superpixels are modeled as hidden variables and involved in a
classification term conditioned to the salient object existence variable, which
in turn depends on both global image and regional saliency features and
saliency label assignment. Experimental results on benchmark datasets validate
the effectiveness of our proposed approach.
| no_new_dataset | 0.951051 |
1506.02108 | Chunhua Shen | Guosheng Lin, Chunhua Shen, Ian Reid, Anton van den Hengel | Deeply Learning the Messages in Message Passing Inference | 11 pages. Appearing in Proc. The Twenty-ninth Annual Conference on
Neural Information Processing Systems (NIPS), 2015, Montreal, Canada | null | null | null | cs.CV cs.LG stat.ML | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | Deep structured output learning shows great promise in tasks like semantic
image segmentation. We proffer a new, efficient deep structured model learning
scheme, in which we show how deep Convolutional Neural Networks (CNNs) can be
used to estimate the messages in message passing inference for structured
prediction with Conditional Random Fields (CRFs). With such CNN message
estimators, we obviate the need to learn or evaluate potential functions for
message calculation. This confers significant efficiency for learning, since
otherwise when performing structured learning for a CRF with CNN potentials it
is necessary to undertake expensive inference for every stochastic gradient
iteration. The network output dimension for message estimation is the same as
the number of classes, in contrast to the network output for general CNN
potential functions in CRFs, which is exponential in the order of the
potentials. Hence CNN message learning has fewer network parameters and is more
scalable for cases that a large number of classes are involved. We apply our
method to semantic image segmentation on the PASCAL VOC 2012 dataset. We
achieve an intersection-over-union score of 73.4 on its test set, which is the
best reported result for methods using the VOC training images alone. This
impressive performance demonstrates the effectiveness and usefulness of our CNN
message learning method.
| [
{
"version": "v1",
"created": "Sat, 6 Jun 2015 02:52:38 GMT"
},
{
"version": "v2",
"created": "Wed, 10 Jun 2015 06:49:06 GMT"
},
{
"version": "v3",
"created": "Tue, 8 Sep 2015 04:29:45 GMT"
}
] | 2015-09-09T00:00:00 | [
[
"Lin",
"Guosheng",
""
],
[
"Shen",
"Chunhua",
""
],
[
"Reid",
"Ian",
""
],
[
"Hengel",
"Anton van den",
""
]
] | TITLE: Deeply Learning the Messages in Message Passing Inference
ABSTRACT: Deep structured output learning shows great promise in tasks like semantic
image segmentation. We proffer a new, efficient deep structured model learning
scheme, in which we show how deep Convolutional Neural Networks (CNNs) can be
used to estimate the messages in message passing inference for structured
prediction with Conditional Random Fields (CRFs). With such CNN message
estimators, we obviate the need to learn or evaluate potential functions for
message calculation. This confers significant efficiency for learning, since
otherwise when performing structured learning for a CRF with CNN potentials it
is necessary to undertake expensive inference for every stochastic gradient
iteration. The network output dimension for message estimation is the same as
the number of classes, in contrast to the network output for general CNN
potential functions in CRFs, which is exponential in the order of the
potentials. Hence CNN message learning has fewer network parameters and is more
scalable for cases that a large number of classes are involved. We apply our
method to semantic image segmentation on the PASCAL VOC 2012 dataset. We
achieve an intersection-over-union score of 73.4 on its test set, which is the
best reported result for methods using the VOC training images alone. This
impressive performance demonstrates the effectiveness and usefulness of our CNN
message learning method.
| no_new_dataset | 0.950041 |
1509.02441 | Subarna Tripathi | Subarna Tripathi, Serge Belongie, Youngbae Hwang, Truong Nguyen | Semantic Video Segmentation : Exploring Inference Efficiency | To appear in proc of ISOCC 2015 | null | null | null | cs.CV | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | We explore the efficiency of the CRF inference beyond image level semantic
segmentation and perform joint inference in video frames. The key idea is to
combine best of two worlds: semantic co-labeling and more expressive models.
Our formulation enables us to perform inference over ten thousand images within
seconds and makes the system amenable to perform video semantic segmentation
most effectively. On CamVid dataset, with TextonBoost unaries, our proposed
method achieves up to 8% improvement in accuracy over individual semantic image
segmentation without additional time overhead. The source code is available at
https://github.com/subtri/video_inference
| [
{
"version": "v1",
"created": "Fri, 4 Sep 2015 22:03:40 GMT"
}
] | 2015-09-09T00:00:00 | [
[
"Tripathi",
"Subarna",
""
],
[
"Belongie",
"Serge",
""
],
[
"Hwang",
"Youngbae",
""
],
[
"Nguyen",
"Truong",
""
]
] | TITLE: Semantic Video Segmentation : Exploring Inference Efficiency
ABSTRACT: We explore the efficiency of the CRF inference beyond image level semantic
segmentation and perform joint inference in video frames. The key idea is to
combine best of two worlds: semantic co-labeling and more expressive models.
Our formulation enables us to perform inference over ten thousand images within
seconds and makes the system amenable to perform video semantic segmentation
most effectively. On CamVid dataset, with TextonBoost unaries, our proposed
method achieves up to 8% improvement in accuracy over individual semantic image
segmentation without additional time overhead. The source code is available at
https://github.com/subtri/video_inference
| no_new_dataset | 0.950503 |
1408.1656 | Shengcai Liao | Shengcai Liao, Anil K. Jain, and Stan Z. Li | A Fast and Accurate Unconstrained Face Detector | This paper has been accepted by TPAMI. The source code is available
on the project page
http://www.cbsr.ia.ac.cn/users/scliao/projects/npdface/index.html | null | 10.1109/TPAMI.2015.2448075 | null | cs.CV | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | We propose a method to address challenges in unconstrained face detection,
such as arbitrary pose variations and occlusions. First, a new image feature
called Normalized Pixel Difference (NPD) is proposed. NPD feature is computed
as the difference to sum ratio between two pixel values, inspired by the Weber
Fraction in experimental psychology. The new feature is scale invariant,
bounded, and is able to reconstruct the original image. Second, we propose a
deep quadratic tree to learn the optimal subset of NPD features and their
combinations, so that complex face manifolds can be partitioned by the learned
rules. This way, only a single soft-cascade classifier is needed to handle
unconstrained face detection. Furthermore, we show that the NPD features can be
efficiently obtained from a look up table, and the detection template can be
easily scaled, making the proposed face detector very fast. Experimental
results on three public face datasets (FDDB, GENKI, and CMU-MIT) show that the
proposed method achieves state-of-the-art performance in detecting
unconstrained faces with arbitrary pose variations and occlusions in cluttered
scenes.
| [
{
"version": "v1",
"created": "Wed, 6 Aug 2014 15:17:33 GMT"
},
{
"version": "v2",
"created": "Tue, 12 Aug 2014 14:24:52 GMT"
},
{
"version": "v3",
"created": "Mon, 7 Sep 2015 08:17:34 GMT"
}
] | 2015-09-08T00:00:00 | [
[
"Liao",
"Shengcai",
""
],
[
"Jain",
"Anil K.",
""
],
[
"Li",
"Stan Z.",
""
]
] | TITLE: A Fast and Accurate Unconstrained Face Detector
ABSTRACT: We propose a method to address challenges in unconstrained face detection,
such as arbitrary pose variations and occlusions. First, a new image feature
called Normalized Pixel Difference (NPD) is proposed. NPD feature is computed
as the difference to sum ratio between two pixel values, inspired by the Weber
Fraction in experimental psychology. The new feature is scale invariant,
bounded, and is able to reconstruct the original image. Second, we propose a
deep quadratic tree to learn the optimal subset of NPD features and their
combinations, so that complex face manifolds can be partitioned by the learned
rules. This way, only a single soft-cascade classifier is needed to handle
unconstrained face detection. Furthermore, we show that the NPD features can be
efficiently obtained from a look up table, and the detection template can be
easily scaled, making the proposed face detector very fast. Experimental
results on three public face datasets (FDDB, GENKI, and CMU-MIT) show that the
proposed method achieves state-of-the-art performance in detecting
unconstrained faces with arbitrary pose variations and occlusions in cluttered
scenes.
| no_new_dataset | 0.946547 |
1410.4449 | Alina S\^irbu | Alina S\^irbu, Ozalp Babaoglu | A Holistic Approach to Log Data Analysis in High-Performance Computing
Systems: The Case of IBM Blue Gene/Q | 12 pages, 7 Figures | null | null | null | cs.DC | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | The complexity and cost of managing high-performance computing
infrastructures are on the rise. Automating management and repair through
predictive models to minimize human interventions is an attempt to increase
system availability and contain these costs. Building predictive models that
are accurate enough to be useful in automatic management cannot be based on
restricted log data from subsystems but requires a holistic approach to data
analysis from disparate sources. Here we provide a detailed multi-scale
characterization study based on four datasets reporting power consumption,
temperature, workload, and hardware/software events for an IBM Blue Gene/Q
installation. We show that the system runs a rich parallel workload, with low
correlation among its components in terms of temperature and power, but higher
correlation in terms of events. As expected, power and temperature correlate
strongly, while events display negative correlations with load and power. Power
and workload show moderate correlations, and only at the scale of components.
The aim of the study is a systematic, integrated characterization of the
computing infrastructure and discovery of correlation sources and levels to
serve as basis for future predictive modeling efforts.
| [
{
"version": "v1",
"created": "Thu, 16 Oct 2014 14:40:00 GMT"
},
{
"version": "v2",
"created": "Tue, 10 Feb 2015 10:41:57 GMT"
},
{
"version": "v3",
"created": "Mon, 7 Sep 2015 11:08:50 GMT"
}
] | 2015-09-08T00:00:00 | [
[
"Sîrbu",
"Alina",
""
],
[
"Babaoglu",
"Ozalp",
""
]
] | TITLE: A Holistic Approach to Log Data Analysis in High-Performance Computing
Systems: The Case of IBM Blue Gene/Q
ABSTRACT: The complexity and cost of managing high-performance computing
infrastructures are on the rise. Automating management and repair through
predictive models to minimize human interventions is an attempt to increase
system availability and contain these costs. Building predictive models that
are accurate enough to be useful in automatic management cannot be based on
restricted log data from subsystems but requires a holistic approach to data
analysis from disparate sources. Here we provide a detailed multi-scale
characterization study based on four datasets reporting power consumption,
temperature, workload, and hardware/software events for an IBM Blue Gene/Q
installation. We show that the system runs a rich parallel workload, with low
correlation among its components in terms of temperature and power, but higher
correlation in terms of events. As expected, power and temperature correlate
strongly, while events display negative correlations with load and power. Power
and workload show moderate correlations, and only at the scale of components.
The aim of the study is a systematic, integrated characterization of the
computing infrastructure and discovery of correlation sources and levels to
serve as basis for future predictive modeling efforts.
| no_new_dataset | 0.934574 |
1505.02108 | Ira Kemelmacher-Shlizerman | D. Miller, E. Brossard, S. Seitz, I. Kemelmacher-Shlizerman | MegaFace: A Million Faces for Recognition at Scale | Please see http://megaface.cs.washington.edu/ for code and data | null | null | null | cs.CV | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | Recent face recognition experiments on the LFW benchmark show that face
recognition is performing stunningly well, surpassing human recognition rates.
In this paper, we study face recognition at scale. Specifically, we have
collected from Flickr a \textbf{Million} faces and evaluated state of the art
face recognition algorithms on this dataset. We found that the performance of
algorithms varies--while all perform great on LFW, once evaluated at scale
recognition rates drop drastically for most algorithms. Interestingly, deep
learning based approach by \cite{schroff2015facenet} performs much better, but
still gets less robust at scale. We consider both verification and
identification problems, and evaluate how pose affects recognition at scale.
Moreover, we ran an extensive human study on Mechanical Turk to evaluate human
recognition at scale, and report results. All the photos are creative commons
photos and is released at \small{\url{http://megaface.cs.washington.edu/}} for
research and further experiments.
| [
{
"version": "v1",
"created": "Fri, 8 May 2015 17:39:23 GMT"
},
{
"version": "v2",
"created": "Mon, 7 Sep 2015 19:45:47 GMT"
}
] | 2015-09-08T00:00:00 | [
[
"Miller",
"D.",
""
],
[
"Brossard",
"E.",
""
],
[
"Seitz",
"S.",
""
],
[
"Kemelmacher-Shlizerman",
"I.",
""
]
] | TITLE: MegaFace: A Million Faces for Recognition at Scale
ABSTRACT: Recent face recognition experiments on the LFW benchmark show that face
recognition is performing stunningly well, surpassing human recognition rates.
In this paper, we study face recognition at scale. Specifically, we have
collected from Flickr a \textbf{Million} faces and evaluated state of the art
face recognition algorithms on this dataset. We found that the performance of
algorithms varies--while all perform great on LFW, once evaluated at scale
recognition rates drop drastically for most algorithms. Interestingly, deep
learning based approach by \cite{schroff2015facenet} performs much better, but
still gets less robust at scale. We consider both verification and
identification problems, and evaluate how pose affects recognition at scale.
Moreover, we ran an extensive human study on Mechanical Turk to evaluate human
recognition at scale, and report results. All the photos are creative commons
photos and is released at \small{\url{http://megaface.cs.washington.edu/}} for
research and further experiments.
| no_new_dataset | 0.926901 |
1509.01602 | Ivan Bogun | Ivan Bogun, Anelia Angelova and Navdeep Jaitly | Object Recognition from Short Videos for Robotic Perception | 7 pages, 6 figures, 3 tables | null | null | null | cs.CV | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | Deep neural networks have become the primary learning technique for object
recognition. Videos, unlike still images, are temporally coherent which makes
the application of deep networks non-trivial. Here, we investigate how motion
can aid object recognition in short videos. Our approach is based on Long
Short-Term Memory (LSTM) deep networks. Unlike previous applications of LSTMs,
we implement each gate as a convolution. We show that convolutional-based LSTM
models are capable of learning motion dependencies and are able to improve the
recognition accuracy when more frames in a sequence are available. We evaluate
our approach on the Washington RGBD Object dataset and on the Washington RGBD
Scenes dataset. Our approach outperforms deep nets applied to still images and
sets a new state-of-the-art in this domain.
| [
{
"version": "v1",
"created": "Fri, 4 Sep 2015 20:48:23 GMT"
}
] | 2015-09-08T00:00:00 | [
[
"Bogun",
"Ivan",
""
],
[
"Angelova",
"Anelia",
""
],
[
"Jaitly",
"Navdeep",
""
]
] | TITLE: Object Recognition from Short Videos for Robotic Perception
ABSTRACT: Deep neural networks have become the primary learning technique for object
recognition. Videos, unlike still images, are temporally coherent which makes
the application of deep networks non-trivial. Here, we investigate how motion
can aid object recognition in short videos. Our approach is based on Long
Short-Term Memory (LSTM) deep networks. Unlike previous applications of LSTMs,
we implement each gate as a convolution. We show that convolutional-based LSTM
models are capable of learning motion dependencies and are able to improve the
recognition accuracy when more frames in a sequence are available. We evaluate
our approach on the Washington RGBD Object dataset and on the Washington RGBD
Scenes dataset. Our approach outperforms deep nets applied to still images and
sets a new state-of-the-art in this domain.
| no_new_dataset | 0.948346 |
1509.01659 | Armen Aghajanyan | Armen Aghajanyan | Gravitational Clustering | null | null | null | null | cs.LG | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | The downfall of many supervised learning algorithms, such as neural networks,
is the inherent need for a large amount of training data. Although there is a
lot of buzz about big data, there is still the problem of doing classification
from a small dataset. Other methods such as support vector machines, although
capable of dealing with few samples, are inherently binary classifiers, and are
in need of learning strategies such as One vs All in the case of
multi-classification. In the presence of a large number of classes this can
become problematic. In this paper we present, a novel approach to supervised
learning through the method of clustering. Unlike traditional methods such as
K-Means, Gravitational Clustering does not require the initial number of
clusters, and automatically builds the clusters, individual samples can be
arbitrarily weighted and it requires only few samples while staying resilient
to over-fitting.
| [
{
"version": "v1",
"created": "Sat, 5 Sep 2015 03:37:50 GMT"
}
] | 2015-09-08T00:00:00 | [
[
"Aghajanyan",
"Armen",
""
]
] | TITLE: Gravitational Clustering
ABSTRACT: The downfall of many supervised learning algorithms, such as neural networks,
is the inherent need for a large amount of training data. Although there is a
lot of buzz about big data, there is still the problem of doing classification
from a small dataset. Other methods such as support vector machines, although
capable of dealing with few samples, are inherently binary classifiers, and are
in need of learning strategies such as One vs All in the case of
multi-classification. In the presence of a large number of classes this can
become problematic. In this paper we present, a novel approach to supervised
learning through the method of clustering. Unlike traditional methods such as
K-Means, Gravitational Clustering does not require the initial number of
clusters, and automatically builds the clusters, individual samples can be
arbitrarily weighted and it requires only few samples while staying resilient
to over-fitting.
| no_new_dataset | 0.949856 |
1509.01719 | Yuewei Lin | Yuewei Lin, Jing Chen, Yu Cao, Youjie Zhou, Lingfeng Zhang, Yuan Yan
Tang, Song Wang | Unsupervised Cross-Domain Recognition by Identifying Compact Joint
Subspaces | ICIP 2015 Top 10% paper | null | null | null | cs.CV | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | This paper introduces a new method to solve the cross-domain recognition
problem. Different from the traditional domain adaption methods which rely on a
global domain shift for all classes between source and target domain, the
proposed method is more flexible to capture individual class variations across
domains. By adopting a natural and widely used assumption -- "the data samples
from the same class should lay on a low-dimensional subspace, even if they come
from different domains", the proposed method circumvents the limitation of the
global domain shift, and solves the cross-domain recognition by finding the
compact joint subspaces of source and target domain. Specifically, given
labeled samples in source domain, we construct subspaces for each of the
classes. Then we construct subspaces in the target domain, called anchor
subspaces, by collecting unlabeled samples that are close to each other and
highly likely all fall into the same class. The corresponding class label is
then assigned by minimizing a cost function which reflects the overlap and
topological structure consistency between subspaces across source and target
domains, and within anchor subspaces, respectively.We further combine the
anchor subspaces to corresponding source subspaces to construct the compact
joint subspaces. Subsequently, one-vs-rest SVM classifiers are trained in the
compact joint subspaces and applied to unlabeled data in the target domain. We
evaluate the proposed method on two widely used datasets: object recognition
dataset for computer vision tasks, and sentiment classification dataset for
natural language processing tasks. Comparison results demonstrate that the
proposed method outperforms the comparison methods on both datasets.
| [
{
"version": "v1",
"created": "Sat, 5 Sep 2015 17:12:21 GMT"
}
] | 2015-09-08T00:00:00 | [
[
"Lin",
"Yuewei",
""
],
[
"Chen",
"Jing",
""
],
[
"Cao",
"Yu",
""
],
[
"Zhou",
"Youjie",
""
],
[
"Zhang",
"Lingfeng",
""
],
[
"Tang",
"Yuan Yan",
""
],
[
"Wang",
"Song",
""
]
] | TITLE: Unsupervised Cross-Domain Recognition by Identifying Compact Joint
Subspaces
ABSTRACT: This paper introduces a new method to solve the cross-domain recognition
problem. Different from the traditional domain adaption methods which rely on a
global domain shift for all classes between source and target domain, the
proposed method is more flexible to capture individual class variations across
domains. By adopting a natural and widely used assumption -- "the data samples
from the same class should lay on a low-dimensional subspace, even if they come
from different domains", the proposed method circumvents the limitation of the
global domain shift, and solves the cross-domain recognition by finding the
compact joint subspaces of source and target domain. Specifically, given
labeled samples in source domain, we construct subspaces for each of the
classes. Then we construct subspaces in the target domain, called anchor
subspaces, by collecting unlabeled samples that are close to each other and
highly likely all fall into the same class. The corresponding class label is
then assigned by minimizing a cost function which reflects the overlap and
topological structure consistency between subspaces across source and target
domains, and within anchor subspaces, respectively.We further combine the
anchor subspaces to corresponding source subspaces to construct the compact
joint subspaces. Subsequently, one-vs-rest SVM classifiers are trained in the
compact joint subspaces and applied to unlabeled data in the target domain. We
evaluate the proposed method on two widely used datasets: object recognition
dataset for computer vision tasks, and sentiment classification dataset for
natural language processing tasks. Comparison results demonstrate that the
proposed method outperforms the comparison methods on both datasets.
| no_new_dataset | 0.949482 |
1509.02094 | Hyun Soo Park | Hyun Soo Park, Yedong Niu, Jianbo Shi | Future Localization from an Egocentric Depth Image | 9 pages | null | null | null | cs.CV | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | This paper presents a method for future localization: to predict a set of
plausible trajectories of ego-motion given a depth image. We predict paths
avoiding obstacles, between objects, even paths turning around a corner into
space behind objects. As a byproduct of the predicted trajectories of
ego-motion, we discover in the image the empty space occluded by foreground
objects. We use no image based features such as semantic labeling/segmentation
or object detection/recognition for this algorithm. Inspired by proxemics, we
represent the space around a person using an EgoSpace map, akin to an
illustrated tourist map, that measures a likelihood of occlusion at the
egocentric coordinate system. A future trajectory of ego-motion is modeled by a
linear combination of compact trajectory bases allowing us to constrain the
predicted trajectory. We learn the relationship between the EgoSpace map and
trajectory from the EgoMotion dataset providing in-situ measurements of the
future trajectory. A cost function that takes into account partial occlusion
due to foreground objects is minimized to predict a trajectory. This cost
function generates a trajectory that passes through the occluded space, which
allows us to discover the empty space behind the foreground objects. We
quantitatively evaluate our method to show predictive validity and apply to
various real world scenes including walking, shopping, and social interactions.
| [
{
"version": "v1",
"created": "Mon, 7 Sep 2015 15:51:11 GMT"
}
] | 2015-09-08T00:00:00 | [
[
"Park",
"Hyun Soo",
""
],
[
"Niu",
"Yedong",
""
],
[
"Shi",
"Jianbo",
""
]
] | TITLE: Future Localization from an Egocentric Depth Image
ABSTRACT: This paper presents a method for future localization: to predict a set of
plausible trajectories of ego-motion given a depth image. We predict paths
avoiding obstacles, between objects, even paths turning around a corner into
space behind objects. As a byproduct of the predicted trajectories of
ego-motion, we discover in the image the empty space occluded by foreground
objects. We use no image based features such as semantic labeling/segmentation
or object detection/recognition for this algorithm. Inspired by proxemics, we
represent the space around a person using an EgoSpace map, akin to an
illustrated tourist map, that measures a likelihood of occlusion at the
egocentric coordinate system. A future trajectory of ego-motion is modeled by a
linear combination of compact trajectory bases allowing us to constrain the
predicted trajectory. We learn the relationship between the EgoSpace map and
trajectory from the EgoMotion dataset providing in-situ measurements of the
future trajectory. A cost function that takes into account partial occlusion
due to foreground objects is minimized to predict a trajectory. This cost
function generates a trajectory that passes through the occluded space, which
allows us to discover the empty space behind the foreground objects. We
quantitatively evaluate our method to show predictive validity and apply to
various real world scenes including walking, shopping, and social interactions.
| no_new_dataset | 0.919643 |
1509.01354 | Jinma Guo | Jinma Guo and Jianmin Li | CNN Based Hashing for Image Retrieval | 16 pages, 6 figures | null | null | null | cs.CV cs.LG | http://creativecommons.org/publicdomain/zero/1.0/ | Along with data on the web increasing dramatically, hashing is becoming more
and more popular as a method of approximate nearest neighbor search. Previous
supervised hashing methods utilized similarity/dissimilarity matrix to get
semantic information. But the matrix is not easy to construct for a new
dataset. Rather than to reconstruct the matrix, we proposed a straightforward
CNN-based hashing method, i.e. binarilizing the activations of a fully
connected layer with threshold 0 and taking the binary result as hash codes.
This method achieved the best performance on CIFAR-10 and was comparable with
the state-of-the-art on MNIST. And our experiments on CIFAR-10 suggested that
the signs of activations may carry more information than the relative values of
activations between samples, and that the co-adaption between feature extractor
and hash functions is important for hashing.
| [
{
"version": "v1",
"created": "Fri, 4 Sep 2015 07:08:44 GMT"
}
] | 2015-09-07T00:00:00 | [
[
"Guo",
"Jinma",
""
],
[
"Li",
"Jianmin",
""
]
] | TITLE: CNN Based Hashing for Image Retrieval
ABSTRACT: Along with data on the web increasing dramatically, hashing is becoming more
and more popular as a method of approximate nearest neighbor search. Previous
supervised hashing methods utilized similarity/dissimilarity matrix to get
semantic information. But the matrix is not easy to construct for a new
dataset. Rather than to reconstruct the matrix, we proposed a straightforward
CNN-based hashing method, i.e. binarilizing the activations of a fully
connected layer with threshold 0 and taking the binary result as hash codes.
This method achieved the best performance on CIFAR-10 and was comparable with
the state-of-the-art on MNIST. And our experiments on CIFAR-10 suggested that
the signs of activations may carry more information than the relative values of
activations between samples, and that the co-adaption between feature extractor
and hash functions is important for hashing.
| no_new_dataset | 0.94868 |
1509.01379 | Balubaid Mohammed | Mohammed A. Balubaid and Umar Manzoor | Ontology Based SMS Controller for Smart Phones | null | null | null | null | cs.AI | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | Text analysis includes lexical analysis of the text and has been widely
studied and used in diverse applications. In the last decade, researchers have
proposed many efficient solutions to analyze / classify large text dataset,
however, analysis / classification of short text is still a challenge because
1) the data is very sparse 2) It contains noise words and 3) It is difficult to
understand the syntactical structure of the text. Short Messaging Service (SMS)
is a text messaging service for mobile/smart phone and this service is
frequently used by all mobile users. Because of the popularity of SMS service,
marketing companies nowadays are also using this service for direct marketing
also known as SMS marketing.In this paper, we have proposed Ontology based SMS
Controller which analyze the text message and classify it using ontology
aslegitimate or spam. The proposed system has been tested on different
scenarios and experimental results shows that the proposed solution is
effective both in terms of efficiency and time.
| [
{
"version": "v1",
"created": "Fri, 4 Sep 2015 09:29:47 GMT"
}
] | 2015-09-07T00:00:00 | [
[
"Balubaid",
"Mohammed A.",
""
],
[
"Manzoor",
"Umar",
""
]
] | TITLE: Ontology Based SMS Controller for Smart Phones
ABSTRACT: Text analysis includes lexical analysis of the text and has been widely
studied and used in diverse applications. In the last decade, researchers have
proposed many efficient solutions to analyze / classify large text dataset,
however, analysis / classification of short text is still a challenge because
1) the data is very sparse 2) It contains noise words and 3) It is difficult to
understand the syntactical structure of the text. Short Messaging Service (SMS)
is a text messaging service for mobile/smart phone and this service is
frequently used by all mobile users. Because of the popularity of SMS service,
marketing companies nowadays are also using this service for direct marketing
also known as SMS marketing.In this paper, we have proposed Ontology based SMS
Controller which analyze the text message and classify it using ontology
aslegitimate or spam. The proposed system has been tested on different
scenarios and experimental results shows that the proposed solution is
effective both in terms of efficiency and time.
| no_new_dataset | 0.944074 |
1509.01469 | Ruiqi Guo | Ruiqi Guo, Sanjiv Kumar, Krzysztof Choromanski and David Simcha | Quantization based Fast Inner Product Search | null | null | null | null | cs.AI cs.LG stat.ML | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | We propose a quantization based approach for fast approximate Maximum Inner
Product Search (MIPS). Each database vector is quantized in multiple subspaces
via a set of codebooks, learned directly by minimizing the inner product
quantization error. Then, the inner product of a query to a database vector is
approximated as the sum of inner products with the subspace quantizers.
Different from recently proposed LSH approaches to MIPS, the database vectors
and queries do not need to be augmented in a higher dimensional feature space.
We also provide a theoretical analysis of the proposed approach, consisting of
the concentration results under mild assumptions. Furthermore, if a small
sample of example queries is given at the training time, we propose a modified
codebook learning procedure which further improves the accuracy. Experimental
results on a variety of datasets including those arising from deep neural
networks show that the proposed approach significantly outperforms the existing
state-of-the-art.
| [
{
"version": "v1",
"created": "Fri, 4 Sep 2015 14:43:11 GMT"
}
] | 2015-09-07T00:00:00 | [
[
"Guo",
"Ruiqi",
""
],
[
"Kumar",
"Sanjiv",
""
],
[
"Choromanski",
"Krzysztof",
""
],
[
"Simcha",
"David",
""
]
] | TITLE: Quantization based Fast Inner Product Search
ABSTRACT: We propose a quantization based approach for fast approximate Maximum Inner
Product Search (MIPS). Each database vector is quantized in multiple subspaces
via a set of codebooks, learned directly by minimizing the inner product
quantization error. Then, the inner product of a query to a database vector is
approximated as the sum of inner products with the subspace quantizers.
Different from recently proposed LSH approaches to MIPS, the database vectors
and queries do not need to be augmented in a higher dimensional feature space.
We also provide a theoretical analysis of the proposed approach, consisting of
the concentration results under mild assumptions. Furthermore, if a small
sample of example queries is given at the training time, we propose a modified
codebook learning procedure which further improves the accuracy. Experimental
results on a variety of datasets including those arising from deep neural
networks show that the proposed approach significantly outperforms the existing
state-of-the-art.
| no_new_dataset | 0.941007 |
1409.6813 | Hossein Rahmani | Hossein Rahmani, Arif Mahmood, Du Huynh, Ajmal Mian | Histogram of Oriented Principal Components for Cross-View Action
Recognition | null | null | null | null | cs.CV | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | Existing techniques for 3D action recognition are sensitive to viewpoint
variations because they extract features from depth images which are viewpoint
dependent. In contrast, we directly process pointclouds for cross-view action
recognition from unknown and unseen views. We propose the Histogram of Oriented
Principal Components (HOPC) descriptor that is robust to noise, viewpoint,
scale and action speed variations. At a 3D point, HOPC is computed by
projecting the three scaled eigenvectors of the pointcloud within its local
spatio-temporal support volume onto the vertices of a regular dodecahedron.
HOPC is also used for the detection of Spatio-Temporal Keypoints (STK) in 3D
pointcloud sequences so that view-invariant STK descriptors (or Local HOPC
descriptors) at these key locations only are used for action recognition. We
also propose a global descriptor computed from the normalized spatio-temporal
distribution of STKs in 4-D, which we refer to as STK-D. We have evaluated the
performance of our proposed descriptors against nine existing techniques on two
cross-view and three single-view human action recognition datasets. The
Experimental results show that our techniques provide significant improvement
over state-of-the-art methods.
| [
{
"version": "v1",
"created": "Wed, 24 Sep 2014 03:57:49 GMT"
},
{
"version": "v2",
"created": "Thu, 3 Sep 2015 05:12:27 GMT"
}
] | 2015-09-04T00:00:00 | [
[
"Rahmani",
"Hossein",
""
],
[
"Mahmood",
"Arif",
""
],
[
"Huynh",
"Du",
""
],
[
"Mian",
"Ajmal",
""
]
] | TITLE: Histogram of Oriented Principal Components for Cross-View Action
Recognition
ABSTRACT: Existing techniques for 3D action recognition are sensitive to viewpoint
variations because they extract features from depth images which are viewpoint
dependent. In contrast, we directly process pointclouds for cross-view action
recognition from unknown and unseen views. We propose the Histogram of Oriented
Principal Components (HOPC) descriptor that is robust to noise, viewpoint,
scale and action speed variations. At a 3D point, HOPC is computed by
projecting the three scaled eigenvectors of the pointcloud within its local
spatio-temporal support volume onto the vertices of a regular dodecahedron.
HOPC is also used for the detection of Spatio-Temporal Keypoints (STK) in 3D
pointcloud sequences so that view-invariant STK descriptors (or Local HOPC
descriptors) at these key locations only are used for action recognition. We
also propose a global descriptor computed from the normalized spatio-temporal
distribution of STKs in 4-D, which we refer to as STK-D. We have evaluated the
performance of our proposed descriptors against nine existing techniques on two
cross-view and three single-view human action recognition datasets. The
Experimental results show that our techniques provide significant improvement
over state-of-the-art methods.
| no_new_dataset | 0.94428 |
1506.00976 | Gautier Marti | Gautier Marti, Philippe Very and Philippe Donnat | Toward a generic representation of random variables for machine learning | submitted to Pattern Recognition Letters | null | null | null | cs.LG stat.ML | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | This paper presents a pre-processing and a distance which improve the
performance of machine learning algorithms working on independent and
identically distributed stochastic processes. We introduce a novel
non-parametric approach to represent random variables which splits apart
dependency and distribution without losing any information. We also propound an
associated metric leveraging this representation and its statistical estimate.
Besides experiments on synthetic datasets, the benefits of our contribution is
illustrated through the example of clustering financial time series, for
instance prices from the credit default swaps market. Results are available on
the website www.datagrapple.com and an IPython Notebook tutorial is available
at www.datagrapple.com/Tech for reproducible research.
| [
{
"version": "v1",
"created": "Tue, 2 Jun 2015 17:58:48 GMT"
},
{
"version": "v2",
"created": "Thu, 3 Sep 2015 19:23:30 GMT"
}
] | 2015-09-04T00:00:00 | [
[
"Marti",
"Gautier",
""
],
[
"Very",
"Philippe",
""
],
[
"Donnat",
"Philippe",
""
]
] | TITLE: Toward a generic representation of random variables for machine learning
ABSTRACT: This paper presents a pre-processing and a distance which improve the
performance of machine learning algorithms working on independent and
identically distributed stochastic processes. We introduce a novel
non-parametric approach to represent random variables which splits apart
dependency and distribution without losing any information. We also propound an
associated metric leveraging this representation and its statistical estimate.
Besides experiments on synthetic datasets, the benefits of our contribution is
illustrated through the example of clustering financial time series, for
instance prices from the credit default swaps market. Results are available on
the website www.datagrapple.com and an IPython Notebook tutorial is available
at www.datagrapple.com/Tech for reproducible research.
| no_new_dataset | 0.944022 |
1509.01074 | Ahmed Mohamed | Ahmed Nabil Mohamed | A Novice Guide towards Human Motion Analysis and Understanding | 35 Pages | null | null | null | cs.CV | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | Human motion analysis and understanding has been, and is still, the focus of
attention of many disciplines which is considered an obvious indicator of the
wide and massive importance of the subject. The purpose of this article is to
shed some light on this very important subject, so it can be a good insight for
a novice computer vision researcher in this field by providing him/her with a
wealth of knowledge about the subject covering many directions. There are two
main contributions of this article. The first one investigates various aspects
of some disciplines (e.g., arts, philosophy, psychology, and neuroscience) that
are interested in the subject and review some of their contributions stressing
on those that can be useful for computer vision researchers. Moreover, many
examples are illustrated to indicate the benefits of integrating concepts and
results among different disciplines. The second contribution is concerned with
the subject from the computer vision aspect where we discuss the following
issues. First, we explore many demanding and promising applications to reveal
the wide and massive importance of the field. Second, we list various types of
sensors that may be used for acquiring various data. Third, we review different
taxonomies used for classifying motions. Fourth, we review various processes
involved in motion analysis. Fifth, we exhibit how different surveys are
structured. Sixth, we examine many of the most cited and recent reviews in the
field that have been published during the past two decades to reveal various
approaches used for implementing different stages of the problem and refer to
various algorithms and their suitability for different situations. Moreover, we
provide a long list of public datasets and discuss briefly some examples of
these datasets. Finally, we provide a general discussion of the subject from
the aspect of computer vision.
| [
{
"version": "v1",
"created": "Thu, 3 Sep 2015 13:25:37 GMT"
}
] | 2015-09-04T00:00:00 | [
[
"Mohamed",
"Ahmed Nabil",
""
]
] | TITLE: A Novice Guide towards Human Motion Analysis and Understanding
ABSTRACT: Human motion analysis and understanding has been, and is still, the focus of
attention of many disciplines which is considered an obvious indicator of the
wide and massive importance of the subject. The purpose of this article is to
shed some light on this very important subject, so it can be a good insight for
a novice computer vision researcher in this field by providing him/her with a
wealth of knowledge about the subject covering many directions. There are two
main contributions of this article. The first one investigates various aspects
of some disciplines (e.g., arts, philosophy, psychology, and neuroscience) that
are interested in the subject and review some of their contributions stressing
on those that can be useful for computer vision researchers. Moreover, many
examples are illustrated to indicate the benefits of integrating concepts and
results among different disciplines. The second contribution is concerned with
the subject from the computer vision aspect where we discuss the following
issues. First, we explore many demanding and promising applications to reveal
the wide and massive importance of the field. Second, we list various types of
sensors that may be used for acquiring various data. Third, we review different
taxonomies used for classifying motions. Fourth, we review various processes
involved in motion analysis. Fifth, we exhibit how different surveys are
structured. Sixth, we examine many of the most cited and recent reviews in the
field that have been published during the past two decades to reveal various
approaches used for implementing different stages of the problem and refer to
various algorithms and their suitability for different situations. Moreover, we
provide a long list of public datasets and discuss briefly some examples of
these datasets. Finally, we provide a general discussion of the subject from
the aspect of computer vision.
| no_new_dataset | 0.924756 |
1412.5129 | David Weyburne | David Weyburne | The Prevalence of Similarity of the Turbulent Wall-bounded Velocity
Profile | 17 pages, 10 figures, 1 appendix. This update adds a paragraph that
shows that the similarity equivalency argument is valid not only for whole
profile similarity but is also valid even for the case where only the outer
region similarity is considered | null | null | null | physics.flu-dyn | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | Castillo and George (Castillo, L. and George, W., AIAA J. 39, 41(2001))
developed a flow governing equation approach for describing the turbulent outer
boundary layer region. The approach was used to develop similarity criteria for
the mean velocity and Reynolds shear stress profiles. Using the criteria as a
guide, Castillo, George, and coworkers examined an extensive set of
experimental datasets and claim that most of these turbulent velocity boundary
layers appear to be similar boundary layers when scaled with the Zagarola and
Smits (Zagarola, M. and Smits, A., J. Fluid Mech. 373, 33(1998)) velocity
parameter. In the work herein it is shown that their success at showing scaled
profile similarity in many of those datasets is flawed due to a similarity
problem that occurs when one combines the defect profile and the Zagarola and
Smits type of velocity scaling parameter. The same problem has been identified
in other papers in the literature and may in fact be widespread. We conclude
that similarity of the turbulent velocity profile is not as prevalent as was
claimed by Castillo, George, and coworkers. The result has implications as to
the accepted paradigm of the scaling of the turbulent boundary layer.
| [
{
"version": "v1",
"created": "Tue, 16 Dec 2014 19:15:03 GMT"
},
{
"version": "v2",
"created": "Fri, 6 Mar 2015 15:50:30 GMT"
},
{
"version": "v3",
"created": "Fri, 3 Apr 2015 16:29:42 GMT"
},
{
"version": "v4",
"created": "Wed, 3 Jun 2015 18:56:47 GMT"
},
{
"version": "v5",
"created": "Wed, 2 Sep 2015 13:46:11 GMT"
}
] | 2015-09-03T00:00:00 | [
[
"Weyburne",
"David",
""
]
] | TITLE: The Prevalence of Similarity of the Turbulent Wall-bounded Velocity
Profile
ABSTRACT: Castillo and George (Castillo, L. and George, W., AIAA J. 39, 41(2001))
developed a flow governing equation approach for describing the turbulent outer
boundary layer region. The approach was used to develop similarity criteria for
the mean velocity and Reynolds shear stress profiles. Using the criteria as a
guide, Castillo, George, and coworkers examined an extensive set of
experimental datasets and claim that most of these turbulent velocity boundary
layers appear to be similar boundary layers when scaled with the Zagarola and
Smits (Zagarola, M. and Smits, A., J. Fluid Mech. 373, 33(1998)) velocity
parameter. In the work herein it is shown that their success at showing scaled
profile similarity in many of those datasets is flawed due to a similarity
problem that occurs when one combines the defect profile and the Zagarola and
Smits type of velocity scaling parameter. The same problem has been identified
in other papers in the literature and may in fact be widespread. We conclude
that similarity of the turbulent velocity profile is not as prevalent as was
claimed by Castillo, George, and coworkers. The result has implications as to
the accepted paradigm of the scaling of the turbulent boundary layer.
| no_new_dataset | 0.954223 |
1509.00511 | Xitong Yang | Xitong Yang, Yuncheng Li, Jiebo Luo | Pinterest Board Recommendation for Twitter Users | null | null | 10.1145/2733373.2806375 | null | cs.SI cs.MM | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | Pinboard on Pinterest is an emerging media to engage online social media
users, on which users post online images for specific topics. Regardless of its
significance, there is little previous work specifically to facilitate
information discovery based on pinboards. This paper proposes a novel pinboard
recommendation system for Twitter users. In order to associate contents from
the two social media platforms, we propose to use MultiLabel classification to
map Twitter user followees to pinboard topics and visual diversification to
recommend pinboards given user interested topics. A preliminary experiment on a
dataset with 2000 users validated our proposed system.
| [
{
"version": "v1",
"created": "Tue, 1 Sep 2015 21:42:27 GMT"
}
] | 2015-09-03T00:00:00 | [
[
"Yang",
"Xitong",
""
],
[
"Li",
"Yuncheng",
""
],
[
"Luo",
"Jiebo",
""
]
] | TITLE: Pinterest Board Recommendation for Twitter Users
ABSTRACT: Pinboard on Pinterest is an emerging media to engage online social media
users, on which users post online images for specific topics. Regardless of its
significance, there is little previous work specifically to facilitate
information discovery based on pinboards. This paper proposes a novel pinboard
recommendation system for Twitter users. In order to associate contents from
the two social media platforms, we propose to use MultiLabel classification to
map Twitter user followees to pinboard topics and visual diversification to
recommend pinboards given user interested topics. A preliminary experiment on a
dataset with 2000 users validated our proposed system.
| no_new_dataset | 0.937211 |
1509.00533 | Scott Wisdom | Scott Wisdom, Thomas Powers, Les Atlas, and James Pitton | Enhancement and Recognition of Reverberant and Noisy Speech by Extending
Its Coherence | 22 pages | null | null | null | cs.SD cs.CL stat.AP | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | Most speech enhancement algorithms make use of the short-time Fourier
transform (STFT), which is a simple and flexible time-frequency decomposition
that estimates the short-time spectrum of a signal. However, the duration of
short STFT frames are inherently limited by the nonstationarity of speech
signals. The main contribution of this paper is a demonstration of speech
enhancement and automatic speech recognition in the presence of reverberation
and noise by extending the length of analysis windows. We accomplish this
extension by performing enhancement in the short-time fan-chirp transform
(STFChT) domain, an overcomplete time-frequency representation that is coherent
with speech signals over longer analysis window durations than the STFT. This
extended coherence is gained by using a linear model of fundamental frequency
variation of voiced speech signals. Our approach centers around using a
single-channel minimum mean-square error log-spectral amplitude (MMSE-LSA)
estimator proposed by Habets, which scales coefficients in a time-frequency
domain to suppress noise and reverberation. In the case of multiple
microphones, we preprocess the data with either a minimum variance
distortionless response (MVDR) beamformer, or a delay-and-sum beamformer (DSB).
We evaluate our algorithm on both speech enhancement and recognition tasks for
the REVERB challenge dataset. Compared to the same processing done in the STFT
domain, our approach achieves significant improvement in terms of objective
enhancement metrics (including PESQ---the ITU-T standard measurement for speech
quality). In terms of automatic speech recognition (ASR) performance as
measured by word error rate (WER), our experiments indicate that the STFT with
a long window is more effective for ASR.
| [
{
"version": "v1",
"created": "Wed, 2 Sep 2015 00:31:40 GMT"
}
] | 2015-09-03T00:00:00 | [
[
"Wisdom",
"Scott",
""
],
[
"Powers",
"Thomas",
""
],
[
"Atlas",
"Les",
""
],
[
"Pitton",
"James",
""
]
] | TITLE: Enhancement and Recognition of Reverberant and Noisy Speech by Extending
Its Coherence
ABSTRACT: Most speech enhancement algorithms make use of the short-time Fourier
transform (STFT), which is a simple and flexible time-frequency decomposition
that estimates the short-time spectrum of a signal. However, the duration of
short STFT frames are inherently limited by the nonstationarity of speech
signals. The main contribution of this paper is a demonstration of speech
enhancement and automatic speech recognition in the presence of reverberation
and noise by extending the length of analysis windows. We accomplish this
extension by performing enhancement in the short-time fan-chirp transform
(STFChT) domain, an overcomplete time-frequency representation that is coherent
with speech signals over longer analysis window durations than the STFT. This
extended coherence is gained by using a linear model of fundamental frequency
variation of voiced speech signals. Our approach centers around using a
single-channel minimum mean-square error log-spectral amplitude (MMSE-LSA)
estimator proposed by Habets, which scales coefficients in a time-frequency
domain to suppress noise and reverberation. In the case of multiple
microphones, we preprocess the data with either a minimum variance
distortionless response (MVDR) beamformer, or a delay-and-sum beamformer (DSB).
We evaluate our algorithm on both speech enhancement and recognition tasks for
the REVERB challenge dataset. Compared to the same processing done in the STFT
domain, our approach achieves significant improvement in terms of objective
enhancement metrics (including PESQ---the ITU-T standard measurement for speech
quality). In terms of automatic speech recognition (ASR) performance as
measured by word error rate (WER), our experiments indicate that the STFT with
a long window is more effective for ASR.
| no_new_dataset | 0.949949 |
1509.00568 | Michael (Micky) Fire | Michael Fire and Jonathan Schler | Exploring Online Ad Images Using a Deep Convolutional Neural Network
Approach | null | null | null | null | cs.CV | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | Online advertising is a huge, rapidly growing advertising market in today's
world. One common form of online advertising is using image ads. A decision is
made (often in real time) every time a user sees an ad, and the advertiser is
eager to determine the best ad to display. Consequently, many algorithms have
been developed that calculate the optimal ad to show to the current user at the
present time. Typically, these algorithms focus on variations of the ad,
optimizing among different properties such as background color, image size, or
set of images. However, there is a more fundamental layer. Our study looks at
new qualities of ads that can be determined before an ad is shown (rather than
online optimization) and defines which ads are most likely to be successful.
We present a set of novel algorithms that utilize deep-learning image
processing, machine learning, and graph theory to investigate online
advertising and to construct prediction models which can foresee an image ad's
success. We evaluated our algorithms on a dataset with over 260,000 ad images,
as well as a smaller dataset specifically related to the automotive industry,
and we succeeded in constructing regression models for ad image click rate
prediction. The obtained results emphasize the great potential of using
deep-learning algorithms to effectively and efficiently analyze image ads and
to create better and more innovative online ads. Moreover, the algorithms
presented in this paper can help predict ad success and can be applied to
analyze other large-scale image corpora.
| [
{
"version": "v1",
"created": "Wed, 2 Sep 2015 06:18:27 GMT"
}
] | 2015-09-03T00:00:00 | [
[
"Fire",
"Michael",
""
],
[
"Schler",
"Jonathan",
""
]
] | TITLE: Exploring Online Ad Images Using a Deep Convolutional Neural Network
Approach
ABSTRACT: Online advertising is a huge, rapidly growing advertising market in today's
world. One common form of online advertising is using image ads. A decision is
made (often in real time) every time a user sees an ad, and the advertiser is
eager to determine the best ad to display. Consequently, many algorithms have
been developed that calculate the optimal ad to show to the current user at the
present time. Typically, these algorithms focus on variations of the ad,
optimizing among different properties such as background color, image size, or
set of images. However, there is a more fundamental layer. Our study looks at
new qualities of ads that can be determined before an ad is shown (rather than
online optimization) and defines which ads are most likely to be successful.
We present a set of novel algorithms that utilize deep-learning image
processing, machine learning, and graph theory to investigate online
advertising and to construct prediction models which can foresee an image ad's
success. We evaluated our algorithms on a dataset with over 260,000 ad images,
as well as a smaller dataset specifically related to the automotive industry,
and we succeeded in constructing regression models for ad image click rate
prediction. The obtained results emphasize the great potential of using
deep-learning algorithms to effectively and efficiently analyze image ads and
to create better and more innovative online ads. Moreover, the algorithms
presented in this paper can help predict ad success and can be applied to
analyze other large-scale image corpora.
| no_new_dataset | 0.938463 |
1502.05680 | Andrea Montanari | Andrea Montanari | Finding One Community in a Sparse Graph | 30 pages, 8 pdf figures | null | 10.1007/s10955-015-1338-2 | null | stat.ML cond-mat.stat-mech cs.SI | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | We consider a random sparse graph with bounded average degree, in which a
subset of vertices has higher connectivity than the background. In particular,
the average degree inside this subset of vertices is larger than outside (but
still bounded). Given a realization of such graph, we aim at identifying the
hidden subset of vertices. This can be regarded as a model for the problem of
finding a tightly knitted community in a social network, or a cluster in a
relational dataset.
In this paper we present two sets of contributions: $(i)$ We use the cavity
method from spin glass theory to derive an exact phase diagram for the
reconstruction problem. In particular, as the difference in edge probability
increases, the problem undergoes two phase transitions, a static phase
transition and a dynamic one. $(ii)$ We establish rigorous bounds on the
dynamic phase transition and prove that, above a certain threshold, a local
algorithm (belief propagation) correctly identify most of the hidden set. Below
the same threshold \emph{no local algorithm} can achieve this goal. However, in
this regime the subset can be identified by exhaustive search.
For small hidden sets and large average degree, the phase transition for
local algorithms takes an intriguingly simple form. Local algorithms succeed
with high probability for ${\rm deg}_{\rm in} - {\rm deg}_{\rm out} >
\sqrt{{\rm deg}_{\rm out}/e}$ and fail for ${\rm deg}_{\rm in} - {\rm deg}_{\rm
out} < \sqrt{{\rm deg}_{\rm out}/e}$ (with ${\rm deg}_{\rm in}$, ${\rm
deg}_{\rm out}$ the average degrees inside and outside the community). We argue
that spectral algorithms are also ineffective in the latter regime.
It is an open problem whether any polynomial time algorithms might succeed
for ${\rm deg}_{\rm in} - {\rm deg}_{\rm out} < \sqrt{{\rm deg}_{\rm out}/e}$.
| [
{
"version": "v1",
"created": "Thu, 19 Feb 2015 19:50:09 GMT"
},
{
"version": "v2",
"created": "Thu, 30 Jul 2015 19:46:13 GMT"
}
] | 2015-09-02T00:00:00 | [
[
"Montanari",
"Andrea",
""
]
] | TITLE: Finding One Community in a Sparse Graph
ABSTRACT: We consider a random sparse graph with bounded average degree, in which a
subset of vertices has higher connectivity than the background. In particular,
the average degree inside this subset of vertices is larger than outside (but
still bounded). Given a realization of such graph, we aim at identifying the
hidden subset of vertices. This can be regarded as a model for the problem of
finding a tightly knitted community in a social network, or a cluster in a
relational dataset.
In this paper we present two sets of contributions: $(i)$ We use the cavity
method from spin glass theory to derive an exact phase diagram for the
reconstruction problem. In particular, as the difference in edge probability
increases, the problem undergoes two phase transitions, a static phase
transition and a dynamic one. $(ii)$ We establish rigorous bounds on the
dynamic phase transition and prove that, above a certain threshold, a local
algorithm (belief propagation) correctly identify most of the hidden set. Below
the same threshold \emph{no local algorithm} can achieve this goal. However, in
this regime the subset can be identified by exhaustive search.
For small hidden sets and large average degree, the phase transition for
local algorithms takes an intriguingly simple form. Local algorithms succeed
with high probability for ${\rm deg}_{\rm in} - {\rm deg}_{\rm out} >
\sqrt{{\rm deg}_{\rm out}/e}$ and fail for ${\rm deg}_{\rm in} - {\rm deg}_{\rm
out} < \sqrt{{\rm deg}_{\rm out}/e}$ (with ${\rm deg}_{\rm in}$, ${\rm
deg}_{\rm out}$ the average degrees inside and outside the community). We argue
that spectral algorithms are also ineffective in the latter regime.
It is an open problem whether any polynomial time algorithms might succeed
for ${\rm deg}_{\rm in} - {\rm deg}_{\rm out} < \sqrt{{\rm deg}_{\rm out}/e}$.
| no_new_dataset | 0.947624 |
1506.01798 | Shyeh Tjing Loi | Shyeh Tjing Loi, Cathryn M. Trott, Tara Murphy, Iver H. Cairns, Martin
Bell, Natasha Hurley-Walker, John Morgan, Emil Lenc, A. R. Offringa, L. Feng,
P. J. Hancock, D. L. Kaplan, N. Kudryavtseva, G. Bernardi, J. D. Bowman, F.
Briggs, R. J. Cappallo, B. E. Corey, A. A. Deshpande, D. Emrich, B. M.
Gaensler, R. Goeke, L. J. Greenhill, B. J. Hazelton, M. Johnston-Hollitt, J.
C. Kasper, E. Kratzenberg, C. J. Lonsdale, M. J. Lynch, S. R. McWhirter, D.
A. Mitchell, M. F. Morales, E. Morgan, D. Oberoi, S. M. Ord, T. Prabu, A. E.
E. Rogers, A. Roshi, N. Udaya Shankar, K. S. Srivani, R. Subrahmanyan, S. J.
Tingay, M. Waterson, R. B. Wayth, R. L. Webster, A. R. Whitney, A. Williams
and C. L. Williams | Power spectrum analysis of ionospheric fluctuations with the Murchison
Widefield Array | Accepted for publication in Radio Science | null | 10.1002/2015RS005711 | null | astro-ph.IM physics.space-ph | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | Low-frequency, wide field-of-view (FoV) radio telescopes such as the
Murchison Widefield Array (MWA) enable the ionosphere to be sampled at high
spatial completeness. We present the results of the first power spectrum
analysis of ionospheric fluctuations in MWA data, where we examined the
position offsets of radio sources appearing in two datasets. The refractive
shifts in the positions of celestial sources are proportional to spatial
gradients in the electron column density transverse to the line of sight. These
can be used to probe plasma structures and waves in the ionosphere. The
regional (10-100 km) scales probed by the MWA, determined by the size of its
FoV and the spatial density of radio sources (typically thousands in a single
FoV), complement the global (100-1000 km) scales of GPS studies and local
(0.01-1 km) scales of radar scattering measurements. Our data exhibit a range
of complex structures and waves. Some fluctuations have the characteristics of
travelling ionospheric disturbances (TIDs), while others take the form of
narrow, slowly-drifting bands aligned along the Earth's magnetic field.
| [
{
"version": "v1",
"created": "Fri, 5 Jun 2015 07:31:32 GMT"
}
] | 2015-09-02T00:00:00 | [
[
"Loi",
"Shyeh Tjing",
""
],
[
"Trott",
"Cathryn M.",
""
],
[
"Murphy",
"Tara",
""
],
[
"Cairns",
"Iver H.",
""
],
[
"Bell",
"Martin",
""
],
[
"Hurley-Walker",
"Natasha",
""
],
[
"Morgan",
"John",
""
],
[
"Lenc",
"Emil",
""
],
[
"Offringa",
"A. R.",
""
],
[
"Feng",
"L.",
""
],
[
"Hancock",
"P. J.",
""
],
[
"Kaplan",
"D. L.",
""
],
[
"Kudryavtseva",
"N.",
""
],
[
"Bernardi",
"G.",
""
],
[
"Bowman",
"J. D.",
""
],
[
"Briggs",
"F.",
""
],
[
"Cappallo",
"R. J.",
""
],
[
"Corey",
"B. E.",
""
],
[
"Deshpande",
"A. A.",
""
],
[
"Emrich",
"D.",
""
],
[
"Gaensler",
"B. M.",
""
],
[
"Goeke",
"R.",
""
],
[
"Greenhill",
"L. J.",
""
],
[
"Hazelton",
"B. J.",
""
],
[
"Johnston-Hollitt",
"M.",
""
],
[
"Kasper",
"J. C.",
""
],
[
"Kratzenberg",
"E.",
""
],
[
"Lonsdale",
"C. J.",
""
],
[
"Lynch",
"M. J.",
""
],
[
"McWhirter",
"S. R.",
""
],
[
"Mitchell",
"D. A.",
""
],
[
"Morales",
"M. F.",
""
],
[
"Morgan",
"E.",
""
],
[
"Oberoi",
"D.",
""
],
[
"Ord",
"S. M.",
""
],
[
"Prabu",
"T.",
""
],
[
"Rogers",
"A. E. E.",
""
],
[
"Roshi",
"A.",
""
],
[
"Shankar",
"N. Udaya",
""
],
[
"Srivani",
"K. S.",
""
],
[
"Subrahmanyan",
"R.",
""
],
[
"Tingay",
"S. J.",
""
],
[
"Waterson",
"M.",
""
],
[
"Wayth",
"R. B.",
""
],
[
"Webster",
"R. L.",
""
],
[
"Whitney",
"A. R.",
""
],
[
"Williams",
"A.",
""
],
[
"Williams",
"C. L.",
""
]
] | TITLE: Power spectrum analysis of ionospheric fluctuations with the Murchison
Widefield Array
ABSTRACT: Low-frequency, wide field-of-view (FoV) radio telescopes such as the
Murchison Widefield Array (MWA) enable the ionosphere to be sampled at high
spatial completeness. We present the results of the first power spectrum
analysis of ionospheric fluctuations in MWA data, where we examined the
position offsets of radio sources appearing in two datasets. The refractive
shifts in the positions of celestial sources are proportional to spatial
gradients in the electron column density transverse to the line of sight. These
can be used to probe plasma structures and waves in the ionosphere. The
regional (10-100 km) scales probed by the MWA, determined by the size of its
FoV and the spatial density of radio sources (typically thousands in a single
FoV), complement the global (100-1000 km) scales of GPS studies and local
(0.01-1 km) scales of radar scattering measurements. Our data exhibit a range
of complex structures and waves. Some fluctuations have the characteristics of
travelling ionospheric disturbances (TIDs), while others take the form of
narrow, slowly-drifting bands aligned along the Earth's magnetic field.
| no_new_dataset | 0.940626 |
1506.07324 | Mariusz Tarnopolski | Mariusz Tarnopolski | Analysis of Fermi gamma-ray burst duration distribution | 6 pages, 3 figures; matches the version to be published | A&A 581, A29 (2015) | 10.1051/0004-6361/201526415 | null | astro-ph.HE astro-ph.CO hep-ph physics.space-ph | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | Two classes of GRBs, short and long, have been determined without any doubts,
and are usually prescribed to different physical scenarios. A third class,
intermediate in $T_{90}$ durations, has been reported to be present in the
datasets of BATSE, Swift, RHESSI and possibly BeppoSAX. The latest release of
$>1500$ GRBs observed by Fermi gives an opportunity to further investigate the
duration distribution. The aim of this paper is to investigate whether a third
class is present in the $\log T_{90}$ distribution, or is it described by a
bimodal distribution. A standard $\chi^2$ fitting of a mixture of Gaussians is
applied to 25 histograms with different binnings. Different binnings give
various values of the fitting parameters, as well as the shape of the fitted
curve. Among five statistically significant fits none is trimodal. Locations of
the Gaussian components are in agreement with previous works. However, a
trimodal distribution, understood in the sense of having three separated peaks,
is not found for any binning. It is concluded that the duration distribution in
Fermi data is well described by a mixture of three log-normal distributions,
but it is intrinsically bimodal, hence no third class is present in the
$T_{90}$ data of Fermi. It is suggested that the log-normal fit may not be an
adequate model.
| [
{
"version": "v1",
"created": "Wed, 24 Jun 2015 11:28:41 GMT"
},
{
"version": "v2",
"created": "Tue, 7 Jul 2015 15:39:44 GMT"
}
] | 2015-09-02T00:00:00 | [
[
"Tarnopolski",
"Mariusz",
""
]
] | TITLE: Analysis of Fermi gamma-ray burst duration distribution
ABSTRACT: Two classes of GRBs, short and long, have been determined without any doubts,
and are usually prescribed to different physical scenarios. A third class,
intermediate in $T_{90}$ durations, has been reported to be present in the
datasets of BATSE, Swift, RHESSI and possibly BeppoSAX. The latest release of
$>1500$ GRBs observed by Fermi gives an opportunity to further investigate the
duration distribution. The aim of this paper is to investigate whether a third
class is present in the $\log T_{90}$ distribution, or is it described by a
bimodal distribution. A standard $\chi^2$ fitting of a mixture of Gaussians is
applied to 25 histograms with different binnings. Different binnings give
various values of the fitting parameters, as well as the shape of the fitted
curve. Among five statistically significant fits none is trimodal. Locations of
the Gaussian components are in agreement with previous works. However, a
trimodal distribution, understood in the sense of having three separated peaks,
is not found for any binning. It is concluded that the duration distribution in
Fermi data is well described by a mixture of three log-normal distributions,
but it is intrinsically bimodal, hence no third class is present in the
$T_{90}$ data of Fermi. It is suggested that the log-normal fit may not be an
adequate model.
| no_new_dataset | 0.947381 |
1509.00083 | Samuel Kadoury | Samuel Kadoury, Eugene Vorontsov, An Tang | Metastatic liver tumour segmentation from discriminant Grassmannian
manifolds | null | Physics in Medicine and Biology 60 (2015) | 10.1088/0031-9155/60/16/6459 | null | cs.LG cs.CV | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | The early detection, diagnosis and monitoring of liver cancer progression can
be achieved with the precise delineation of metastatic tumours. However,
accurate automated segmentation remains challenging due to the presence of
noise, inhomogeneity and the high appearance variability of malignant tissue.
In this paper, we propose an unsupervised metastatic liver tumour segmentation
framework using a machine learning approach based on discriminant Grassmannian
manifolds which learns the appearance of tumours with respect to normal tissue.
First, the framework learns within-class and between-class similarity
distributions from a training set of images to discover the optimal manifold
discrimination between normal and pathological tissue in the liver. Second, a
conditional optimisation scheme computes nonlocal pairwise as well as
pattern-based clique potentials from the manifold subspace to recognise regions
with similar labelings and to incorporate global consistency in the
segmentation process. The proposed framework was validated on a clinical
database of 43 CT images from patients with metastatic liver cancer. Compared
to state-of-the-art methods, our method achieves a better performance on two
separate datasets of metastatic liver tumours from different clinical sites,
yielding an overall mean Dice similarity coefficient of 90.7 +/- 2.4 in over 50
tumours with an average volume of 27.3 mm3.
| [
{
"version": "v1",
"created": "Mon, 31 Aug 2015 21:45:40 GMT"
}
] | 2015-09-02T00:00:00 | [
[
"Kadoury",
"Samuel",
""
],
[
"Vorontsov",
"Eugene",
""
],
[
"Tang",
"An",
""
]
] | TITLE: Metastatic liver tumour segmentation from discriminant Grassmannian
manifolds
ABSTRACT: The early detection, diagnosis and monitoring of liver cancer progression can
be achieved with the precise delineation of metastatic tumours. However,
accurate automated segmentation remains challenging due to the presence of
noise, inhomogeneity and the high appearance variability of malignant tissue.
In this paper, we propose an unsupervised metastatic liver tumour segmentation
framework using a machine learning approach based on discriminant Grassmannian
manifolds which learns the appearance of tumours with respect to normal tissue.
First, the framework learns within-class and between-class similarity
distributions from a training set of images to discover the optimal manifold
discrimination between normal and pathological tissue in the liver. Second, a
conditional optimisation scheme computes nonlocal pairwise as well as
pattern-based clique potentials from the manifold subspace to recognise regions
with similar labelings and to incorporate global consistency in the
segmentation process. The proposed framework was validated on a clinical
database of 43 CT images from patients with metastatic liver cancer. Compared
to state-of-the-art methods, our method achieves a better performance on two
separate datasets of metastatic liver tumours from different clinical sites,
yielding an overall mean Dice similarity coefficient of 90.7 +/- 2.4 in over 50
tumours with an average volume of 27.3 mm3.
| no_new_dataset | 0.948155 |
1509.00313 | Amit K.C. | Amit Kumar K.C., Damien Delannay and Christophe De Vleeschouwer | Iterative hypothesis testing for multi-object tracking in presence of
features with variable reliability | 21 pages, 8 figures, submitted to CVIU: Special Issue on Visual
Tracking | null | null | null | cs.CV | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | This paper assumes prior detections of multiple targets at each time instant,
and uses a graph-based approach to connect those detections across time, based
on their position and appearance estimates. In contrast to most earlier works
in the field, our framework has been designed to exploit the appearance
features, even when they are only sporadically available, or affected by a
non-stationary noise, along the sequence of detections. This is done by
implementing an iterative hypothesis testing strategy to progressively
aggregate the detections into short trajectories, named tracklets.
Specifically, each iteration considers a node, named key-node, and investigates
how to link this key-node with other nodes in its neighborhood, under the
assumption that the target appearance is defined by the key-node appearance
estimate. This is done through shortest path computation in a temporal
neighborhood of the key-node. The approach is conservative in that it only
aggregates the shortest paths that are sufficiently better compared to
alternative paths. It is also multi-scale in that the size of the investigated
neighborhood is increased proportionally to the number of detections already
aggregated into the key-node. The multi-scale nature of the process and the
progressive relaxation of its conservativeness makes it both computationally
efficient and effective.
Experimental validations are performed extensively on a toy example, a 15
minutes long multi-view basketball dataset, and other monocular pedestrian
datasets.
| [
{
"version": "v1",
"created": "Tue, 1 Sep 2015 14:27:50 GMT"
}
] | 2015-09-02T00:00:00 | [
[
"C.",
"Amit Kumar K.",
""
],
[
"Delannay",
"Damien",
""
],
[
"De Vleeschouwer",
"Christophe",
""
]
] | TITLE: Iterative hypothesis testing for multi-object tracking in presence of
features with variable reliability
ABSTRACT: This paper assumes prior detections of multiple targets at each time instant,
and uses a graph-based approach to connect those detections across time, based
on their position and appearance estimates. In contrast to most earlier works
in the field, our framework has been designed to exploit the appearance
features, even when they are only sporadically available, or affected by a
non-stationary noise, along the sequence of detections. This is done by
implementing an iterative hypothesis testing strategy to progressively
aggregate the detections into short trajectories, named tracklets.
Specifically, each iteration considers a node, named key-node, and investigates
how to link this key-node with other nodes in its neighborhood, under the
assumption that the target appearance is defined by the key-node appearance
estimate. This is done through shortest path computation in a temporal
neighborhood of the key-node. The approach is conservative in that it only
aggregates the shortest paths that are sufficiently better compared to
alternative paths. It is also multi-scale in that the size of the investigated
neighborhood is increased proportionally to the number of detections already
aggregated into the key-node. The multi-scale nature of the process and the
progressive relaxation of its conservativeness makes it both computationally
efficient and effective.
Experimental validations are performed extensively on a toy example, a 15
minutes long multi-view basketball dataset, and other monocular pedestrian
datasets.
| no_new_dataset | 0.941547 |
1509.00386 | Antonio Spanu | A. Spanu (1,2,3), M. de' Michieli Vitturi (1) and S. Barsotti (1,4)
((1) Istituto Nazionale di Geofisica e Vulcanologia, Sezione di Pisa, Italy,
(2) Scuola Normale Superiore di Pisa, Italy, (3) Now at Deutschen Zentrums
fur Luft- und Raumfahrt, Germany, (4) Now at Icelandic Meteorological Office,
Iceland) | Reconstructing eruptive source parameters from tephra deposit: a
numerical approach for medium-sized explosive eruptions | Article:24 pages, 9 figures, 1 table, Auxiliary Material:xi pages, 6
figures, 1 table | null | null | null | physics.geo-ph | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | Since the seventies, several reconstruction techniques have been proposed,
and are currently used, to extrapolate and quantify eruptive parameters from
sampled deposit datasets. Discrete numbers of tephra ground loadings or
stratigraphic records are usually processed to estimate source eruptive values.
Reconstruction techniques like Pyle, Power law and Weibull are adopted as
standard to quantify the erupted mass (or volume) whereas Voronoi for
reconstructing the granulometry. Reconstructed values can be affected by large
uncertainty due to complexities occurring within the atmospheric dispersion and
deposition of volcanic particles. Here we want to quantify the sensitivity of
reconstruction techniques, and to quantify how much estimated values of mass
and grain size differ from emitted and deposited ones. We adopted a numerical
approach simulating with a dispersal code a mild explosive event occurring at
Mt. Etna, with eruptive parameters similar to those estimated for eruptions
occurred in the last decade. Then we created a synthetic deposit by integrating
the mass on the ground computed by the model over the computational domain
(>50000 km2). Multiple samplings of the simulated deposit are used for
generating a large dataset of sampling tests afterwards processed with standard
reconstruction techniques. Results are then compared and evaluated through a
statistical analysis, based on 2000 sampling tests of 100 samplings points. On
average, all the used techniques underestimate deposited and emitted mass. A
similar analysis, carried on Voronoi results, shows that information on the
total grain size distribution is strongly deteriorated. Here we present a new
method allowing an estimate of the deficiency in deposited mass for each
simulated class. Finally a sensitivity study on eruptive parameters is
presented in order to generalize our results to a wider range of eruptive
conditions.
| [
{
"version": "v1",
"created": "Tue, 1 Sep 2015 16:44:29 GMT"
}
] | 2015-09-02T00:00:00 | [
[
"Spanu",
"A.",
""
],
[
"Vitturi",
"M. de' Michieli",
""
],
[
"Barsotti",
"S.",
""
]
] | TITLE: Reconstructing eruptive source parameters from tephra deposit: a
numerical approach for medium-sized explosive eruptions
ABSTRACT: Since the seventies, several reconstruction techniques have been proposed,
and are currently used, to extrapolate and quantify eruptive parameters from
sampled deposit datasets. Discrete numbers of tephra ground loadings or
stratigraphic records are usually processed to estimate source eruptive values.
Reconstruction techniques like Pyle, Power law and Weibull are adopted as
standard to quantify the erupted mass (or volume) whereas Voronoi for
reconstructing the granulometry. Reconstructed values can be affected by large
uncertainty due to complexities occurring within the atmospheric dispersion and
deposition of volcanic particles. Here we want to quantify the sensitivity of
reconstruction techniques, and to quantify how much estimated values of mass
and grain size differ from emitted and deposited ones. We adopted a numerical
approach simulating with a dispersal code a mild explosive event occurring at
Mt. Etna, with eruptive parameters similar to those estimated for eruptions
occurred in the last decade. Then we created a synthetic deposit by integrating
the mass on the ground computed by the model over the computational domain
(>50000 km2). Multiple samplings of the simulated deposit are used for
generating a large dataset of sampling tests afterwards processed with standard
reconstruction techniques. Results are then compared and evaluated through a
statistical analysis, based on 2000 sampling tests of 100 samplings points. On
average, all the used techniques underestimate deposited and emitted mass. A
similar analysis, carried on Voronoi results, shows that information on the
total grain size distribution is strongly deteriorated. Here we present a new
method allowing an estimate of the deficiency in deposited mass for each
simulated class. Finally a sensitivity study on eruptive parameters is
presented in order to generalize our results to a wider range of eruptive
conditions.
| no_new_dataset | 0.950273 |
1502.05243 | Shanmuganathan Raman | Aalok Gangopadhyay, Shivam Mani Tripathi, Ishan Jindal, Shanmuganathan
Raman | SA-CNN: Dynamic Scene Classification using Convolutional Neural Networks | null | null | null | null | cs.CV | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | The task of classifying videos of natural dynamic scenes into appropriate
classes has gained lot of attention in recent years. The problem especially
becomes challenging when the camera used to capture the video is dynamic. In
this paper, we analyse the performance of statistical aggregation (SA)
techniques on various pre-trained convolutional neural network(CNN) models to
address this problem. The proposed approach works by extracting CNN activation
features for a number of frames in a video and then uses an aggregation scheme
in order to obtain a robust feature descriptor for the video. We show through
results that the proposed approach performs better than the-state-of-the arts
for the Maryland and YUPenn dataset. The final descriptor obtained is powerful
enough to distinguish among dynamic scenes and is even capable of addressing
the scenario where the camera motion is dominant and the scene dynamics are
complex. Further, this paper shows an extensive study on the performance of
various aggregation methods and their combinations. We compare the proposed
approach with other dynamic scene classification algorithms on two publicly
available datasets - Maryland and YUPenn to demonstrate the superior
performance of the proposed approach.
| [
{
"version": "v1",
"created": "Tue, 17 Feb 2015 12:25:27 GMT"
},
{
"version": "v2",
"created": "Sat, 29 Aug 2015 06:01:02 GMT"
}
] | 2015-09-01T00:00:00 | [
[
"Gangopadhyay",
"Aalok",
""
],
[
"Tripathi",
"Shivam Mani",
""
],
[
"Jindal",
"Ishan",
""
],
[
"Raman",
"Shanmuganathan",
""
]
] | TITLE: SA-CNN: Dynamic Scene Classification using Convolutional Neural Networks
ABSTRACT: The task of classifying videos of natural dynamic scenes into appropriate
classes has gained lot of attention in recent years. The problem especially
becomes challenging when the camera used to capture the video is dynamic. In
this paper, we analyse the performance of statistical aggregation (SA)
techniques on various pre-trained convolutional neural network(CNN) models to
address this problem. The proposed approach works by extracting CNN activation
features for a number of frames in a video and then uses an aggregation scheme
in order to obtain a robust feature descriptor for the video. We show through
results that the proposed approach performs better than the-state-of-the arts
for the Maryland and YUPenn dataset. The final descriptor obtained is powerful
enough to distinguish among dynamic scenes and is even capable of addressing
the scenario where the camera motion is dominant and the scene dynamics are
complex. Further, this paper shows an extensive study on the performance of
various aggregation methods and their combinations. We compare the proposed
approach with other dynamic scene classification algorithms on two publicly
available datasets - Maryland and YUPenn to demonstrate the superior
performance of the proposed approach.
| no_new_dataset | 0.951051 |
1503.06289 | Pritheega Magalingam | Pritheega Magalingam, Stephen Davis, Asha Rao | Using shortest path to discover criminal community | null | DIIN584 2015 | 10.1016/j.diin.2015.08.002 | null | cs.SI physics.soc-ph | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | Extracting communities using existing community detection algorithms yields
dense sub-networks that are difficult to analyse. Extracting a smaller sample
that embodies the relationships of a list of suspects is an important part of
the beginning of an investigation. In this paper, we present the efficacy of
our shortest paths network search algorithm (SPNSA) that begins with an
"algorithm feed", a small subset of nodes of particular interest, and builds an
investigative sub-network. The algorithm feed may consist of known criminals or
suspects, or persons of influence. This sets our approach apart from existing
community detection algorithms. We apply the SPNSA on the Enron Dataset of
e-mail communications starting with those convicted of money laundering in
relation to the collapse of Enron as the algorithm feed. The algorithm produces
sparse and small sub-networks that could feasibly identify a list of persons
and relationships to be further investigated. In contrast, we show that
identifying sub-networks of interest using either community detection
algorithms or a k-Neighbourhood approach produces sub-networks of much larger
size and complexity. When the 18 top managers of Enron were used as the
algorithm feed, the resulting sub-network identified 4 convicted criminals that
were not managers and so not part of the algorithm feed. We also directly
tested the SPNSA by removing one of the convicted criminals from the algorithm
feed and re-running the algorithm; in 5 out of 9 cases the left out criminal
occurred in the resulting sub-network.
| [
{
"version": "v1",
"created": "Sat, 21 Mar 2015 12:27:49 GMT"
}
] | 2015-09-01T00:00:00 | [
[
"Magalingam",
"Pritheega",
""
],
[
"Davis",
"Stephen",
""
],
[
"Rao",
"Asha",
""
]
] | TITLE: Using shortest path to discover criminal community
ABSTRACT: Extracting communities using existing community detection algorithms yields
dense sub-networks that are difficult to analyse. Extracting a smaller sample
that embodies the relationships of a list of suspects is an important part of
the beginning of an investigation. In this paper, we present the efficacy of
our shortest paths network search algorithm (SPNSA) that begins with an
"algorithm feed", a small subset of nodes of particular interest, and builds an
investigative sub-network. The algorithm feed may consist of known criminals or
suspects, or persons of influence. This sets our approach apart from existing
community detection algorithms. We apply the SPNSA on the Enron Dataset of
e-mail communications starting with those convicted of money laundering in
relation to the collapse of Enron as the algorithm feed. The algorithm produces
sparse and small sub-networks that could feasibly identify a list of persons
and relationships to be further investigated. In contrast, we show that
identifying sub-networks of interest using either community detection
algorithms or a k-Neighbourhood approach produces sub-networks of much larger
size and complexity. When the 18 top managers of Enron were used as the
algorithm feed, the resulting sub-network identified 4 convicted criminals that
were not managers and so not part of the algorithm feed. We also directly
tested the SPNSA by removing one of the convicted criminals from the algorithm
feed and re-running the algorithm; in 5 out of 9 cases the left out criminal
occurred in the resulting sub-network.
| no_new_dataset | 0.948106 |
1508.07551 | Karim Awudu | Awudu Karim and Shangbo Zhou | X-TREPAN: a multi class regression and adapted extraction of
comprehensible decision tree in artificial neural networks | 17 Pages, 8 Tables, 8 Figures, 6 Equations | null | null | null | cs.LG cs.NE | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | In this work, the TREPAN algorithm is enhanced and extended for extracting
decision trees from neural networks. We empirically evaluated the performance
of the algorithm on a set of databases from real world events. This benchmark
enhancement was achieved by adapting Single-test TREPAN and C4.5 decision tree
induction algorithms to analyze the datasets. The models are then compared with
X-TREPAN for comprehensibility and classification accuracy. Furthermore, we
validate the experimentations by applying statistical methods. Finally, the
modified algorithm is extended to work with multi-class regression problems and
the ability to comprehend generalized feed forward networks is achieved.
| [
{
"version": "v1",
"created": "Sun, 30 Aug 2015 10:14:48 GMT"
}
] | 2015-09-01T00:00:00 | [
[
"Karim",
"Awudu",
""
],
[
"Zhou",
"Shangbo",
""
]
] | TITLE: X-TREPAN: a multi class regression and adapted extraction of
comprehensible decision tree in artificial neural networks
ABSTRACT: In this work, the TREPAN algorithm is enhanced and extended for extracting
decision trees from neural networks. We empirically evaluated the performance
of the algorithm on a set of databases from real world events. This benchmark
enhancement was achieved by adapting Single-test TREPAN and C4.5 decision tree
induction algorithms to analyze the datasets. The models are then compared with
X-TREPAN for comprehensibility and classification accuracy. Furthermore, we
validate the experimentations by applying statistical methods. Finally, the
modified algorithm is extended to work with multi-class regression problems and
the ability to comprehend generalized feed forward networks is achieved.
| no_new_dataset | 0.949482 |
1106.2233 | Xiaowen Dong | Xiaowen Dong, Pascal Frossard, Pierre Vandergheynst and Nikolai
Nefedov | Clustering with Multi-Layer Graphs: A Spectral Perspective | null | IEEE Transactions on Signal Processing, vol. 60, no. 11, pp.
5820-5831, November 2012 | 10.1109/TSP.2012.2212886 | null | cs.LG cs.CV cs.SI stat.ML | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | Observational data usually comes with a multimodal nature, which means that
it can be naturally represented by a multi-layer graph whose layers share the
same set of vertices (users) with different edges (pairwise relationships). In
this paper, we address the problem of combining different layers of the
multi-layer graph for improved clustering of the vertices compared to using
layers independently. We propose two novel methods, which are based on joint
matrix factorization and graph regularization framework respectively, to
efficiently combine the spectrum of the multiple graph layers, namely the
eigenvectors of the graph Laplacian matrices. In each case, the resulting
combination, which we call a "joint spectrum" of multiple graphs, is used for
clustering the vertices. We evaluate our approaches by simulations with several
real world social network datasets. Results demonstrate the superior or
competitive performance of the proposed methods over state-of-the-art technique
and common baseline methods, such as co-regularization and summation of
information from individual graphs.
| [
{
"version": "v1",
"created": "Sat, 11 Jun 2011 12:43:18 GMT"
}
] | 2015-08-31T00:00:00 | [
[
"Dong",
"Xiaowen",
""
],
[
"Frossard",
"Pascal",
""
],
[
"Vandergheynst",
"Pierre",
""
],
[
"Nefedov",
"Nikolai",
""
]
] | TITLE: Clustering with Multi-Layer Graphs: A Spectral Perspective
ABSTRACT: Observational data usually comes with a multimodal nature, which means that
it can be naturally represented by a multi-layer graph whose layers share the
same set of vertices (users) with different edges (pairwise relationships). In
this paper, we address the problem of combining different layers of the
multi-layer graph for improved clustering of the vertices compared to using
layers independently. We propose two novel methods, which are based on joint
matrix factorization and graph regularization framework respectively, to
efficiently combine the spectrum of the multiple graph layers, namely the
eigenvectors of the graph Laplacian matrices. In each case, the resulting
combination, which we call a "joint spectrum" of multiple graphs, is used for
clustering the vertices. We evaluate our approaches by simulations with several
real world social network datasets. Results demonstrate the superior or
competitive performance of the proposed methods over state-of-the-art technique
and common baseline methods, such as co-regularization and summation of
information from individual graphs.
| no_new_dataset | 0.947284 |
1303.2221 | Xiaowen Dong | Xiaowen Dong, Pascal Frossard, Pierre Vandergheynst, Nikolai Nefedov | Clustering on Multi-Layer Graphs via Subspace Analysis on Grassmann
Manifolds | null | IEEE Transactions on Signal Processing, vol. 62, no. 4, pp.
905-918, February 2014 | 10.1109/TSP.2013.2295553 | null | cs.LG cs.CV cs.SI stat.ML | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | Relationships between entities in datasets are often of multiple nature, like
geographical distance, social relationships, or common interests among people
in a social network, for example. This information can naturally be modeled by
a set of weighted and undirected graphs that form a global multilayer graph,
where the common vertex set represents the entities and the edges on different
layers capture the similarities of the entities in term of the different
modalities. In this paper, we address the problem of analyzing multi-layer
graphs and propose methods for clustering the vertices by efficiently merging
the information provided by the multiple modalities. To this end, we propose to
combine the characteristics of individual graph layers using tools from
subspace analysis on a Grassmann manifold. The resulting combination can then
be viewed as a low dimensional representation of the original data which
preserves the most important information from diverse relationships between
entities. We use this information in new clustering methods and test our
algorithm on several synthetic and real world datasets where we demonstrate
superior or competitive performances compared to baseline and state-of-the-art
techniques. Our generic framework further extends to numerous analysis and
learning problems that involve different types of information on graphs.
| [
{
"version": "v1",
"created": "Sat, 9 Mar 2013 15:31:48 GMT"
}
] | 2015-08-31T00:00:00 | [
[
"Dong",
"Xiaowen",
""
],
[
"Frossard",
"Pascal",
""
],
[
"Vandergheynst",
"Pierre",
""
],
[
"Nefedov",
"Nikolai",
""
]
] | TITLE: Clustering on Multi-Layer Graphs via Subspace Analysis on Grassmann
Manifolds
ABSTRACT: Relationships between entities in datasets are often of multiple nature, like
geographical distance, social relationships, or common interests among people
in a social network, for example. This information can naturally be modeled by
a set of weighted and undirected graphs that form a global multilayer graph,
where the common vertex set represents the entities and the edges on different
layers capture the similarities of the entities in term of the different
modalities. In this paper, we address the problem of analyzing multi-layer
graphs and propose methods for clustering the vertices by efficiently merging
the information provided by the multiple modalities. To this end, we propose to
combine the characteristics of individual graph layers using tools from
subspace analysis on a Grassmann manifold. The resulting combination can then
be viewed as a low dimensional representation of the original data which
preserves the most important information from diverse relationships between
entities. We use this information in new clustering methods and test our
algorithm on several synthetic and real world datasets where we demonstrate
superior or competitive performances compared to baseline and state-of-the-art
techniques. Our generic framework further extends to numerous analysis and
learning problems that involve different types of information on graphs.
| no_new_dataset | 0.946349 |
1508.02593 | Denis Krompass | Denis Krompa{\ss} and Stephan Baier and Volker Tresp | Type-Constrained Representation Learning in Knowledge Graphs | null | null | null | null | cs.AI cs.LG | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | Large knowledge graphs increasingly add value to various applications that
require machines to recognize and understand queries and their semantics, as in
search or question answering systems. Latent variable models have increasingly
gained attention for the statistical modeling of knowledge graphs, showing
promising results in tasks related to knowledge graph completion and cleaning.
Besides storing facts about the world, schema-based knowledge graphs are backed
by rich semantic descriptions of entities and relation-types that allow
machines to understand the notion of things and their semantic relationships.
In this work, we study how type-constraints can generally support the
statistical modeling with latent variable models. More precisely, we integrated
prior knowledge in form of type-constraints in various state of the art latent
variable approaches. Our experimental results show that prior knowledge on
relation-types significantly improves these models up to 77% in link-prediction
tasks. The achieved improvements are especially prominent when a low model
complexity is enforced, a crucial requirement when these models are applied to
very large datasets. Unfortunately, type-constraints are neither always
available nor always complete e.g., they can become fuzzy when entities lack
proper typing. We show that in these cases, it can be beneficial to apply a
local closed-world assumption that approximates the semantics of relation-types
based on observations made in the data.
| [
{
"version": "v1",
"created": "Tue, 11 Aug 2015 13:49:07 GMT"
},
{
"version": "v2",
"created": "Fri, 28 Aug 2015 09:00:31 GMT"
}
] | 2015-08-31T00:00:00 | [
[
"Krompaß",
"Denis",
""
],
[
"Baier",
"Stephan",
""
],
[
"Tresp",
"Volker",
""
]
] | TITLE: Type-Constrained Representation Learning in Knowledge Graphs
ABSTRACT: Large knowledge graphs increasingly add value to various applications that
require machines to recognize and understand queries and their semantics, as in
search or question answering systems. Latent variable models have increasingly
gained attention for the statistical modeling of knowledge graphs, showing
promising results in tasks related to knowledge graph completion and cleaning.
Besides storing facts about the world, schema-based knowledge graphs are backed
by rich semantic descriptions of entities and relation-types that allow
machines to understand the notion of things and their semantic relationships.
In this work, we study how type-constraints can generally support the
statistical modeling with latent variable models. More precisely, we integrated
prior knowledge in form of type-constraints in various state of the art latent
variable approaches. Our experimental results show that prior knowledge on
relation-types significantly improves these models up to 77% in link-prediction
tasks. The achieved improvements are especially prominent when a low model
complexity is enforced, a crucial requirement when these models are applied to
very large datasets. Unfortunately, type-constraints are neither always
available nor always complete e.g., they can become fuzzy when entities lack
proper typing. We show that in these cases, it can be beneficial to apply a
local closed-world assumption that approximates the semantics of relation-types
based on observations made in the data.
| no_new_dataset | 0.944228 |
1508.07053 | Kenji Hata | Kenji Hata, Sherman Leung, Ranjay Krishna, Michael S. Bernstein, Li
Fei-Fei | SentenceRacer: A Game with a Purpose for Image Sentence Annotation | 2 pages, 2 figures, 2 tables, potential CSCW poster submission | null | null | null | cs.HC | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | Recently datasets that contain sentence descriptions of images have enabled
models that can automatically generate image captions. However, collecting
these datasets are still very expensive. Here, we present SentenceRacer, an
online game that gathers and verifies descriptions of images at no cost.
Similar to the game hangman, players compete to uncover words in a sentence
that ultimately describes an image. SentenceRacer both generates and verifies
that the sentences are accurate descriptions. We show that SentenceRacer
generates annotations of higher quality than those generated on Amazon
Mechanical Turk (AMT).
| [
{
"version": "v1",
"created": "Thu, 27 Aug 2015 23:03:17 GMT"
}
] | 2015-08-31T00:00:00 | [
[
"Hata",
"Kenji",
""
],
[
"Leung",
"Sherman",
""
],
[
"Krishna",
"Ranjay",
""
],
[
"Bernstein",
"Michael S.",
""
],
[
"Fei-Fei",
"Li",
""
]
] | TITLE: SentenceRacer: A Game with a Purpose for Image Sentence Annotation
ABSTRACT: Recently datasets that contain sentence descriptions of images have enabled
models that can automatically generate image captions. However, collecting
these datasets are still very expensive. Here, we present SentenceRacer, an
online game that gathers and verifies descriptions of images at no cost.
Similar to the game hangman, players compete to uncover words in a sentence
that ultimately describes an image. SentenceRacer both generates and verifies
that the sentences are accurate descriptions. We show that SentenceRacer
generates annotations of higher quality than those generated on Amazon
Mechanical Turk (AMT).
| no_new_dataset | 0.89974 |
1508.07148 | Thanh-Toan Do | Thanh-Toan Do, Anh-Zung Doan, Ngai-Man Cheung | Discrete Hashing with Deep Neural Network | null | null | null | null | cs.CV | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | This paper addresses the problem of learning binary hash codes for large
scale image search by proposing a novel hashing method based on deep neural
network. The advantage of our deep model over previous deep model used in
hashing is that our model contains necessary criteria for producing good codes
such as similarity preserving, balance and independence. Another advantage of
our method is that instead of relaxing the binary constraint of codes during
the learning process as most previous works, in this paper, by introducing the
auxiliary variable, we reformulate the optimization into two sub-optimization
steps allowing us to efficiently solve binary constraints without any
relaxation.
The proposed method is also extended to the supervised hashing by leveraging
the label information such that the learned binary codes preserve the pairwise
label of inputs.
The experimental results on three benchmark datasets show the proposed
methods outperform state-of-the-art hashing methods.
| [
{
"version": "v1",
"created": "Fri, 28 Aug 2015 09:38:05 GMT"
}
] | 2015-08-31T00:00:00 | [
[
"Do",
"Thanh-Toan",
""
],
[
"Doan",
"Anh-Zung",
""
],
[
"Cheung",
"Ngai-Man",
""
]
] | TITLE: Discrete Hashing with Deep Neural Network
ABSTRACT: This paper addresses the problem of learning binary hash codes for large
scale image search by proposing a novel hashing method based on deep neural
network. The advantage of our deep model over previous deep model used in
hashing is that our model contains necessary criteria for producing good codes
such as similarity preserving, balance and independence. Another advantage of
our method is that instead of relaxing the binary constraint of codes during
the learning process as most previous works, in this paper, by introducing the
auxiliary variable, we reformulate the optimization into two sub-optimization
steps allowing us to efficiently solve binary constraints without any
relaxation.
The proposed method is also extended to the supervised hashing by leveraging
the label information such that the learned binary codes preserve the pairwise
label of inputs.
The experimental results on three benchmark datasets show the proposed
methods outperform state-of-the-art hashing methods.
| no_new_dataset | 0.947527 |
1508.07275 | Luiz Capretz Dr. | Ali Bou Nassif, Mohammad Azzeh, Luiz Fernando Capretz, Danny Ho | A Comparison Between Decision Trees and Decision Tree Forest Models for
Software Development Effort Estimation | 3rd International Conference on Communications and Information
Technology (ICCIT), Beirut, Lebanon, pp. 220-224, 2013 | null | 10.1109/ICCITechnology.2013.6579553 | null | cs.SE cs.AI | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | Accurate software effort estimation has been a challenge for many software
practitioners and project managers. Underestimation leads to disruption in the
projects estimated cost and delivery. On the other hand, overestimation causes
outbidding and financial losses in business. Many software estimation models
exist; however, none have been proven to be the best in all situations. In this
paper, a decision tree forest (DTF) model is compared to a traditional decision
tree (DT) model, as well as a multiple linear regression model (MLR). The
evaluation was conducted using ISBSG and Desharnais industrial datasets.
Results show that the DTF model is competitive and can be used as an
alternative in software effort prediction.
| [
{
"version": "v1",
"created": "Fri, 28 Aug 2015 16:52:21 GMT"
}
] | 2015-08-31T00:00:00 | [
[
"Nassif",
"Ali Bou",
""
],
[
"Azzeh",
"Mohammad",
""
],
[
"Capretz",
"Luiz Fernando",
""
],
[
"Ho",
"Danny",
""
]
] | TITLE: A Comparison Between Decision Trees and Decision Tree Forest Models for
Software Development Effort Estimation
ABSTRACT: Accurate software effort estimation has been a challenge for many software
practitioners and project managers. Underestimation leads to disruption in the
projects estimated cost and delivery. On the other hand, overestimation causes
outbidding and financial losses in business. Many software estimation models
exist; however, none have been proven to be the best in all situations. In this
paper, a decision tree forest (DTF) model is compared to a traditional decision
tree (DT) model, as well as a multiple linear regression model (MLR). The
evaluation was conducted using ISBSG and Desharnais industrial datasets.
Results show that the DTF model is competitive and can be used as an
alternative in software effort prediction.
| no_new_dataset | 0.951953 |
1508.07292 | Anastasios Noulas Anastasios Noulas | Anastasios Noulas, Vsevolod Salnikov, Renaud Lambiotte, Cecilia
Mascolo | Mining Open Datasets for Transparency in Taxi Transport in Metropolitan
Environments | null | null | null | null | cs.CY | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | Uber has recently been introducing novel practices in urban taxi transport.
Journey prices can change dynamically in almost real time and also vary
geographically from one area to another in a city, a strategy known as surge
pricing. In this paper, we explore the power of the new generation of open
datasets towards understanding the impact of the new disruption technologies
that emerge in the area of public transport. With our primary goal being a more
transparent economic landscape for urban commuters, we provide a direct price
comparison between Uber and the Yellow Cab company in New York. We discover
that Uber, despite its lower standard pricing rates, effectively charges higher
fares on average, especially during short in length, but frequent in
occurrence, taxi journeys. Building on this insight, we develop a smartphone
application, OpenStreetCab, that offers a personalized consultation to mobile
users on which taxi provider is cheaper for their journey. Almost five months
after its launch, the app has attracted more than three thousand users in a
single city. Their journey queries have provided additional insights on the
potential savings similar technologies can have for urban commuters, with a
highlight being that on average, a user in New York saves 6 U.S. Dollars per
taxi journey if they pick the cheapest taxi provider. We run extensive
experiments to show how Uber's surge pricing is the driving factor of higher
journey prices and therefore higher potential savings for our application's
users. Finally, motivated by the observation that Uber's surge pricing is
occurring more frequently that intuitively expected, we formulate a prediction
task where the aim becomes to predict a geographic area's tendency to surge.
Using exogenous to Uber datasets we show how it is possible to estimate
customer demand within an area, and by extension surge pricing, with high
accuracy.
| [
{
"version": "v1",
"created": "Thu, 27 Aug 2015 16:33:52 GMT"
}
] | 2015-08-31T00:00:00 | [
[
"Noulas",
"Anastasios",
""
],
[
"Salnikov",
"Vsevolod",
""
],
[
"Lambiotte",
"Renaud",
""
],
[
"Mascolo",
"Cecilia",
""
]
] | TITLE: Mining Open Datasets for Transparency in Taxi Transport in Metropolitan
Environments
ABSTRACT: Uber has recently been introducing novel practices in urban taxi transport.
Journey prices can change dynamically in almost real time and also vary
geographically from one area to another in a city, a strategy known as surge
pricing. In this paper, we explore the power of the new generation of open
datasets towards understanding the impact of the new disruption technologies
that emerge in the area of public transport. With our primary goal being a more
transparent economic landscape for urban commuters, we provide a direct price
comparison between Uber and the Yellow Cab company in New York. We discover
that Uber, despite its lower standard pricing rates, effectively charges higher
fares on average, especially during short in length, but frequent in
occurrence, taxi journeys. Building on this insight, we develop a smartphone
application, OpenStreetCab, that offers a personalized consultation to mobile
users on which taxi provider is cheaper for their journey. Almost five months
after its launch, the app has attracted more than three thousand users in a
single city. Their journey queries have provided additional insights on the
potential savings similar technologies can have for urban commuters, with a
highlight being that on average, a user in New York saves 6 U.S. Dollars per
taxi journey if they pick the cheapest taxi provider. We run extensive
experiments to show how Uber's surge pricing is the driving factor of higher
journey prices and therefore higher potential savings for our application's
users. Finally, motivated by the observation that Uber's surge pricing is
occurring more frequently that intuitively expected, we formulate a prediction
task where the aim becomes to predict a geographic area's tendency to surge.
Using exogenous to Uber datasets we show how it is possible to estimate
customer demand within an area, and by extension surge pricing, with high
accuracy.
| no_new_dataset | 0.91957 |
1508.07306 | Ashwin Machanavajjhala | Yan Chen and Ashwin Machanavajjhala | On the Privacy Properties of Variants on the Sparse Vector Technique | 8 pages | null | null | null | cs.DB cs.CR | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | The sparse vector technique is a powerful differentially private primitive
that allows an analyst to check whether queries in a stream are greater or
lesser than a threshold. This technique has a unique property -- the algorithm
works by adding noise with a finite variance to the queries and the threshold,
and guarantees privacy that only degrades with (a) the maximum sensitivity of
any one query in stream, and (b) the number of positive answers output by the
algorithm. Recent work has developed variants of this algorithm, which we call
{\em generalized private threshold testing}, and are claimed to have privacy
guarantees that do not depend on the number of positive or negative answers
output by the algorithm. These algorithms result in a significant improvement
in utility over the sparse vector technique for a given privacy budget, and
have found applications in frequent itemset mining, feature selection in
machine learning and generating synthetic data.
In this paper we critically analyze the privacy properties of generalized
private threshold testing. We show that generalized private threshold testing
does not satisfy \epsilon-differential privacy for any finite \epsilon. We
identify a subtle error in the privacy analysis of this technique in prior
work. Moreover, we show an adversary can use generalized private threshold
testing to recover counts from the datasets (especially small counts) exactly
with high accuracy, and thus can result in individuals being reidentified. We
demonstrate our attacks empirically on real datasets.
| [
{
"version": "v1",
"created": "Fri, 28 Aug 2015 18:42:56 GMT"
}
] | 2015-08-31T00:00:00 | [
[
"Chen",
"Yan",
""
],
[
"Machanavajjhala",
"Ashwin",
""
]
] | TITLE: On the Privacy Properties of Variants on the Sparse Vector Technique
ABSTRACT: The sparse vector technique is a powerful differentially private primitive
that allows an analyst to check whether queries in a stream are greater or
lesser than a threshold. This technique has a unique property -- the algorithm
works by adding noise with a finite variance to the queries and the threshold,
and guarantees privacy that only degrades with (a) the maximum sensitivity of
any one query in stream, and (b) the number of positive answers output by the
algorithm. Recent work has developed variants of this algorithm, which we call
{\em generalized private threshold testing}, and are claimed to have privacy
guarantees that do not depend on the number of positive or negative answers
output by the algorithm. These algorithms result in a significant improvement
in utility over the sparse vector technique for a given privacy budget, and
have found applications in frequent itemset mining, feature selection in
machine learning and generating synthetic data.
In this paper we critically analyze the privacy properties of generalized
private threshold testing. We show that generalized private threshold testing
does not satisfy \epsilon-differential privacy for any finite \epsilon. We
identify a subtle error in the privacy analysis of this technique in prior
work. Moreover, we show an adversary can use generalized private threshold
testing to recover counts from the datasets (especially small counts) exactly
with high accuracy, and thus can result in individuals being reidentified. We
demonstrate our attacks empirically on real datasets.
| no_new_dataset | 0.949809 |
1502.06719 | Anamika Chhabra | Anamika Chhabra, S. R. S. Iyengar, Poonam Saini, Rajesh Shreedhar
Bhat, Vijay Kumar | Ecosystem: A Characteristic Of Crowdsourced Environments | 21 pages, 9 figures, 7 tables | null | null | null | cs.CY | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | The phenomenal success of certain crowdsourced online platforms, such as
Wikipedia, is accredited to their ability to tap the crowd's potential to
collaboratively build knowledge. While it is well known that the crowd's
collective wisdom surpasses the cumulative individual expertise, little is
understood on the dynamics of knowledge building in a crowdsourced environment.
A proper understanding of the dynamics of knowledge building in a crowdsourced
environment would enable one in the better designing of such environments to
solicit knowledge from the crowd. Our experiment on crowdsourced systems based
on annotations shows that an important reason for the rapid knowledge building
in such environments is due to variance in expertise. First, we used as our
test bed, a customized Crowdsourced Annotation System (CAS) which provides a
group of users the facility to annotate a given document while trying to
understand it. Our results showed the presence of different genres of
proficiency amongst the users of an annotation system. We observed that the
ecosystem in crowdsourced annotation system comprised of mainly four categories
of contributors, namely: Probers, Solvers, Articulators and Explorers. We
inferred from our experiment that the knowledge garnering mainly happens due to
the synergetic interaction across these categories. Further, we conducted an
analysis on the dataset of Wikipedia and Stack Overflow and noticed the
ecosystem presence in these portals as well. From this study, we claim that the
ecosystem is a universal characteristic of all crowdsourced portals.
| [
{
"version": "v1",
"created": "Tue, 24 Feb 2015 09:11:19 GMT"
},
{
"version": "v2",
"created": "Sat, 16 May 2015 09:35:01 GMT"
},
{
"version": "v3",
"created": "Tue, 11 Aug 2015 06:11:37 GMT"
},
{
"version": "v4",
"created": "Thu, 27 Aug 2015 16:46:22 GMT"
}
] | 2015-08-28T00:00:00 | [
[
"Chhabra",
"Anamika",
""
],
[
"Iyengar",
"S. R. S.",
""
],
[
"Saini",
"Poonam",
""
],
[
"Bhat",
"Rajesh Shreedhar",
""
],
[
"Kumar",
"Vijay",
""
]
] | TITLE: Ecosystem: A Characteristic Of Crowdsourced Environments
ABSTRACT: The phenomenal success of certain crowdsourced online platforms, such as
Wikipedia, is accredited to their ability to tap the crowd's potential to
collaboratively build knowledge. While it is well known that the crowd's
collective wisdom surpasses the cumulative individual expertise, little is
understood on the dynamics of knowledge building in a crowdsourced environment.
A proper understanding of the dynamics of knowledge building in a crowdsourced
environment would enable one in the better designing of such environments to
solicit knowledge from the crowd. Our experiment on crowdsourced systems based
on annotations shows that an important reason for the rapid knowledge building
in such environments is due to variance in expertise. First, we used as our
test bed, a customized Crowdsourced Annotation System (CAS) which provides a
group of users the facility to annotate a given document while trying to
understand it. Our results showed the presence of different genres of
proficiency amongst the users of an annotation system. We observed that the
ecosystem in crowdsourced annotation system comprised of mainly four categories
of contributors, namely: Probers, Solvers, Articulators and Explorers. We
inferred from our experiment that the knowledge garnering mainly happens due to
the synergetic interaction across these categories. Further, we conducted an
analysis on the dataset of Wikipedia and Stack Overflow and noticed the
ecosystem presence in these portals as well. From this study, we claim that the
ecosystem is a universal characteristic of all crowdsourced portals.
| no_new_dataset | 0.940079 |
1506.03487 | John Wieting | John Wieting, Mohit Bansal, Kevin Gimpel, Karen Livescu, and Dan Roth | From Paraphrase Database to Compositional Paraphrase Model and Back | 2015 TACL paper updated with an appendix describing new 300
dimensional embeddings. Submitted 1/2015. Accepted 2/2015. Published 6/2015 | TACL Vol 3 (2015) pg 345-358 | null | null | cs.CL | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | The Paraphrase Database (PPDB; Ganitkevitch et al., 2013) is an extensive
semantic resource, consisting of a list of phrase pairs with (heuristic)
confidence estimates. However, it is still unclear how it can best be used, due
to the heuristic nature of the confidences and its necessarily incomplete
coverage. We propose models to leverage the phrase pairs from the PPDB to build
parametric paraphrase models that score paraphrase pairs more accurately than
the PPDB's internal scores while simultaneously improving its coverage. They
allow for learning phrase embeddings as well as improved word embeddings.
Moreover, we introduce two new, manually annotated datasets to evaluate
short-phrase paraphrasing models. Using our paraphrase model trained using
PPDB, we achieve state-of-the-art results on standard word and bigram
similarity tasks and beat strong baselines on our new short phrase paraphrase
tasks.
| [
{
"version": "v1",
"created": "Wed, 10 Jun 2015 21:29:28 GMT"
},
{
"version": "v2",
"created": "Wed, 26 Aug 2015 21:18:00 GMT"
}
] | 2015-08-28T00:00:00 | [
[
"Wieting",
"John",
""
],
[
"Bansal",
"Mohit",
""
],
[
"Gimpel",
"Kevin",
""
],
[
"Livescu",
"Karen",
""
],
[
"Roth",
"Dan",
""
]
] | TITLE: From Paraphrase Database to Compositional Paraphrase Model and Back
ABSTRACT: The Paraphrase Database (PPDB; Ganitkevitch et al., 2013) is an extensive
semantic resource, consisting of a list of phrase pairs with (heuristic)
confidence estimates. However, it is still unclear how it can best be used, due
to the heuristic nature of the confidences and its necessarily incomplete
coverage. We propose models to leverage the phrase pairs from the PPDB to build
parametric paraphrase models that score paraphrase pairs more accurately than
the PPDB's internal scores while simultaneously improving its coverage. They
allow for learning phrase embeddings as well as improved word embeddings.
Moreover, we introduce two new, manually annotated datasets to evaluate
short-phrase paraphrasing models. Using our paraphrase model trained using
PPDB, we achieve state-of-the-art results on standard word and bigram
similarity tasks and beat strong baselines on our new short phrase paraphrase
tasks.
| new_dataset | 0.95877 |
1508.06708 | Sijin Li | Sijin Li, Weichen Zhang, Antoni B. Chan | Maximum-Margin Structured Learning with Deep Networks for 3D Human Pose
Estimation | null | null | null | null | cs.CV | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | This paper focuses on structured-output learning using deep neural networks
for 3D human pose estimation from monocular images. Our network takes an image
and 3D pose as inputs and outputs a score value, which is high when the
image-pose pair matches and low otherwise. The network structure consists of a
convolutional neural network for image feature extraction, followed by two
sub-networks for transforming the image features and pose into a joint
embedding. The score function is then the dot-product between the image and
pose embeddings. The image-pose embedding and score function are jointly
trained using a maximum-margin cost function. Our proposed framework can be
interpreted as a special form of structured support vector machines where the
joint feature space is discriminatively learned using deep neural networks. We
test our framework on the Human3.6m dataset and obtain state-of-the-art results
compared to other recent methods. Finally, we present visualizations of the
image-pose embedding space, demonstrating the network has learned a high-level
embedding of body-orientation and pose-configuration.
| [
{
"version": "v1",
"created": "Thu, 27 Aug 2015 03:21:15 GMT"
}
] | 2015-08-28T00:00:00 | [
[
"Li",
"Sijin",
""
],
[
"Zhang",
"Weichen",
""
],
[
"Chan",
"Antoni B.",
""
]
] | TITLE: Maximum-Margin Structured Learning with Deep Networks for 3D Human Pose
Estimation
ABSTRACT: This paper focuses on structured-output learning using deep neural networks
for 3D human pose estimation from monocular images. Our network takes an image
and 3D pose as inputs and outputs a score value, which is high when the
image-pose pair matches and low otherwise. The network structure consists of a
convolutional neural network for image feature extraction, followed by two
sub-networks for transforming the image features and pose into a joint
embedding. The score function is then the dot-product between the image and
pose embeddings. The image-pose embedding and score function are jointly
trained using a maximum-margin cost function. Our proposed framework can be
interpreted as a special form of structured support vector machines where the
joint feature space is discriminatively learned using deep neural networks. We
test our framework on the Human3.6m dataset and obtain state-of-the-art results
compared to other recent methods. Finally, we present visualizations of the
image-pose embedding space, demonstrating the network has learned a high-level
embedding of body-orientation and pose-configuration.
| no_new_dataset | 0.946745 |
1508.06878 | Kunal Bhattacharya | Kunal Bhattacharya, Asim Ghosh, Daniel Monsivais, Robin I. M. Dunbar
and Kimmo Kaski | Sex differences in social focus across the lifecycle in humans | 11 pages, 6 figures | null | null | null | physics.soc-ph cs.SI | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | Age and gender are two important factors that play crucial roles in the way
organisms allocate their social effort. In this study, we analyse a large
mobile phone dataset to explore the way lifehistory influences human sociality
and the way social networks are structured. Our results indicate that these
aspects of human behaviour are strongly related to the age and gender such that
younger individuals have more contacts and, among them, males more than
females. However, the rate of decrease in the number of contacts with age
differs between males and females, such that there is a reversal in the number
of contacts around the late 30s. We suggest that this pattern can be attributed
to the difference in reproductive investments that are made by the two sexes.
We analyse the inequality in social investment patterns and suggest that the
age and gender-related differences that we find reflect the constraints imposed
by reproduction in a context where time (a form of social capital) is limited.
| [
{
"version": "v1",
"created": "Thu, 27 Aug 2015 14:40:38 GMT"
}
] | 2015-08-28T00:00:00 | [
[
"Bhattacharya",
"Kunal",
""
],
[
"Ghosh",
"Asim",
""
],
[
"Monsivais",
"Daniel",
""
],
[
"Dunbar",
"Robin I. M.",
""
],
[
"Kaski",
"Kimmo",
""
]
] | TITLE: Sex differences in social focus across the lifecycle in humans
ABSTRACT: Age and gender are two important factors that play crucial roles in the way
organisms allocate their social effort. In this study, we analyse a large
mobile phone dataset to explore the way lifehistory influences human sociality
and the way social networks are structured. Our results indicate that these
aspects of human behaviour are strongly related to the age and gender such that
younger individuals have more contacts and, among them, males more than
females. However, the rate of decrease in the number of contacts with age
differs between males and females, such that there is a reversal in the number
of contacts around the late 30s. We suggest that this pattern can be attributed
to the difference in reproductive investments that are made by the two sexes.
We analyse the inequality in social investment patterns and suggest that the
age and gender-related differences that we find reflect the constraints imposed
by reproduction in a context where time (a form of social capital) is limited.
| no_new_dataset | 0.928668 |
1508.06976 | Byung Suk Lee | Saurav Acharya, Byung Suk Lee and Paul Hines | Real-time Top-K Predictive Query Processing over Event Streams | null | null | null | null | cs.DB cs.AI | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | This paper addresses the problem of predicting the k events that are most
likely to occur next, over historical real-time event streams. Existing
approaches to causal prediction queries have a number of limitations. First,
they exhaustively search over an acyclic causal network to find the most likely
k effect events; however, data from real event streams frequently reflect
cyclic causality. Second, they contain conservative assumptions intended to
exclude all possible non-causal links in the causal network; it leads to the
omission of many less-frequent but important causal links. We overcome these
limitations by proposing a novel event precedence model and a run-time causal
inference mechanism. The event precedence model constructs a first order
absorbing Markov chain incrementally over event streams, where an edge between
two events signifies a temporal precedence relationship between them, which is
a necessary condition for causality. Then, the run-time causal inference
mechanism learns causal relationships dynamically during query processing. This
is done by removing some of the temporal precedence relationships that do not
exhibit causality in the presence of other events in the event precedence
model. This paper presents two query processing algorithms -- one performs
exhaustive search on the model and the other performs a more efficient reduced
search with early termination. Experiments using two real datasets (cascading
blackouts in power systems and web page views) verify the effectiveness of the
probabilistic top-k prediction queries and the efficiency of the algorithms.
Specifically, the reduced search algorithm reduced runtime, relative to
exhaustive search, by 25-80% (depending on the application) with only a small
reduction in accuracy.
| [
{
"version": "v1",
"created": "Wed, 26 Aug 2015 15:02:09 GMT"
}
] | 2015-08-28T00:00:00 | [
[
"Acharya",
"Saurav",
""
],
[
"Lee",
"Byung Suk",
""
],
[
"Hines",
"Paul",
""
]
] | TITLE: Real-time Top-K Predictive Query Processing over Event Streams
ABSTRACT: This paper addresses the problem of predicting the k events that are most
likely to occur next, over historical real-time event streams. Existing
approaches to causal prediction queries have a number of limitations. First,
they exhaustively search over an acyclic causal network to find the most likely
k effect events; however, data from real event streams frequently reflect
cyclic causality. Second, they contain conservative assumptions intended to
exclude all possible non-causal links in the causal network; it leads to the
omission of many less-frequent but important causal links. We overcome these
limitations by proposing a novel event precedence model and a run-time causal
inference mechanism. The event precedence model constructs a first order
absorbing Markov chain incrementally over event streams, where an edge between
two events signifies a temporal precedence relationship between them, which is
a necessary condition for causality. Then, the run-time causal inference
mechanism learns causal relationships dynamically during query processing. This
is done by removing some of the temporal precedence relationships that do not
exhibit causality in the presence of other events in the event precedence
model. This paper presents two query processing algorithms -- one performs
exhaustive search on the model and the other performs a more efficient reduced
search with early termination. Experiments using two real datasets (cascading
blackouts in power systems and web page views) verify the effectiveness of the
probabilistic top-k prediction queries and the efficiency of the algorithms.
Specifically, the reduced search algorithm reduced runtime, relative to
exhaustive search, by 25-80% (depending on the application) with only a small
reduction in accuracy.
| no_new_dataset | 0.957517 |
1502.01199 | Reza Farrahi Moghaddam | Reza Farrahi Moghaddam and Mohamed Cheriet | A Multiple-Expert Binarization Framework for Multispectral Images | 12 pages, 8 figures, 6 tables. Presented at ICDAR'15 | null | null | null | cs.CV | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | In this work, a multiple-expert binarization framework for multispectral
images is proposed. The framework is based on a constrained subspace selection
limited to the spectral bands combined with state-of-the-art gray-level
binarization methods. The framework uses a binarization wrapper to enhance the
performance of the gray-level binarization. Nonlinear preprocessing of the
individual spectral bands is used to enhance the textual information. An
evolutionary optimizer is considered to obtain the optimal and some suboptimal
3-band subspaces from which an ensemble of experts is then formed. The
framework is applied to a ground truth multispectral dataset with promising
results. In addition, a generalization to the cross-validation approach is
developed that not only evaluates generalizability of the framework, it also
provides a practical instance of the selected experts that could be then
applied to unseen inputs despite the small size of the given ground truth
dataset.
| [
{
"version": "v1",
"created": "Wed, 4 Feb 2015 14:01:38 GMT"
},
{
"version": "v2",
"created": "Thu, 5 Feb 2015 18:56:40 GMT"
},
{
"version": "v3",
"created": "Mon, 9 Feb 2015 17:42:37 GMT"
},
{
"version": "v4",
"created": "Wed, 11 Feb 2015 18:04:21 GMT"
},
{
"version": "v5",
"created": "Mon, 13 Apr 2015 14:49:16 GMT"
},
{
"version": "v6",
"created": "Wed, 26 Aug 2015 13:27:54 GMT"
}
] | 2015-08-27T00:00:00 | [
[
"Moghaddam",
"Reza Farrahi",
""
],
[
"Cheriet",
"Mohamed",
""
]
] | TITLE: A Multiple-Expert Binarization Framework for Multispectral Images
ABSTRACT: In this work, a multiple-expert binarization framework for multispectral
images is proposed. The framework is based on a constrained subspace selection
limited to the spectral bands combined with state-of-the-art gray-level
binarization methods. The framework uses a binarization wrapper to enhance the
performance of the gray-level binarization. Nonlinear preprocessing of the
individual spectral bands is used to enhance the textual information. An
evolutionary optimizer is considered to obtain the optimal and some suboptimal
3-band subspaces from which an ensemble of experts is then formed. The
framework is applied to a ground truth multispectral dataset with promising
results. In addition, a generalization to the cross-validation approach is
developed that not only evaluates generalizability of the framework, it also
provides a practical instance of the selected experts that could be then
applied to unseen inputs despite the small size of the given ground truth
dataset.
| no_new_dataset | 0.944587 |
1508.06314 | Maher Salloum | Maher Salloum, Nathan Fabian, David M. Hensinger, Jeremy A. Templeton | Compressed Sensing and Reconstruction of Unstructured Mesh Datasets | 18 pages, 7 figures | null | null | SAND2015-4995C | cs.IT cs.DC cs.SY math.IT math.OC | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | Exascale computing promises quantities of data too large to efficiently store
and transfer across networks in order to be able to analyze and visualize the
results. We investigate Compressive Sensing (CS) as a way to reduce the size of
the data as it is being stored. CS works by sampling the data on the
computational cluster within an alternative function space such as wavelet
bases, and then reconstructing back to the original space on visualization
platforms. While much work has gone into exploring CS on structured data sets,
such as image data, we investigate its usefulness for point clouds such as
unstructured mesh datasets found in many finite element simulations. We sample
using second generation wavelets (SGW) and reconstruct using the Stagewise
Orthogonal Matching Pursuit (StOMP) algorithm. We analyze the compression
ratios achievable and quality of reconstructed results at each compression
rate. We are able to achieve compression ratios between 10 and 30 on moderate
size datasets with minimal visual deterioration as a result of the lossy
compression.
| [
{
"version": "v1",
"created": "Tue, 25 Aug 2015 21:46:30 GMT"
}
] | 2015-08-27T00:00:00 | [
[
"Salloum",
"Maher",
""
],
[
"Fabian",
"Nathan",
""
],
[
"Hensinger",
"David M.",
""
],
[
"Templeton",
"Jeremy A.",
""
]
] | TITLE: Compressed Sensing and Reconstruction of Unstructured Mesh Datasets
ABSTRACT: Exascale computing promises quantities of data too large to efficiently store
and transfer across networks in order to be able to analyze and visualize the
results. We investigate Compressive Sensing (CS) as a way to reduce the size of
the data as it is being stored. CS works by sampling the data on the
computational cluster within an alternative function space such as wavelet
bases, and then reconstructing back to the original space on visualization
platforms. While much work has gone into exploring CS on structured data sets,
such as image data, we investigate its usefulness for point clouds such as
unstructured mesh datasets found in many finite element simulations. We sample
using second generation wavelets (SGW) and reconstruct using the Stagewise
Orthogonal Matching Pursuit (StOMP) algorithm. We analyze the compression
ratios achievable and quality of reconstructed results at each compression
rate. We are able to achieve compression ratios between 10 and 30 on moderate
size datasets with minimal visual deterioration as a result of the lossy
compression.
| no_new_dataset | 0.9463 |
1508.06380 | Suman Saha | Suman Saha and Satya P. Ghrera | Network Community Detection on Metric Space | null | Algorithms 2015, 8(3), 680-696 | 10.3390/a8030680 | null | cs.SI physics.soc-ph | http://creativecommons.org/licenses/by/4.0/ | Community detection in a complex network is an important problem of much
interest in recent years. In general, a community detection algorithm chooses
an objective function and captures the communities of the network by optimizing
the objective function, and then, one uses various heuristics to solve the
optimization problem to extract the interesting communities for the user. In
this article, we demonstrate the procedure to transform a graph into points of
a metric space and develop the methods of community detection with the help of
a metric defined for a pair of points. We have also studied and analyzed the
community structure of the network therein. The results obtained with our
approach are very competitive with most of the well-known algorithms in the
literature, and this is justified over the large collection of datasets. On the
other hand, it can be observed that time taken by our algorithm is quite less
compared to other methods and justifies the theoretical findings.
| [
{
"version": "v1",
"created": "Wed, 26 Aug 2015 06:55:20 GMT"
}
] | 2015-08-27T00:00:00 | [
[
"Saha",
"Suman",
""
],
[
"Ghrera",
"Satya P.",
""
]
] | TITLE: Network Community Detection on Metric Space
ABSTRACT: Community detection in a complex network is an important problem of much
interest in recent years. In general, a community detection algorithm chooses
an objective function and captures the communities of the network by optimizing
the objective function, and then, one uses various heuristics to solve the
optimization problem to extract the interesting communities for the user. In
this article, we demonstrate the procedure to transform a graph into points of
a metric space and develop the methods of community detection with the help of
a metric defined for a pair of points. We have also studied and analyzed the
community structure of the network therein. The results obtained with our
approach are very competitive with most of the well-known algorithms in the
literature, and this is justified over the large collection of datasets. On the
other hand, it can be observed that time taken by our algorithm is quite less
compared to other methods and justifies the theoretical findings.
| no_new_dataset | 0.951323 |
1409.0080 | Wei Lu | Wei Lu, Shanshan Chen, Keqian Li, Laks V.S. Lakshmanan | Show Me the Money: Dynamic Recommendations for Revenue Maximization | Conference version published in PVLDB 7(14). To be presented in the
VLDB Conference 2015, in Hawaii. This version gives a detailed submodularity
proof | null | null | null | cs.DB cs.GT cs.IR | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | Recommender Systems (RS) play a vital role in applications such as e-commerce
and on-demand content streaming. Research on RS has mainly focused on the
customer perspective, i.e., accurate prediction of user preferences and
maximization of user utilities. As a result, most existing techniques are not
explicitly built for revenue maximization, the primary business goal of
enterprises. In this work, we explore and exploit a novel connection between RS
and the profitability of a business. As recommendations can be seen as an
information channel between a business and its customers, it is interesting and
important to investigate how to make strategic dynamic recommendations leading
to maximum possible revenue. To this end, we propose a novel \model that takes
into account a variety of factors including prices, valuations, saturation
effects, and competition amongst products. Under this model, we study the
problem of finding revenue-maximizing recommendation strategies over a finite
time horizon. We show that this problem is NP-hard, but approximation
guarantees can be obtained for a slightly relaxed version, by establishing an
elegant connection to matroid theory. Given the prohibitively high complexity
of the approximation algorithm, we also design intelligent heuristics for the
original problem. Finally, we conduct extensive experiments on two real and
synthetic datasets and demonstrate the efficiency, scalability, and
effectiveness our algorithms, and that they significantly outperform several
intuitive baselines.
| [
{
"version": "v1",
"created": "Sat, 30 Aug 2014 04:15:15 GMT"
},
{
"version": "v2",
"created": "Sat, 6 Sep 2014 01:37:15 GMT"
},
{
"version": "v3",
"created": "Tue, 25 Aug 2015 18:21:48 GMT"
}
] | 2015-08-26T00:00:00 | [
[
"Lu",
"Wei",
""
],
[
"Chen",
"Shanshan",
""
],
[
"Li",
"Keqian",
""
],
[
"Lakshmanan",
"Laks V. S.",
""
]
] | TITLE: Show Me the Money: Dynamic Recommendations for Revenue Maximization
ABSTRACT: Recommender Systems (RS) play a vital role in applications such as e-commerce
and on-demand content streaming. Research on RS has mainly focused on the
customer perspective, i.e., accurate prediction of user preferences and
maximization of user utilities. As a result, most existing techniques are not
explicitly built for revenue maximization, the primary business goal of
enterprises. In this work, we explore and exploit a novel connection between RS
and the profitability of a business. As recommendations can be seen as an
information channel between a business and its customers, it is interesting and
important to investigate how to make strategic dynamic recommendations leading
to maximum possible revenue. To this end, we propose a novel \model that takes
into account a variety of factors including prices, valuations, saturation
effects, and competition amongst products. Under this model, we study the
problem of finding revenue-maximizing recommendation strategies over a finite
time horizon. We show that this problem is NP-hard, but approximation
guarantees can be obtained for a slightly relaxed version, by establishing an
elegant connection to matroid theory. Given the prohibitively high complexity
of the approximation algorithm, we also design intelligent heuristics for the
original problem. Finally, we conduct extensive experiments on two real and
synthetic datasets and demonstrate the efficiency, scalability, and
effectiveness our algorithms, and that they significantly outperform several
intuitive baselines.
| no_new_dataset | 0.939081 |
1505.02445 | Tomaso Aste | Guido Previde Massara, T. Di Matteo, Tomaso Aste | Network Filtering for Big Data: Triangulated Maximally Filtered Graph | 16 pages, 7 Figures, 2 Tables | null | null | null | cs.DS cond-mat.stat-mech cs.IR | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | We propose a network-filtering method, the Triangulated Maximally Filtered
Graph (TMFG), that provides an approximate solution to the Weighted Maximal
Planar Graph problem. The underlying idea of TMFG consists in building a
triangulation that maximizes a score function associated with the amount of
information retained by the network. TMFG uses as weights any arbitrary
similarity measure to arrange data into a meaningful network structure that can
be used for clustering, community detection and modeling. The method is fast,
adaptable and scalable to very large datasets, it allows online updating and
learning as new data can be inserted and deleted with combinations of local and
non-local moves. TMFG permits readjustments of the network in consequence of
changes in the strength of the similarity measure. The method is based on local
topological moves and can therefore take advantage of parallel and GPUs
computing. We discuss how this network-filtering method can be used intuitively
and efficiently for big data studies and its significance from an
information-theoretic perspective.
| [
{
"version": "v1",
"created": "Sun, 10 May 2015 21:47:38 GMT"
},
{
"version": "v2",
"created": "Tue, 25 Aug 2015 16:02:37 GMT"
}
] | 2015-08-26T00:00:00 | [
[
"Massara",
"Guido Previde",
""
],
[
"Di Matteo",
"T.",
""
],
[
"Aste",
"Tomaso",
""
]
] | TITLE: Network Filtering for Big Data: Triangulated Maximally Filtered Graph
ABSTRACT: We propose a network-filtering method, the Triangulated Maximally Filtered
Graph (TMFG), that provides an approximate solution to the Weighted Maximal
Planar Graph problem. The underlying idea of TMFG consists in building a
triangulation that maximizes a score function associated with the amount of
information retained by the network. TMFG uses as weights any arbitrary
similarity measure to arrange data into a meaningful network structure that can
be used for clustering, community detection and modeling. The method is fast,
adaptable and scalable to very large datasets, it allows online updating and
learning as new data can be inserted and deleted with combinations of local and
non-local moves. TMFG permits readjustments of the network in consequence of
changes in the strength of the similarity measure. The method is based on local
topological moves and can therefore take advantage of parallel and GPUs
computing. We discuss how this network-filtering method can be used intuitively
and efficiently for big data studies and its significance from an
information-theoretic perspective.
| no_new_dataset | 0.947866 |
1508.06206 | Sahar Vahdati | Angelo Di Iorio, Christoph Lange, Anastasia Dimou, Sahar Vahdati | Semantic Publishing Challenge - Assessing the Quality of Scientific
Output by Information Extraction and Interlinking | To appear in: E. Cabrio and M. Stankovic and M. Dragoni and A.
Gangemi and R. Navigli and V. Presutti and D. Garigliotti and A. L. Gentile
and A. Nuzzolese and A. Di Iorio and A. Dimou and C. Lange and S. Vahdati and
A. Freitas and C. Unger and D. Reforgiato Recupero (eds.). Semantic Web
Evaluation Challenges 2015. Communications in Computer and Information
Science, Springer, 2015. arXiv admin note: text overlap with arXiv:1408.3863 | null | null | null | cs.DL | http://creativecommons.org/licenses/by/4.0/ | The Semantic Publishing Challenge series aims at investigating novel
approaches for improving scholarly publishing using Linked Data technology. In
2014 we had bootstrapped this effort with a focus on extracting information
from non-semantic publications - computer science workshop proceedings volumes
and their papers - to assess their quality. The objective of this second
edition was to improve information extraction but also to interlink the 2014
dataset with related ones in the LOD Cloud, thus paving the way for
sophisticated end-user services.
| [
{
"version": "v1",
"created": "Tue, 25 Aug 2015 16:17:24 GMT"
}
] | 2015-08-26T00:00:00 | [
[
"Di Iorio",
"Angelo",
""
],
[
"Lange",
"Christoph",
""
],
[
"Dimou",
"Anastasia",
""
],
[
"Vahdati",
"Sahar",
""
]
] | TITLE: Semantic Publishing Challenge - Assessing the Quality of Scientific
Output by Information Extraction and Interlinking
ABSTRACT: The Semantic Publishing Challenge series aims at investigating novel
approaches for improving scholarly publishing using Linked Data technology. In
2014 we had bootstrapped this effort with a focus on extracting information
from non-semantic publications - computer science workshop proceedings volumes
and their papers - to assess their quality. The objective of this second
edition was to improve information extraction but also to interlink the 2014
dataset with related ones in the LOD Cloud, thus paving the way for
sophisticated end-user services.
| no_new_dataset | 0.941708 |
1501.04505 | Kaihua Zhang | Kaihua Zhang, Qingshan Liu, Yi Wu, Ming-Hsuan Yang | Robust Visual Tracking via Convolutional Networks | null | null | null | null | cs.CV | http://creativecommons.org/licenses/by-nc-sa/4.0/ | Deep networks have been successfully applied to visual tracking by learning a
generic representation offline from numerous training images. However the
offline training is time-consuming and the learned generic representation may
be less discriminative for tracking specific objects. In this paper we present
that, even without offline training with a large amount of auxiliary data,
simple two-layer convolutional networks can be powerful enough to develop a
robust representation for visual tracking. In the first frame, we employ the
k-means algorithm to extract a set of normalized patches from the target region
as fixed filters, which integrate a series of adaptive contextual filters
surrounding the target to define a set of feature maps in the subsequent
frames. These maps measure similarities between each filter and the useful
local intensity patterns across the target, thereby encoding its local
structural information. Furthermore, all the maps form together a global
representation, which is built on mid-level features, thereby remaining close
to image-level information, and hence the inner geometric layout of the target
is also well preserved. A simple soft shrinkage method with an adaptive
threshold is employed to de-noise the global representation, resulting in a
robust sparse representation. The representation is updated via a simple and
effective online strategy, allowing it to robustly adapt to target appearance
variations. Our convolution networks have surprisingly lightweight structure,
yet perform favorably against several state-of-the-art methods on the CVPR2013
tracking benchmark dataset with 50 challenging videos.
| [
{
"version": "v1",
"created": "Mon, 19 Jan 2015 14:39:51 GMT"
},
{
"version": "v2",
"created": "Mon, 24 Aug 2015 06:07:22 GMT"
}
] | 2015-08-25T00:00:00 | [
[
"Zhang",
"Kaihua",
""
],
[
"Liu",
"Qingshan",
""
],
[
"Wu",
"Yi",
""
],
[
"Yang",
"Ming-Hsuan",
""
]
] | TITLE: Robust Visual Tracking via Convolutional Networks
ABSTRACT: Deep networks have been successfully applied to visual tracking by learning a
generic representation offline from numerous training images. However the
offline training is time-consuming and the learned generic representation may
be less discriminative for tracking specific objects. In this paper we present
that, even without offline training with a large amount of auxiliary data,
simple two-layer convolutional networks can be powerful enough to develop a
robust representation for visual tracking. In the first frame, we employ the
k-means algorithm to extract a set of normalized patches from the target region
as fixed filters, which integrate a series of adaptive contextual filters
surrounding the target to define a set of feature maps in the subsequent
frames. These maps measure similarities between each filter and the useful
local intensity patterns across the target, thereby encoding its local
structural information. Furthermore, all the maps form together a global
representation, which is built on mid-level features, thereby remaining close
to image-level information, and hence the inner geometric layout of the target
is also well preserved. A simple soft shrinkage method with an adaptive
threshold is employed to de-noise the global representation, resulting in a
robust sparse representation. The representation is updated via a simple and
effective online strategy, allowing it to robustly adapt to target appearance
variations. Our convolution networks have surprisingly lightweight structure,
yet perform favorably against several state-of-the-art methods on the CVPR2013
tracking benchmark dataset with 50 challenging videos.
| no_new_dataset | 0.946646 |
1506.02275 | Jacob Eisenstein | Umashanthi Pavalanathan and Jacob Eisenstein | Confounds and Consequences in Geotagged Twitter Data | final version for EMNLP 2015 | null | null | null | cs.CL | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | Twitter is often used in quantitative studies that identify
geographically-preferred topics, writing styles, and entities. These studies
rely on either GPS coordinates attached to individual messages, or on the
user-supplied location field in each profile. In this paper, we compare these
data acquisition techniques and quantify the biases that they introduce; we
also measure their effects on linguistic analysis and text-based geolocation.
GPS-tagging and self-reported locations yield measurably different corpora, and
these linguistic differences are partially attributable to differences in
dataset composition by age and gender. Using a latent variable model to induce
age and gender, we show how these demographic variables interact with geography
to affect language use. We also show that the accuracy of text-based
geolocation varies with population demographics, giving the best results for
men above the age of 40.
| [
{
"version": "v1",
"created": "Sun, 7 Jun 2015 15:29:26 GMT"
},
{
"version": "v2",
"created": "Sat, 22 Aug 2015 15:25:59 GMT"
}
] | 2015-08-25T00:00:00 | [
[
"Pavalanathan",
"Umashanthi",
""
],
[
"Eisenstein",
"Jacob",
""
]
] | TITLE: Confounds and Consequences in Geotagged Twitter Data
ABSTRACT: Twitter is often used in quantitative studies that identify
geographically-preferred topics, writing styles, and entities. These studies
rely on either GPS coordinates attached to individual messages, or on the
user-supplied location field in each profile. In this paper, we compare these
data acquisition techniques and quantify the biases that they introduce; we
also measure their effects on linguistic analysis and text-based geolocation.
GPS-tagging and self-reported locations yield measurably different corpora, and
these linguistic differences are partially attributable to differences in
dataset composition by age and gender. Using a latent variable model to induce
age and gender, we show how these demographic variables interact with geography
to affect language use. We also show that the accuracy of text-based
geolocation varies with population demographics, giving the best results for
men above the age of 40.
| no_new_dataset | 0.951188 |
1508.05710 | Zixuan Zhuang | Zixuan Zhuang | An Experimental Study of Distributed Quantile Estimation | M.S. Thesis | null | null | null | cs.DB | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | Quantiles are very important statistics information used to describe the
distribution of datasets. Given the quantiles of a dataset, we can easily know
the distribution of the dataset, which is a fundamental problem in data
analysis. However, quite often, computing quantiles directly is inappropriate
due to the memory limitations. Further, in many settings such as data streaming
and sensor network model, even the data size is unpredictable. Although the
quantiles computation has been widely studied, it was mostly in the sequential
setting. In this paper, we study several quantile computation algorithms in the
distributed setting and compare them in terms of space usage, running time, and
accuracy. Moreover, we provide detailed experimental comparisons between
several popular algorithms. Our work focuses on the approximate quantile
algorithms which provide error bounds. Approximate quantiles have received more
attentions than exact ones since they are often faster, can be more easily
adapted to the distributed setting while giving sufficiently good statistical
information on the data sets.
| [
{
"version": "v1",
"created": "Mon, 24 Aug 2015 07:49:38 GMT"
}
] | 2015-08-25T00:00:00 | [
[
"Zhuang",
"Zixuan",
""
]
] | TITLE: An Experimental Study of Distributed Quantile Estimation
ABSTRACT: Quantiles are very important statistics information used to describe the
distribution of datasets. Given the quantiles of a dataset, we can easily know
the distribution of the dataset, which is a fundamental problem in data
analysis. However, quite often, computing quantiles directly is inappropriate
due to the memory limitations. Further, in many settings such as data streaming
and sensor network model, even the data size is unpredictable. Although the
quantiles computation has been widely studied, it was mostly in the sequential
setting. In this paper, we study several quantile computation algorithms in the
distributed setting and compare them in terms of space usage, running time, and
accuracy. Moreover, we provide detailed experimental comparisons between
several popular algorithms. Our work focuses on the approximate quantile
algorithms which provide error bounds. Approximate quantiles have received more
attentions than exact ones since they are often faster, can be more easily
adapted to the distributed setting while giving sufficiently good statistical
information on the data sets.
| no_new_dataset | 0.9455 |
1508.05817 | Marco Guerini | Marco Guerini, G\"ozde \"Ozbal, Carlo Strapparava | Echoes of Persuasion: The Effect of Euphony in Persuasive Communication | null | null | null | null | cs.CL cs.CY cs.SI | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | While the effect of various lexical, syntactic, semantic and stylistic
features have been addressed in persuasive language from a computational point
of view, the persuasive effect of phonetics has received little attention. By
modeling a notion of euphony and analyzing four datasets comprising persuasive
and non-persuasive sentences in different domains (political speeches, movie
quotes, slogans and tweets), we explore the impact of sounds on different forms
of persuasiveness. We conduct a series of analyses and prediction experiments
within and across datasets. Our results highlight the positive role of phonetic
devices on persuasion.
| [
{
"version": "v1",
"created": "Mon, 24 Aug 2015 14:15:39 GMT"
}
] | 2015-08-25T00:00:00 | [
[
"Guerini",
"Marco",
""
],
[
"Özbal",
"Gözde",
""
],
[
"Strapparava",
"Carlo",
""
]
] | TITLE: Echoes of Persuasion: The Effect of Euphony in Persuasive Communication
ABSTRACT: While the effect of various lexical, syntactic, semantic and stylistic
features have been addressed in persuasive language from a computational point
of view, the persuasive effect of phonetics has received little attention. By
modeling a notion of euphony and analyzing four datasets comprising persuasive
and non-persuasive sentences in different domains (political speeches, movie
quotes, slogans and tweets), we explore the impact of sounds on different forms
of persuasiveness. We conduct a series of analyses and prediction experiments
within and across datasets. Our results highlight the positive role of phonetic
devices on persuasion.
| no_new_dataset | 0.940024 |
1406.5975 | Yogesh Simmhan | Yogesh Simmhan, Charith Wickramaarachchi, Alok Kumbhare, Marc Frincu,
Soonil Nagarkar, Santosh Ravi, Cauligi Raghavendra, Viktor Prasanna | Scalable Analytics over Distributed Time-series Graphs using GoFFish | null | Proceedings of the IEEE International Parallel and Distributed
Processing Symposium (IPDPS) (2015) pp. 809-818 | 10.1109/IPDPS.2015.66 | null | cs.DC | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | Graphs are a key form of Big Data, and performing scalable analytics over
them is invaluable to many domains. As our ability to collect data grows, there
is an emerging class of inter-connected data which accumulates or varies over
time, and on which novel analytics - both over the network structure and across
the time-variant attribute values - is necessary. We introduce the notion of
time-series graph analytics and propose Gopher, a scalable programming
abstraction to develop algorithms and analytics on such datasets. Our
abstraction leverages a sub-graph centric programming model and extends it to
the temporal dimension using an iterative BSP (Bulk Synchronous Parallel)
approach. Gopher is co-designed with GoFS, a distributed storage specialized
for time-series graphs, as part of the GoFFish distributed analytics platform.
We examine storage optimizations for GoFS, design patterns in Gopher to
leverage the distributed data layout, and evaluate the GoFFish platform using
time-series graph data and applications on a commodity cluster.
| [
{
"version": "v1",
"created": "Mon, 23 Jun 2014 16:48:03 GMT"
}
] | 2015-08-21T00:00:00 | [
[
"Simmhan",
"Yogesh",
""
],
[
"Wickramaarachchi",
"Charith",
""
],
[
"Kumbhare",
"Alok",
""
],
[
"Frincu",
"Marc",
""
],
[
"Nagarkar",
"Soonil",
""
],
[
"Ravi",
"Santosh",
""
],
[
"Raghavendra",
"Cauligi",
""
],
[
"Prasanna",
"Viktor",
""
]
] | TITLE: Scalable Analytics over Distributed Time-series Graphs using GoFFish
ABSTRACT: Graphs are a key form of Big Data, and performing scalable analytics over
them is invaluable to many domains. As our ability to collect data grows, there
is an emerging class of inter-connected data which accumulates or varies over
time, and on which novel analytics - both over the network structure and across
the time-variant attribute values - is necessary. We introduce the notion of
time-series graph analytics and propose Gopher, a scalable programming
abstraction to develop algorithms and analytics on such datasets. Our
abstraction leverages a sub-graph centric programming model and extends it to
the temporal dimension using an iterative BSP (Bulk Synchronous Parallel)
approach. Gopher is co-designed with GoFS, a distributed storage specialized
for time-series graphs, as part of the GoFFish distributed analytics platform.
We examine storage optimizations for GoFS, design patterns in Gopher to
leverage the distributed data layout, and evaluate the GoFFish platform using
time-series graph data and applications on a commodity cluster.
| no_new_dataset | 0.942454 |
1411.4046 | Mohammad Ali Keyvanrad | Mohammad Ali Keyvanrad, Mohammad Mehdi Homayounpour | Deep Belief Network Training Improvement Using Elite Samples Minimizing
Free Energy | 18 pages. arXiv admin note: substantial text overlap with
arXiv:1408.3264 | Int. J. Patt. Recogn. Artif. Intell. 29, 1551006 (2015) | 10.1142/S0218001415510064 | null | cs.LG cs.CV | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | Nowadays this is very popular to use deep architectures in machine learning.
Deep Belief Networks (DBNs) are deep architectures that use stack of Restricted
Boltzmann Machines (RBM) to create a powerful generative model using training
data. In this paper we present an improvement in a common method that is
usually used in training of RBMs. The new method uses free energy as a
criterion to obtain elite samples from generative model. We argue that these
samples can more accurately compute gradient of log probability of training
data. According to the results, an error rate of 0.99% was achieved on MNIST
test set. This result shows that the proposed method outperforms the method
presented in the first paper introducing DBN (1.25% error rate) and general
classification methods such as SVM (1.4% error rate) and KNN (with 1.6% error
rate). In another test using ISOLET dataset, letter classification error
dropped to 3.59% compared to 5.59% error rate achieved in those papers using
this dataset. The implemented method is available online at
"http://ceit.aut.ac.ir/~keyvanrad/DeeBNet Toolbox.html".
| [
{
"version": "v1",
"created": "Fri, 14 Nov 2014 16:57:48 GMT"
}
] | 2015-08-21T00:00:00 | [
[
"Keyvanrad",
"Mohammad Ali",
""
],
[
"Homayounpour",
"Mohammad Mehdi",
""
]
] | TITLE: Deep Belief Network Training Improvement Using Elite Samples Minimizing
Free Energy
ABSTRACT: Nowadays this is very popular to use deep architectures in machine learning.
Deep Belief Networks (DBNs) are deep architectures that use stack of Restricted
Boltzmann Machines (RBM) to create a powerful generative model using training
data. In this paper we present an improvement in a common method that is
usually used in training of RBMs. The new method uses free energy as a
criterion to obtain elite samples from generative model. We argue that these
samples can more accurately compute gradient of log probability of training
data. According to the results, an error rate of 0.99% was achieved on MNIST
test set. This result shows that the proposed method outperforms the method
presented in the first paper introducing DBN (1.25% error rate) and general
classification methods such as SVM (1.4% error rate) and KNN (with 1.6% error
rate). In another test using ISOLET dataset, letter classification error
dropped to 3.59% compared to 5.59% error rate achieved in those papers using
this dataset. The implemented method is available online at
"http://ceit.aut.ac.ir/~keyvanrad/DeeBNet Toolbox.html".
| no_new_dataset | 0.949576 |
1508.04785 | KuanTing Chen | KuanTing Chen, Kezhen Chen, Peizhong Cong, Winston H. Hsu, Jiebo Luo | Who are the Devils Wearing Prada in New York City? | null | null | 10.1145/2733373.2809930 | null | cs.CV cs.CY | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | Fashion is a perpetual topic in human social life, and the mass has the
penchant to emulate what large city residents and celebrities wear. Undeniably,
New York City is such a bellwether large city with all kinds of fashion
leadership. Consequently, to study what the fashion trends are during this
year, it is very helpful to learn the fashion trends of New York City.
Discovering fashion trends in New York City could boost many applications such
as clothing recommendation and advertising. Does the fashion trend in the New
York Fashion Show actually influence the clothing styles on the public? To
answer this question, we design a novel system that consists of three major
components: (1) constructing a large dataset from the New York Fashion Shows
and New York street chic in order to understand the likely clothing fashion
trends in New York, (2) utilizing a learning-based approach to discover fashion
attributes as the representative characteristics of fashion trends, and (3)
comparing the analysis results from the New York Fashion Shows and street-chic
images to verify whether the fashion shows have actual influence on the people
in New York City. Through the preliminary experiments over a large clothing
dataset, we demonstrate the effectiveness of our proposed system, and obtain
useful insights on fashion trends and fashion influence.
| [
{
"version": "v1",
"created": "Wed, 19 Aug 2015 20:28:31 GMT"
}
] | 2015-08-21T00:00:00 | [
[
"Chen",
"KuanTing",
""
],
[
"Chen",
"Kezhen",
""
],
[
"Cong",
"Peizhong",
""
],
[
"Hsu",
"Winston H.",
""
],
[
"Luo",
"Jiebo",
""
]
] | TITLE: Who are the Devils Wearing Prada in New York City?
ABSTRACT: Fashion is a perpetual topic in human social life, and the mass has the
penchant to emulate what large city residents and celebrities wear. Undeniably,
New York City is such a bellwether large city with all kinds of fashion
leadership. Consequently, to study what the fashion trends are during this
year, it is very helpful to learn the fashion trends of New York City.
Discovering fashion trends in New York City could boost many applications such
as clothing recommendation and advertising. Does the fashion trend in the New
York Fashion Show actually influence the clothing styles on the public? To
answer this question, we design a novel system that consists of three major
components: (1) constructing a large dataset from the New York Fashion Shows
and New York street chic in order to understand the likely clothing fashion
trends in New York, (2) utilizing a learning-based approach to discover fashion
attributes as the representative characteristics of fashion trends, and (3)
comparing the analysis results from the New York Fashion Shows and street-chic
images to verify whether the fashion shows have actual influence on the people
in New York City. Through the preliminary experiments over a large clothing
dataset, we demonstrate the effectiveness of our proposed system, and obtain
useful insights on fashion trends and fashion influence.
| no_new_dataset | 0.66861 |
1508.04870 | Natasha Holmes | N.G. Holmes, Carl E. Wieman, and D.A. Bonn | Teaching Critical Thinking | Proceedings of the National Academy of Sciences (2015) | null | 10.1073/pnas.1505329112 | null | physics.ed-ph | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | The ability to make decisions based on data, with its inherent uncertainties
and variability, is a complex and vital skill in the modern world. The need for
such quantitative critical thinking occurs in many different contexts, and
while it is an important goal of education, that goal is seldom being achieved.
We argue that the key element for developing this ability is repeated practice
in making decisions based on data, with feedback on those decisions. We
demonstrate a structure for providing suitable practice that can be applied in
any instructional setting that involves the acquisition of data and relating
that data to scientific models. This study reports the results of applying that
structure in an introductory physics lab course. Students in an experimental
condition were repeatedly instructed to make and act on quantitative
comparisons between datasets, and between data and models, an approach that is
common to all science disciplines. These instructions were slowly faded across
the course. After the instructions had been removed, students in the
experimental condition were 12 times more likely to spontaneously propose or
make changes to improve their experimental methods than a control group, who
performed traditional experimental activities. They were also four times more
likely to identify and explain a limitation of a physical model using their
data. Students in the experimental condition also showed much more
sophisticated reasoning about their data. These differences between the groups
were seen to persist into a subsequent course taken the following year.
| [
{
"version": "v1",
"created": "Thu, 20 Aug 2015 03:51:42 GMT"
}
] | 2015-08-21T00:00:00 | [
[
"Holmes",
"N. G.",
""
],
[
"Wieman",
"Carl E.",
""
],
[
"Bonn",
"D. A.",
""
]
] | TITLE: Teaching Critical Thinking
ABSTRACT: The ability to make decisions based on data, with its inherent uncertainties
and variability, is a complex and vital skill in the modern world. The need for
such quantitative critical thinking occurs in many different contexts, and
while it is an important goal of education, that goal is seldom being achieved.
We argue that the key element for developing this ability is repeated practice
in making decisions based on data, with feedback on those decisions. We
demonstrate a structure for providing suitable practice that can be applied in
any instructional setting that involves the acquisition of data and relating
that data to scientific models. This study reports the results of applying that
structure in an introductory physics lab course. Students in an experimental
condition were repeatedly instructed to make and act on quantitative
comparisons between datasets, and between data and models, an approach that is
common to all science disciplines. These instructions were slowly faded across
the course. After the instructions had been removed, students in the
experimental condition were 12 times more likely to spontaneously propose or
make changes to improve their experimental methods than a control group, who
performed traditional experimental activities. They were also four times more
likely to identify and explain a limitation of a physical model using their
data. Students in the experimental condition also showed much more
sophisticated reasoning about their data. These differences between the groups
were seen to persist into a subsequent course taken the following year.
| no_new_dataset | 0.942082 |
1508.04909 | Alain Rakotomamonjy | Alain Rakotomamonjy (LITIS), Gilles Gasso (LITIS) | Histogram of gradients of Time-Frequency Representations for Audio scene
detection | null | null | null | null | cs.SD cs.LG | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | This paper addresses the problem of audio scenes classification and
contributes to the state of the art by proposing a novel feature. We build this
feature by considering histogram of gradients (HOG) of time-frequency
representation of an audio scene. Contrarily to classical audio features like
MFCC, we make the hypothesis that histogram of gradients are able to encode
some relevant informations in a time-frequency {representation:} namely, the
local direction of variation (in time and frequency) of the signal spectral
power. In addition, in order to gain more invariance and robustness, histogram
of gradients are locally pooled. We have evaluated the relevance of {the novel
feature} by comparing its performances with state-of-the-art competitors, on
several datasets, including a novel one that we provide, as part of our
contribution. This dataset, that we make publicly available, involves $19$
classes and contains about $900$ minutes of audio scene recording. We thus
believe that it may be the next standard dataset for evaluating audio scene
classification algorithms. Our comparison results clearly show that our
HOG-based features outperform its competitors
| [
{
"version": "v1",
"created": "Thu, 20 Aug 2015 08:07:10 GMT"
}
] | 2015-08-21T00:00:00 | [
[
"Rakotomamonjy",
"Alain",
"",
"LITIS"
],
[
"Gasso",
"Gilles",
"",
"LITIS"
]
] | TITLE: Histogram of gradients of Time-Frequency Representations for Audio scene
detection
ABSTRACT: This paper addresses the problem of audio scenes classification and
contributes to the state of the art by proposing a novel feature. We build this
feature by considering histogram of gradients (HOG) of time-frequency
representation of an audio scene. Contrarily to classical audio features like
MFCC, we make the hypothesis that histogram of gradients are able to encode
some relevant informations in a time-frequency {representation:} namely, the
local direction of variation (in time and frequency) of the signal spectral
power. In addition, in order to gain more invariance and robustness, histogram
of gradients are locally pooled. We have evaluated the relevance of {the novel
feature} by comparing its performances with state-of-the-art competitors, on
several datasets, including a novel one that we provide, as part of our
contribution. This dataset, that we make publicly available, involves $19$
classes and contains about $900$ minutes of audio scene recording. We thus
believe that it may be the next standard dataset for evaluating audio scene
classification algorithms. Our comparison results clearly show that our
HOG-based features outperform its competitors
| new_dataset | 0.964119 |
1508.04957 | Aggeliki Dimitriou | Aggeliki Dimitriou, Ananya Dass, Dimitri Theodoratos | Cohesiveness Relationships to Empower Keyword Search on Tree Data on the
Web | null | null | null | null | cs.DB | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | Keyword search is the most popular querying technique on semistructured data.
Keyword queries are simple and con- venient. However, as a consequence of their
imprecision, the quality of their answers is poor and the existing algorithms
do not scale satisfactorily.
In this paper, we introduce the novel concept of cohesive keyword queries for
tree data. Intuitively, a cohesiveness relationship on keywords indicates that
they should form a cohesive whole in a query result. Cohesive keyword queries
allow term nesting and keyword repetition. Although more expressive, they are
as simple as flat keyword queries. We provide formal semantics for cohesive
keyword queries rank- ing query results on the proximity of the keyword
instances. We design a stack based algorithm which efficiently evaluates
cohesive keyword queries. Our experiments demonstrate that our approach
outperforms in quality previous filtering semantics and our algorithm scales
smoothly on queries of even 20 keywords on large datasets.
| [
{
"version": "v1",
"created": "Thu, 20 Aug 2015 11:17:04 GMT"
}
] | 2015-08-21T00:00:00 | [
[
"Dimitriou",
"Aggeliki",
""
],
[
"Dass",
"Ananya",
""
],
[
"Theodoratos",
"Dimitri",
""
]
] | TITLE: Cohesiveness Relationships to Empower Keyword Search on Tree Data on the
Web
ABSTRACT: Keyword search is the most popular querying technique on semistructured data.
Keyword queries are simple and con- venient. However, as a consequence of their
imprecision, the quality of their answers is poor and the existing algorithms
do not scale satisfactorily.
In this paper, we introduce the novel concept of cohesive keyword queries for
tree data. Intuitively, a cohesiveness relationship on keywords indicates that
they should form a cohesive whole in a query result. Cohesive keyword queries
allow term nesting and keyword repetition. Although more expressive, they are
as simple as flat keyword queries. We provide formal semantics for cohesive
keyword queries rank- ing query results on the proximity of the keyword
instances. We design a stack based algorithm which efficiently evaluates
cohesive keyword queries. Our experiments demonstrate that our approach
outperforms in quality previous filtering semantics and our algorithm scales
smoothly on queries of even 20 keywords on large datasets.
| no_new_dataset | 0.951278 |
1508.05003 | Suvrit Sra | Suvrit Sra, Adams Wei Yu, Mu Li, Alexander J. Smola | AdaDelay: Delay Adaptive Distributed Stochastic Convex Optimization | 19 pages | null | null | null | stat.ML cs.LG math.OC | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | We study distributed stochastic convex optimization under the delayed
gradient model where the server nodes perform parameter updates, while the
worker nodes compute stochastic gradients. We discuss, analyze, and experiment
with a setup motivated by the behavior of real-world distributed computation
networks, where the machines are differently slow at different time. Therefore,
we allow the parameter updates to be sensitive to the actual delays
experienced, rather than to worst-case bounds on the maximum delay. This
sensitivity leads to larger stepsizes, that can help gain rapid initial
convergence without having to wait too long for slower machines, while
maintaining the same asymptotic complexity. We obtain encouraging improvements
to overall convergence for distributed experiments on real datasets with up to
billions of examples and features.
| [
{
"version": "v1",
"created": "Thu, 20 Aug 2015 15:11:11 GMT"
}
] | 2015-08-21T00:00:00 | [
[
"Sra",
"Suvrit",
""
],
[
"Yu",
"Adams Wei",
""
],
[
"Li",
"Mu",
""
],
[
"Smola",
"Alexander J.",
""
]
] | TITLE: AdaDelay: Delay Adaptive Distributed Stochastic Convex Optimization
ABSTRACT: We study distributed stochastic convex optimization under the delayed
gradient model where the server nodes perform parameter updates, while the
worker nodes compute stochastic gradients. We discuss, analyze, and experiment
with a setup motivated by the behavior of real-world distributed computation
networks, where the machines are differently slow at different time. Therefore,
we allow the parameter updates to be sensitive to the actual delays
experienced, rather than to worst-case bounds on the maximum delay. This
sensitivity leads to larger stepsizes, that can help gain rapid initial
convergence without having to wait too long for slower machines, while
maintaining the same asymptotic complexity. We obtain encouraging improvements
to overall convergence for distributed experiments on real datasets with up to
billions of examples and features.
| no_new_dataset | 0.946051 |
1508.04525 | Wei Zhang | Wei Zhang, Yang Yu, Osho Gupta, Judith Gelernter | Recognizing Extended Spatiotemporal Expressions by Actively Trained
Average Perceptron Ensembles | 10 pages | null | null | null | cs.CL cs.LG | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | Precise geocoding and time normalization for text requires that location and
time phrases be identified. Many state-of-the-art geoparsers and temporal
parsers suffer from low recall. Categories commonly missed by parsers are:
nouns used in a non- spatiotemporal sense, adjectival and adverbial phrases,
prepositional phrases, and numerical phrases. We collected and annotated data
set by querying commercial web searches API with such spatiotemporal
expressions as were missed by state-of-the- art parsers. Due to the high cost
of sentence annotation, active learning was used to label training data, and a
new strategy was designed to better select training examples to reduce labeling
cost. For the learning algorithm, we applied an average perceptron trained
Featurized Hidden Markov Model (FHMM). Five FHMM instances were used to create
an ensemble, with the output phrase selected by voting. Our ensemble model was
tested on a range of sequential labeling tasks, and has shown competitive
performance. Our contributions include (1) an new dataset annotated with named
entities and expanded spatiotemporal expressions; (2) a comparison of inference
algorithms for ensemble models showing the superior accuracy of Belief
Propagation over Viterbi Decoding; (3) a new example re-weighting method for
active ensemble learning that 'memorizes' the latest examples trained; (4) a
spatiotemporal parser that jointly recognizes expanded spatiotemporal
expressions as well as named entities.
| [
{
"version": "v1",
"created": "Wed, 19 Aug 2015 04:17:47 GMT"
}
] | 2015-08-20T00:00:00 | [
[
"Zhang",
"Wei",
""
],
[
"Yu",
"Yang",
""
],
[
"Gupta",
"Osho",
""
],
[
"Gelernter",
"Judith",
""
]
] | TITLE: Recognizing Extended Spatiotemporal Expressions by Actively Trained
Average Perceptron Ensembles
ABSTRACT: Precise geocoding and time normalization for text requires that location and
time phrases be identified. Many state-of-the-art geoparsers and temporal
parsers suffer from low recall. Categories commonly missed by parsers are:
nouns used in a non- spatiotemporal sense, adjectival and adverbial phrases,
prepositional phrases, and numerical phrases. We collected and annotated data
set by querying commercial web searches API with such spatiotemporal
expressions as were missed by state-of-the- art parsers. Due to the high cost
of sentence annotation, active learning was used to label training data, and a
new strategy was designed to better select training examples to reduce labeling
cost. For the learning algorithm, we applied an average perceptron trained
Featurized Hidden Markov Model (FHMM). Five FHMM instances were used to create
an ensemble, with the output phrase selected by voting. Our ensemble model was
tested on a range of sequential labeling tasks, and has shown competitive
performance. Our contributions include (1) an new dataset annotated with named
entities and expanded spatiotemporal expressions; (2) a comparison of inference
algorithms for ensemble models showing the superior accuracy of Belief
Propagation over Viterbi Decoding; (3) a new example re-weighting method for
active ensemble learning that 'memorizes' the latest examples trained; (4) a
spatiotemporal parser that jointly recognizes expanded spatiotemporal
expressions as well as named entities.
| no_new_dataset | 0.559079 |
1508.04537 | Jun He | Hao Wu, Jun He, Bo Li, Yijian Pei | Personalized QoS Prediction of Cloud Services via Learning
Neighborhood-based Model | null | null | null | null | cs.DC cs.PF | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | The explosion of cloud services on the Internet brings new challenges in
service discovery and selection. Particularly, the demand for efficient
quality-of-service (QoS) evaluation is becoming urgently strong. To address
this issue, this paper proposes neighborhood-based approach for QoS prediction
of cloud services by taking advantages of collaborative intelligence. Different
from heuristic collaborative filtering and matrix factorization, we define a
formal neighborhood-based prediction framework which allows an efficient global
optimization scheme, and then exploit different baseline estimate component to
improve predictive performance. To validate the proposed methods, a large-scale
QoS-specific dataset which consists of invocation records from 339 service
users on 5,825 web services on a world-scale distributed network is used.
Experimental results demonstrate that the learned neighborhood-based models can
overcome existing difficulties of heuristic collaborative filtering methods and
achieve superior performance than state-of-the-art prediction methods.
| [
{
"version": "v1",
"created": "Wed, 19 Aug 2015 06:32:54 GMT"
}
] | 2015-08-20T00:00:00 | [
[
"Wu",
"Hao",
""
],
[
"He",
"Jun",
""
],
[
"Li",
"Bo",
""
],
[
"Pei",
"Yijian",
""
]
] | TITLE: Personalized QoS Prediction of Cloud Services via Learning
Neighborhood-based Model
ABSTRACT: The explosion of cloud services on the Internet brings new challenges in
service discovery and selection. Particularly, the demand for efficient
quality-of-service (QoS) evaluation is becoming urgently strong. To address
this issue, this paper proposes neighborhood-based approach for QoS prediction
of cloud services by taking advantages of collaborative intelligence. Different
from heuristic collaborative filtering and matrix factorization, we define a
formal neighborhood-based prediction framework which allows an efficient global
optimization scheme, and then exploit different baseline estimate component to
improve predictive performance. To validate the proposed methods, a large-scale
QoS-specific dataset which consists of invocation records from 339 service
users on 5,825 web services on a world-scale distributed network is used.
Experimental results demonstrate that the learned neighborhood-based models can
overcome existing difficulties of heuristic collaborative filtering methods and
achieve superior performance than state-of-the-art prediction methods.
| no_new_dataset | 0.802633 |
1508.04546 | Alexander Krull | Alexander Krull, Eric Brachmann, Frank Michel, Michael Ying Yang,
Stefan Gumhold, Carsten Rother | Learning Analysis-by-Synthesis for 6D Pose Estimation in RGB-D Images | 16 pages, 8 figures | null | null | null | cs.CV | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | Analysis-by-synthesis has been a successful approach for many tasks in
computer vision, such as 6D pose estimation of an object in an RGB-D image
which is the topic of this work. The idea is to compare the observation with
the output of a forward process, such as a rendered image of the object of
interest in a particular pose. Due to occlusion or complicated sensor noise, it
can be difficult to perform this comparison in a meaningful way. We propose an
approach that "learns to compare", while taking these difficulties into
account. This is done by describing the posterior density of a particular
object pose with a convolutional neural network (CNN) that compares an observed
and rendered image. The network is trained with the maximum likelihood
paradigm. We observe empirically that the CNN does not specialize to the
geometry or appearance of specific objects, and it can be used with objects of
vastly different shapes and appearances, and in different backgrounds. Compared
to state-of-the-art, we demonstrate a significant improvement on two different
datasets which include a total of eleven objects, cluttered background, and
heavy occlusion.
| [
{
"version": "v1",
"created": "Wed, 19 Aug 2015 07:24:14 GMT"
}
] | 2015-08-20T00:00:00 | [
[
"Krull",
"Alexander",
""
],
[
"Brachmann",
"Eric",
""
],
[
"Michel",
"Frank",
""
],
[
"Yang",
"Michael Ying",
""
],
[
"Gumhold",
"Stefan",
""
],
[
"Rother",
"Carsten",
""
]
] | TITLE: Learning Analysis-by-Synthesis for 6D Pose Estimation in RGB-D Images
ABSTRACT: Analysis-by-synthesis has been a successful approach for many tasks in
computer vision, such as 6D pose estimation of an object in an RGB-D image
which is the topic of this work. The idea is to compare the observation with
the output of a forward process, such as a rendered image of the object of
interest in a particular pose. Due to occlusion or complicated sensor noise, it
can be difficult to perform this comparison in a meaningful way. We propose an
approach that "learns to compare", while taking these difficulties into
account. This is done by describing the posterior density of a particular
object pose with a convolutional neural network (CNN) that compares an observed
and rendered image. The network is trained with the maximum likelihood
paradigm. We observe empirically that the CNN does not specialize to the
geometry or appearance of specific objects, and it can be used with objects of
vastly different shapes and appearances, and in different backgrounds. Compared
to state-of-the-art, we demonstrate a significant improvement on two different
datasets which include a total of eleven objects, cluttered background, and
heavy occlusion.
| no_new_dataset | 0.949106 |
1508.04586 | Veronica Vilaplana | Ver\'onica Vilaplana | Saliency maps on image hierarchies | Accepted for publication in Signal Processing: Image Communications,
2015 | null | 10.1016/j.image.2015.07.012 | null | cs.CV | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | In this paper we propose two saliency models for salient object segmentation
based on a hierarchical image segmentation, a tree-like structure that
represents regions at different scales from the details to the whole image
(e.g. gPb-UCM, BPT). The first model is based on a hierarchy of image
partitions. The saliency at each level is computed on a region basis, taking
into account the contrast between regions. The maps obtained for the different
partitions are then integrated into a final saliency map. The second model
directly works on the structure created by the segmentation algorithm,
computing saliency at each node and integrating these cues in a straightforward
manner into a single saliency map. We show that the proposed models produce
high quality saliency maps. Objective evaluation demonstrates that the two
methods achieve state-of-the-art performance in several benchmark datasets.
| [
{
"version": "v1",
"created": "Wed, 19 Aug 2015 10:07:07 GMT"
}
] | 2015-08-20T00:00:00 | [
[
"Vilaplana",
"Verónica",
""
]
] | TITLE: Saliency maps on image hierarchies
ABSTRACT: In this paper we propose two saliency models for salient object segmentation
based on a hierarchical image segmentation, a tree-like structure that
represents regions at different scales from the details to the whole image
(e.g. gPb-UCM, BPT). The first model is based on a hierarchy of image
partitions. The saliency at each level is computed on a region basis, taking
into account the contrast between regions. The maps obtained for the different
partitions are then integrated into a final saliency map. The second model
directly works on the structure created by the segmentation algorithm,
computing saliency at each node and integrating these cues in a straightforward
manner into a single saliency map. We show that the proposed models produce
high quality saliency maps. Objective evaluation demonstrates that the two
methods achieve state-of-the-art performance in several benchmark datasets.
| no_new_dataset | 0.956145 |
1410.8616 | Abhijit Chandra | Abhijit Chandra and Oliva Kar | Data Driven Prognosis: A multi-physics approach verified via balloon
burst experiment | null | null | 10.1098/rspa.2014.0525 | null | cs.CE | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | A multi-physics formulation for Data Driven Prognosis (DDP) is developed.
Unlike traditional predictive strategies that require controlled off-line
measurements or training for determination of constitutive parameters to derive
the transitional statistics, the proposed DDP algorithm relies solely on in
situ measurements. It utilizes a deterministic mechanics framework, but the
stochastic nature of the solution arises naturally from the underlying
assumptions regarding the order of the conservation potential as well as the
number of dimensions involved. The proposed DDP scheme is capable of predicting
onset of instabilities. Since the need for off-line testing (or training) is
obviated, it can be easily implemented for systems where such a priori testing
is difficult or even impossible to conduct. The prognosis capability is
demonstrated here via a balloon burst experiment where the instability is
predicted utilizing only on-line visual observations. The DDP scheme never
failed to predict the incipient failure, and no false positives were issued.
The DDP algorithm is applicable to others types of datasets. Time horizons of
DDP predictions can be adjusted by using memory over different time windows.
Thus, a big dataset can be parsed in time to make a range of predictions over
varying time horizons.
| [
{
"version": "v1",
"created": "Fri, 31 Oct 2014 02:05:09 GMT"
}
] | 2015-08-19T00:00:00 | [
[
"Chandra",
"Abhijit",
""
],
[
"Kar",
"Oliva",
""
]
] | TITLE: Data Driven Prognosis: A multi-physics approach verified via balloon
burst experiment
ABSTRACT: A multi-physics formulation for Data Driven Prognosis (DDP) is developed.
Unlike traditional predictive strategies that require controlled off-line
measurements or training for determination of constitutive parameters to derive
the transitional statistics, the proposed DDP algorithm relies solely on in
situ measurements. It utilizes a deterministic mechanics framework, but the
stochastic nature of the solution arises naturally from the underlying
assumptions regarding the order of the conservation potential as well as the
number of dimensions involved. The proposed DDP scheme is capable of predicting
onset of instabilities. Since the need for off-line testing (or training) is
obviated, it can be easily implemented for systems where such a priori testing
is difficult or even impossible to conduct. The prognosis capability is
demonstrated here via a balloon burst experiment where the instability is
predicted utilizing only on-line visual observations. The DDP scheme never
failed to predict the incipient failure, and no false positives were issued.
The DDP algorithm is applicable to others types of datasets. Time horizons of
DDP predictions can be adjusted by using memory over different time windows.
Thus, a big dataset can be parsed in time to make a range of predictions over
varying time horizons.
| no_new_dataset | 0.949295 |
1412.5808 | Johannes Niedermayer | Johannes Niedermayer, Peer Kr\"oger | Minimizing the Number of Matching Queries for Object Retrieval | null | null | null | null | cs.CV | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | To increase the computational efficiency of interest-point based object
retrieval, researchers have put remarkable research efforts into improving the
efficiency of kNN-based feature matching, pursuing to match thousands of
features against a database within fractions of a second. However, due to the
high-dimensional nature of image features that reduces the effectivity of index
structures (curse of dimensionality), due to the vast amount of features stored
in image databases (images are often represented by up to several thousand
features), this ultimate goal demanded to trade query runtimes for query
precision. In this paper we address an approach complementary to indexing in
order to improve the runtimes of retrieval by querying only the most promising
keypoint descriptors, as this affects matching runtimes linearly and can
therefore lead to increased efficiency. As this reduction of kNN queries
reduces the number of tentative correspondences, a loss of query precision is
minimized by an additional image-level correspondence generation stage with a
computational performance independent of the underlying indexing structure. We
evaluate such an adaption of the standard recognition pipeline on a variety of
datasets using both SIFT and state-of-the-art binary descriptors. Our results
suggest that decreasing the number of queried descriptors does not necessarily
imply a reduction in the result quality as long as alternative ways of
increasing query recall (by thoroughly selecting k) and MAP (using image-level
correspondence generation) are considered.
| [
{
"version": "v1",
"created": "Thu, 18 Dec 2014 11:20:39 GMT"
},
{
"version": "v2",
"created": "Mon, 30 Mar 2015 10:10:14 GMT"
},
{
"version": "v3",
"created": "Tue, 18 Aug 2015 07:08:05 GMT"
}
] | 2015-08-19T00:00:00 | [
[
"Niedermayer",
"Johannes",
""
],
[
"Kröger",
"Peer",
""
]
] | TITLE: Minimizing the Number of Matching Queries for Object Retrieval
ABSTRACT: To increase the computational efficiency of interest-point based object
retrieval, researchers have put remarkable research efforts into improving the
efficiency of kNN-based feature matching, pursuing to match thousands of
features against a database within fractions of a second. However, due to the
high-dimensional nature of image features that reduces the effectivity of index
structures (curse of dimensionality), due to the vast amount of features stored
in image databases (images are often represented by up to several thousand
features), this ultimate goal demanded to trade query runtimes for query
precision. In this paper we address an approach complementary to indexing in
order to improve the runtimes of retrieval by querying only the most promising
keypoint descriptors, as this affects matching runtimes linearly and can
therefore lead to increased efficiency. As this reduction of kNN queries
reduces the number of tentative correspondences, a loss of query precision is
minimized by an additional image-level correspondence generation stage with a
computational performance independent of the underlying indexing structure. We
evaluate such an adaption of the standard recognition pipeline on a variety of
datasets using both SIFT and state-of-the-art binary descriptors. Our results
suggest that decreasing the number of queried descriptors does not necessarily
imply a reduction in the result quality as long as alternative ways of
increasing query recall (by thoroughly selecting k) and MAP (using image-level
correspondence generation) are considered.
| no_new_dataset | 0.950686 |
1506.07862 | Mariusz Tarnopolski | Mariusz Tarnopolski | On the limit between short and long GRBs | 6 pages, 1 figure; matches the version to published in Ap&SS | null | 10.1007/s10509-015-2473-6 | null | astro-ph.HE physics.space-ph | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | Two classes of GRBs have been identified thus far without doubt and are
prescribed to different physical scenarios -- NS-NS or NS-BH mergers, and
collapse of massive stars, for short and long GRBs, respectively. The existence
of two distinct populations was inferred through a bimodal distribution of the
observed durations $T_{90}$, and the commonly applied $2\,{\rm s}$ limit
between short and long GRBs was obtained by fitting a parabola between the two
peaks in binned data from BATSE 1B. Herein, by means of a maximum likelihood
(ML) method a mixture of two Gaussians is fitted to the datasets from BATSE,
$Swift$, $BeppoSAX$, and $Fermi$ in search for a local minimum that might serve
as a new, more proper, limit for the two GRB classes. It is found that $Swift$
and $BeppoSAX$ distributions are unimodal, hence no local minimum is present,
$Fermi$ is consistent with the conventional limit, whereas BATSE gives the
limit significantly longer (equal to $3.38\pm 0.27\,{\rm s}$) than $2\,{\rm
s}$. These new values change the fractions of short and long GRBs in the
samples examined, and imply that the observed $T_{90}$ durations are detector
dependent, hence no universal limiting value may be applied to all satellites
due to their different instrument specifications. Because of this, and due to
the strong overlap of the two-Gaussian components, the straightforward
association of short GRBs to mergers and long ones to collapsars is ambiguous.
| [
{
"version": "v1",
"created": "Thu, 25 Jun 2015 19:29:43 GMT"
},
{
"version": "v2",
"created": "Thu, 30 Jul 2015 20:15:13 GMT"
}
] | 2015-08-19T00:00:00 | [
[
"Tarnopolski",
"Mariusz",
""
]
] | TITLE: On the limit between short and long GRBs
ABSTRACT: Two classes of GRBs have been identified thus far without doubt and are
prescribed to different physical scenarios -- NS-NS or NS-BH mergers, and
collapse of massive stars, for short and long GRBs, respectively. The existence
of two distinct populations was inferred through a bimodal distribution of the
observed durations $T_{90}$, and the commonly applied $2\,{\rm s}$ limit
between short and long GRBs was obtained by fitting a parabola between the two
peaks in binned data from BATSE 1B. Herein, by means of a maximum likelihood
(ML) method a mixture of two Gaussians is fitted to the datasets from BATSE,
$Swift$, $BeppoSAX$, and $Fermi$ in search for a local minimum that might serve
as a new, more proper, limit for the two GRB classes. It is found that $Swift$
and $BeppoSAX$ distributions are unimodal, hence no local minimum is present,
$Fermi$ is consistent with the conventional limit, whereas BATSE gives the
limit significantly longer (equal to $3.38\pm 0.27\,{\rm s}$) than $2\,{\rm
s}$. These new values change the fractions of short and long GRBs in the
samples examined, and imply that the observed $T_{90}$ durations are detector
dependent, hence no universal limiting value may be applied to all satellites
due to their different instrument specifications. Because of this, and due to
the strong overlap of the two-Gaussian components, the straightforward
association of short GRBs to mergers and long ones to collapsars is ambiguous.
| no_new_dataset | 0.945298 |
1507.06821 | Andreas Eitel | Andreas Eitel, Jost Tobias Springenberg, Luciano Spinello, Martin
Riedmiller, Wolfram Burgard | Multimodal Deep Learning for Robust RGB-D Object Recognition | Final version submitted to IROS'2015, results unchanged,
reformulation of some text passages in abstract and introduction | null | null | null | cs.CV cs.LG cs.NE cs.RO | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | Robust object recognition is a crucial ingredient of many, if not all,
real-world robotics applications. This paper leverages recent progress on
Convolutional Neural Networks (CNNs) and proposes a novel RGB-D architecture
for object recognition. Our architecture is composed of two separate CNN
processing streams - one for each modality - which are consecutively combined
with a late fusion network. We focus on learning with imperfect sensor data, a
typical problem in real-world robotics tasks. For accurate learning, we
introduce a multi-stage training methodology and two crucial ingredients for
handling depth data with CNNs. The first, an effective encoding of depth
information for CNNs that enables learning without the need for large depth
datasets. The second, a data augmentation scheme for robust learning with depth
images by corrupting them with realistic noise patterns. We present
state-of-the-art results on the RGB-D object dataset and show recognition in
challenging RGB-D real-world noisy settings.
| [
{
"version": "v1",
"created": "Fri, 24 Jul 2015 12:20:19 GMT"
},
{
"version": "v2",
"created": "Tue, 18 Aug 2015 13:04:29 GMT"
}
] | 2015-08-19T00:00:00 | [
[
"Eitel",
"Andreas",
""
],
[
"Springenberg",
"Jost Tobias",
""
],
[
"Spinello",
"Luciano",
""
],
[
"Riedmiller",
"Martin",
""
],
[
"Burgard",
"Wolfram",
""
]
] | TITLE: Multimodal Deep Learning for Robust RGB-D Object Recognition
ABSTRACT: Robust object recognition is a crucial ingredient of many, if not all,
real-world robotics applications. This paper leverages recent progress on
Convolutional Neural Networks (CNNs) and proposes a novel RGB-D architecture
for object recognition. Our architecture is composed of two separate CNN
processing streams - one for each modality - which are consecutively combined
with a late fusion network. We focus on learning with imperfect sensor data, a
typical problem in real-world robotics tasks. For accurate learning, we
introduce a multi-stage training methodology and two crucial ingredients for
handling depth data with CNNs. The first, an effective encoding of depth
information for CNNs that enables learning without the need for large depth
datasets. The second, a data augmentation scheme for robust learning with depth
images by corrupting them with realistic noise patterns. We present
state-of-the-art results on the RGB-D object dataset and show recognition in
challenging RGB-D real-world noisy settings.
| no_new_dataset | 0.94545 |
1508.04190 | Zongbo Hao | Hao Zongbo, Lu Linlin, Zhang Qianni, Wu Jie, Izquierdo Ebroul, Yang
Juanyu, Zhao Jun | Action Recognition based on Subdivision-Fusion Model | Accepted by BMVC2015 | null | null | null | cs.CV | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | This paper proposes a novel Subdivision-Fusion Model (SFM) to recognize human
actions. In most action recognition tasks, overlapping feature distribution is
a common problem leading to overfitting. In the subdivision stage of the
proposed SFM, samples in each category are clustered. Then, such samples are
grouped into multiple more concentrated subcategories. Boundaries for the
subcategories are easier to find and as consequence overfitting is avoided. In
the subsequent fusion stage, the multi-subcategories classification results are
converted back to the original category recognition problem. Two methods to
determine the number of clusters are provided. The proposed model has been
thoroughly tested with four popular datasets. In the Hollywood2 dataset, an
accuracy of 79.4% is achieved, outperforming the state-of-the-art accuracy of
64.3%. The performance on the YouTube Action dataset has been improved from
75.8% to 82.5%, while considerably improvements are also observed on the KTH
and UCF50 datasets.
| [
{
"version": "v1",
"created": "Tue, 18 Aug 2015 01:38:08 GMT"
}
] | 2015-08-19T00:00:00 | [
[
"Zongbo",
"Hao",
""
],
[
"Linlin",
"Lu",
""
],
[
"Qianni",
"Zhang",
""
],
[
"Jie",
"Wu",
""
],
[
"Ebroul",
"Izquierdo",
""
],
[
"Juanyu",
"Yang",
""
],
[
"Jun",
"Zhao",
""
]
] | TITLE: Action Recognition based on Subdivision-Fusion Model
ABSTRACT: This paper proposes a novel Subdivision-Fusion Model (SFM) to recognize human
actions. In most action recognition tasks, overlapping feature distribution is
a common problem leading to overfitting. In the subdivision stage of the
proposed SFM, samples in each category are clustered. Then, such samples are
grouped into multiple more concentrated subcategories. Boundaries for the
subcategories are easier to find and as consequence overfitting is avoided. In
the subsequent fusion stage, the multi-subcategories classification results are
converted back to the original category recognition problem. Two methods to
determine the number of clusters are provided. The proposed model has been
thoroughly tested with four popular datasets. In the Hollywood2 dataset, an
accuracy of 79.4% is achieved, outperforming the state-of-the-art accuracy of
64.3%. The performance on the YouTube Action dataset has been improved from
75.8% to 82.5%, while considerably improvements are also observed on the KTH
and UCF50 datasets.
| no_new_dataset | 0.955026 |
1508.04198 | Yifan Fu | Yifan Fu and Junbin Gao and Xia Hong and David Tien | Low Rank Representation on Riemannian Manifold of Square Root Densities | null | null | null | null | cs.CV | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | In this paper, we present a novel low rank representation (LRR) algorithm for
data lying on the manifold of square root densities. Unlike traditional LRR
methods which rely on the assumption that the data points are vectors in the
Euclidean space, our new algorithm is designed to incorporate the intrinsic
geometric structure and geodesic distance of the manifold. Experiments on
several computer vision datasets showcase its noise robustness and superior
performance on classification and subspace clustering compared to other
state-of-the-art approaches.
| [
{
"version": "v1",
"created": "Tue, 18 Aug 2015 02:33:30 GMT"
}
] | 2015-08-19T00:00:00 | [
[
"Fu",
"Yifan",
""
],
[
"Gao",
"Junbin",
""
],
[
"Hong",
"Xia",
""
],
[
"Tien",
"David",
""
]
] | TITLE: Low Rank Representation on Riemannian Manifold of Square Root Densities
ABSTRACT: In this paper, we present a novel low rank representation (LRR) algorithm for
data lying on the manifold of square root densities. Unlike traditional LRR
methods which rely on the assumption that the data points are vectors in the
Euclidean space, our new algorithm is designed to incorporate the intrinsic
geometric structure and geodesic distance of the manifold. Experiments on
several computer vision datasets showcase its noise robustness and superior
performance on classification and subspace clustering compared to other
state-of-the-art approaches.
| no_new_dataset | 0.956553 |
1508.04333 | Shouvick Mondal | Shouvick Mondal and Arko Banerjee | ESDF: Ensemble Selection using Diversity and Frequency | Conference: National Conference on Research Trends in Computer
Science and Application (NCRTCSA-2014) Date: 8th February 2014 Organized by:
Dept. of Computer Application, Siliguri Institute of Technology, India In
Association With: Computer Society of India, Siliguri Chapter Technically
Sponsored By: IEEE, Kolkata Section Paper Id: NCRTCSA118. Shouvick Mondal et
al.; ESDF: Ensemble Selection using Diversity and Frequency; Proceedings of
NCRTCSA 2014; pp. 28-33, 2014 | null | null | null | cs.LG | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | Recently ensemble selection for consensus clustering has emerged as a
research problem in Machine Intelligence. Normally consensus clustering
algorithms take into account the entire ensemble of clustering, where there is
a tendency of generating a very large size ensemble before computing its
consensus. One can avoid considering the entire ensemble and can judiciously
select few partitions in the ensemble without compromising on the quality of
the consensus. This may result in an efficient consensus computation technique
and may save unnecessary computational overheads. The ensemble selection
problem addresses this issue of consensus clustering. In this paper, we propose
an efficient method of ensemble selection for a large ensemble. We prioritize
the partitions in the ensemble based on diversity and frequency. Our method
selects top K of the partitions in order of priority, where K is decided by the
user. We observe that considering jointly the diversity and frequency helps in
identifying few representative partitions whose consensus is qualitatively
better than the consensus of the entire ensemble. Experimental analysis on a
large number of datasets shows our method gives better results than earlier
ensemble selection methods.
| [
{
"version": "v1",
"created": "Tue, 18 Aug 2015 14:43:57 GMT"
}
] | 2015-08-19T00:00:00 | [
[
"Mondal",
"Shouvick",
""
],
[
"Banerjee",
"Arko",
""
]
] | TITLE: ESDF: Ensemble Selection using Diversity and Frequency
ABSTRACT: Recently ensemble selection for consensus clustering has emerged as a
research problem in Machine Intelligence. Normally consensus clustering
algorithms take into account the entire ensemble of clustering, where there is
a tendency of generating a very large size ensemble before computing its
consensus. One can avoid considering the entire ensemble and can judiciously
select few partitions in the ensemble without compromising on the quality of
the consensus. This may result in an efficient consensus computation technique
and may save unnecessary computational overheads. The ensemble selection
problem addresses this issue of consensus clustering. In this paper, we propose
an efficient method of ensemble selection for a large ensemble. We prioritize
the partitions in the ensemble based on diversity and frequency. Our method
selects top K of the partitions in order of priority, where K is decided by the
user. We observe that considering jointly the diversity and frequency helps in
identifying few representative partitions whose consensus is qualitatively
better than the consensus of the entire ensemble. Experimental analysis on a
large number of datasets shows our method gives better results than earlier
ensemble selection methods.
| no_new_dataset | 0.954984 |
1508.04389 | Rajeev Ranjan | Rajeev Ranjan, Vishal M. Patel, Rama Chellappa | A Deep Pyramid Deformable Part Model for Face Detection | null | null | null | null | cs.CV | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | We present a face detection algorithm based on Deformable Part Models and
deep pyramidal features. The proposed method called DP2MFD is able to detect
faces of various sizes and poses in unconstrained conditions. It reduces the
gap in training and testing of DPM on deep features by adding a normalization
layer to the deep convolutional neural network (CNN). Extensive experiments on
four publicly available unconstrained face detection datasets show that our
method is able to capture the meaningful structure of faces and performs
significantly better than many competitive face detection algorithms.
| [
{
"version": "v1",
"created": "Tue, 18 Aug 2015 17:24:09 GMT"
}
] | 2015-08-19T00:00:00 | [
[
"Ranjan",
"Rajeev",
""
],
[
"Patel",
"Vishal M.",
""
],
[
"Chellappa",
"Rama",
""
]
] | TITLE: A Deep Pyramid Deformable Part Model for Face Detection
ABSTRACT: We present a face detection algorithm based on Deformable Part Models and
deep pyramidal features. The proposed method called DP2MFD is able to detect
faces of various sizes and poses in unconstrained conditions. It reduces the
gap in training and testing of DPM on deep features by adding a normalization
layer to the deep convolutional neural network (CNN). Extensive experiments on
four publicly available unconstrained face detection datasets show that our
method is able to capture the meaningful structure of faces and performs
significantly better than many competitive face detection algorithms.
| no_new_dataset | 0.948917 |
1506.00379 | Yankai Lin | Yankai Lin, Zhiyuan Liu, Huanbo Luan, Maosong Sun, Siwei Rao, Song Liu | Modeling Relation Paths for Representation Learning of Knowledge Bases | 10 pages | null | null | null | cs.CL | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | Representation learning of knowledge bases (KBs) aims to embed both entities
and relations into a low-dimensional space. Most existing methods only consider
direct relations in representation learning. We argue that multiple-step
relation paths also contain rich inference patterns between entities, and
propose a path-based representation learning model. This model considers
relation paths as translations between entities for representation learning,
and addresses two key challenges: (1) Since not all relation paths are
reliable, we design a path-constraint resource allocation algorithm to measure
the reliability of relation paths. (2) We represent relation paths via semantic
composition of relation embeddings. Experimental results on real-world datasets
show that, as compared with baselines, our model achieves significant and
consistent improvements on knowledge base completion and relation extraction
from text.
| [
{
"version": "v1",
"created": "Mon, 1 Jun 2015 08:22:49 GMT"
},
{
"version": "v2",
"created": "Sat, 15 Aug 2015 09:28:49 GMT"
}
] | 2015-08-18T00:00:00 | [
[
"Lin",
"Yankai",
""
],
[
"Liu",
"Zhiyuan",
""
],
[
"Luan",
"Huanbo",
""
],
[
"Sun",
"Maosong",
""
],
[
"Rao",
"Siwei",
""
],
[
"Liu",
"Song",
""
]
] | TITLE: Modeling Relation Paths for Representation Learning of Knowledge Bases
ABSTRACT: Representation learning of knowledge bases (KBs) aims to embed both entities
and relations into a low-dimensional space. Most existing methods only consider
direct relations in representation learning. We argue that multiple-step
relation paths also contain rich inference patterns between entities, and
propose a path-based representation learning model. This model considers
relation paths as translations between entities for representation learning,
and addresses two key challenges: (1) Since not all relation paths are
reliable, we design a path-constraint resource allocation algorithm to measure
the reliability of relation paths. (2) We represent relation paths via semantic
composition of relation embeddings. Experimental results on real-world datasets
show that, as compared with baselines, our model achieves significant and
consistent improvements on knowledge base completion and relation extraction
from text.
| no_new_dataset | 0.948058 |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.