id
stringlengths 9
16
| submitter
stringlengths 3
64
⌀ | authors
stringlengths 5
6.63k
| title
stringlengths 7
245
| comments
stringlengths 1
482
⌀ | journal-ref
stringlengths 4
382
⌀ | doi
stringlengths 9
151
⌀ | report-no
stringclasses 984
values | categories
stringlengths 5
108
| license
stringclasses 9
values | abstract
stringlengths 83
3.41k
| versions
listlengths 1
20
| update_date
timestamp[s]date 2007-05-23 00:00:00
2025-04-11 00:00:00
| authors_parsed
sequencelengths 1
427
| prompt
stringlengths 166
3.49k
| label
stringclasses 2
values | prob
float64 0.5
0.98
|
---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
1412.3040 | Albert Kim | Albert Kim, Eric Blais, Aditya Parameswaran, Piotr Indyk, Sam Madden,
Ronitt Rubinfeld | Rapid Sampling for Visualizations with Ordering Guarantees | Tech Report. 17 pages. Condensed version to appear in VLDB Vol. 8 No.
5 | null | null | null | cs.DB | http://creativecommons.org/licenses/by-nc-sa/3.0/ | Visualizations are frequently used as a means to understand trends and gather
insights from datasets, but often take a long time to generate. In this paper,
we focus on the problem of rapidly generating approximate visualizations while
preserving crucial visual proper- ties of interest to analysts. Our primary
focus will be on sampling algorithms that preserve the visual property of
ordering; our techniques will also apply to some other visual properties. For
instance, our algorithms can be used to generate an approximate visualization
of a bar chart very rapidly, where the comparisons between any two bars are
correct. We formally show that our sampling algorithms are generally applicable
and provably optimal in theory, in that they do not take more samples than
necessary to generate the visualizations with ordering guarantees. They also
work well in practice, correctly ordering output groups while taking orders of
magnitude fewer samples and much less time than conventional sampling schemes.
| [
{
"version": "v1",
"created": "Tue, 9 Dec 2014 18:08:26 GMT"
}
] | 2014-12-10T00:00:00 | [
[
"Kim",
"Albert",
""
],
[
"Blais",
"Eric",
""
],
[
"Parameswaran",
"Aditya",
""
],
[
"Indyk",
"Piotr",
""
],
[
"Madden",
"Sam",
""
],
[
"Rubinfeld",
"Ronitt",
""
]
] | TITLE: Rapid Sampling for Visualizations with Ordering Guarantees
ABSTRACT: Visualizations are frequently used as a means to understand trends and gather
insights from datasets, but often take a long time to generate. In this paper,
we focus on the problem of rapidly generating approximate visualizations while
preserving crucial visual proper- ties of interest to analysts. Our primary
focus will be on sampling algorithms that preserve the visual property of
ordering; our techniques will also apply to some other visual properties. For
instance, our algorithms can be used to generate an approximate visualization
of a bar chart very rapidly, where the comparisons between any two bars are
correct. We formally show that our sampling algorithms are generally applicable
and provably optimal in theory, in that they do not take more samples than
necessary to generate the visualizations with ordering guarantees. They also
work well in practice, correctly ordering output groups while taking orders of
magnitude fewer samples and much less time than conventional sampling schemes.
| no_new_dataset | 0.951594 |
1406.1485 | Li Yao | Tapani Raiko, Li Yao, Kyunghyun Cho and Yoshua Bengio | Iterative Neural Autoregressive Distribution Estimator (NADE-k) | Accepted at Neural Information Processing Systems (NIPS) 2014 | null | null | null | stat.ML cs.LG | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | Training of the neural autoregressive density estimator (NADE) can be viewed
as doing one step of probabilistic inference on missing values in data. We
propose a new model that extends this inference scheme to multiple steps,
arguing that it is easier to learn to improve a reconstruction in $k$ steps
rather than to learn to reconstruct in a single inference step. The proposed
model is an unsupervised building block for deep learning that combines the
desirable properties of NADE and multi-predictive training: (1) Its test
likelihood can be computed analytically, (2) it is easy to generate independent
samples from it, and (3) it uses an inference engine that is a superset of
variational inference for Boltzmann machines. The proposed NADE-k is
competitive with the state-of-the-art in density estimation on the two datasets
tested.
| [
{
"version": "v1",
"created": "Thu, 5 Jun 2014 19:13:51 GMT"
},
{
"version": "v2",
"created": "Wed, 11 Jun 2014 11:28:36 GMT"
},
{
"version": "v3",
"created": "Sat, 6 Dec 2014 00:22:00 GMT"
}
] | 2014-12-09T00:00:00 | [
[
"Raiko",
"Tapani",
""
],
[
"Yao",
"Li",
""
],
[
"Cho",
"Kyunghyun",
""
],
[
"Bengio",
"Yoshua",
""
]
] | TITLE: Iterative Neural Autoregressive Distribution Estimator (NADE-k)
ABSTRACT: Training of the neural autoregressive density estimator (NADE) can be viewed
as doing one step of probabilistic inference on missing values in data. We
propose a new model that extends this inference scheme to multiple steps,
arguing that it is easier to learn to improve a reconstruction in $k$ steps
rather than to learn to reconstruct in a single inference step. The proposed
model is an unsupervised building block for deep learning that combines the
desirable properties of NADE and multi-predictive training: (1) Its test
likelihood can be computed analytically, (2) it is easy to generate independent
samples from it, and (3) it uses an inference engine that is a superset of
variational inference for Boltzmann machines. The proposed NADE-k is
competitive with the state-of-the-art in density estimation on the two datasets
tested.
| no_new_dataset | 0.946892 |
1411.1792 | Jason Yosinski | Jason Yosinski, Jeff Clune, Yoshua Bengio, Hod Lipson | How transferable are features in deep neural networks? | To appear in Advances in Neural Information Processing Systems 27
(NIPS 2014) | Advances in Neural Information Processing Systems 27, pages
3320-3328. Dec. 2014 | null | null | cs.LG cs.NE | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | Many deep neural networks trained on natural images exhibit a curious
phenomenon in common: on the first layer they learn features similar to Gabor
filters and color blobs. Such first-layer features appear not to be specific to
a particular dataset or task, but general in that they are applicable to many
datasets and tasks. Features must eventually transition from general to
specific by the last layer of the network, but this transition has not been
studied extensively. In this paper we experimentally quantify the generality
versus specificity of neurons in each layer of a deep convolutional neural
network and report a few surprising results. Transferability is negatively
affected by two distinct issues: (1) the specialization of higher layer neurons
to their original task at the expense of performance on the target task, which
was expected, and (2) optimization difficulties related to splitting networks
between co-adapted neurons, which was not expected. In an example network
trained on ImageNet, we demonstrate that either of these two issues may
dominate, depending on whether features are transferred from the bottom,
middle, or top of the network. We also document that the transferability of
features decreases as the distance between the base task and target task
increases, but that transferring features even from distant tasks can be better
than using random features. A final surprising result is that initializing a
network with transferred features from almost any number of layers can produce
a boost to generalization that lingers even after fine-tuning to the target
dataset.
| [
{
"version": "v1",
"created": "Thu, 6 Nov 2014 23:09:37 GMT"
}
] | 2014-12-09T00:00:00 | [
[
"Yosinski",
"Jason",
""
],
[
"Clune",
"Jeff",
""
],
[
"Bengio",
"Yoshua",
""
],
[
"Lipson",
"Hod",
""
]
] | TITLE: How transferable are features in deep neural networks?
ABSTRACT: Many deep neural networks trained on natural images exhibit a curious
phenomenon in common: on the first layer they learn features similar to Gabor
filters and color blobs. Such first-layer features appear not to be specific to
a particular dataset or task, but general in that they are applicable to many
datasets and tasks. Features must eventually transition from general to
specific by the last layer of the network, but this transition has not been
studied extensively. In this paper we experimentally quantify the generality
versus specificity of neurons in each layer of a deep convolutional neural
network and report a few surprising results. Transferability is negatively
affected by two distinct issues: (1) the specialization of higher layer neurons
to their original task at the expense of performance on the target task, which
was expected, and (2) optimization difficulties related to splitting networks
between co-adapted neurons, which was not expected. In an example network
trained on ImageNet, we demonstrate that either of these two issues may
dominate, depending on whether features are transferred from the bottom,
middle, or top of the network. We also document that the transferability of
features decreases as the distance between the base task and target task
increases, but that transferring features even from distant tasks can be better
than using random features. A final surprising result is that initializing a
network with transferred features from almost any number of layers can produce
a boost to generalization that lingers even after fine-tuning to the target
dataset.
| no_new_dataset | 0.942135 |
1412.2485 | Vikram Nathan | Vikram Nathan, Sharath Raghvendra | Accurate Streaming Support Vector Machines | 2 figures, 8 pages | null | null | null | cs.LG | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | A widely-used tool for binary classification is the Support Vector Machine
(SVM), a supervised learning technique that finds the "maximum margin" linear
separator between the two classes. While SVMs have been well studied in the
batch (offline) setting, there is considerably less work on the streaming
(online) setting, which requires only a single pass over the data using
sub-linear space. Existing streaming algorithms are not yet competitive with
the batch implementation. In this paper, we use the formulation of the SVM as a
minimum enclosing ball (MEB) problem to provide a streaming SVM algorithm based
off of the blurred ball cover originally proposed by Agarwal and Sharathkumar.
Our implementation consistently outperforms existing streaming SVM approaches
and provides higher accuracies than libSVM on several datasets, thus making it
competitive with the standard SVM batch implementation.
| [
{
"version": "v1",
"created": "Mon, 8 Dec 2014 08:46:07 GMT"
}
] | 2014-12-09T00:00:00 | [
[
"Nathan",
"Vikram",
""
],
[
"Raghvendra",
"Sharath",
""
]
] | TITLE: Accurate Streaming Support Vector Machines
ABSTRACT: A widely-used tool for binary classification is the Support Vector Machine
(SVM), a supervised learning technique that finds the "maximum margin" linear
separator between the two classes. While SVMs have been well studied in the
batch (offline) setting, there is considerably less work on the streaming
(online) setting, which requires only a single pass over the data using
sub-linear space. Existing streaming algorithms are not yet competitive with
the batch implementation. In this paper, we use the formulation of the SVM as a
minimum enclosing ball (MEB) problem to provide a streaming SVM algorithm based
off of the blurred ball cover originally proposed by Agarwal and Sharathkumar.
Our implementation consistently outperforms existing streaming SVM approaches
and provides higher accuracies than libSVM on several datasets, thus making it
competitive with the standard SVM batch implementation.
| no_new_dataset | 0.948202 |
1412.2697 | Hocine Cherifi | Abdelkaher Ait Abdelouahad, Mohammed El Hassouni, Hocine Cherifi, and
Driss Aboutajdine | Image quality assessment measure based on natural image statistics in
the Tetrolet domain | null | null | null | null | cs.CV | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | This paper deals with a reduced reference (RR) image quality measure based on
natural image statistics modeling. For this purpose, Tetrolet transform is used
since it provides a convenient way to capture local geometric structures. This
transform is applied to both reference and distorted images. Then, Gaussian
Scale Mixture (GSM) is proposed to model subbands in order to take account
statistical dependencies between tetrolet coefficients. In order to quantify
the visual degradation, a measure based on Kullback Leibler Divergence (KLD) is
provided. The proposed measure was tested on the Cornell VCL A-57 dataset and
compared with other measures according to FR-TV1 VQEG framework.
| [
{
"version": "v1",
"created": "Mon, 8 Dec 2014 18:48:26 GMT"
}
] | 2014-12-09T00:00:00 | [
[
"Abdelouahad",
"Abdelkaher Ait",
""
],
[
"Hassouni",
"Mohammed El",
""
],
[
"Cherifi",
"Hocine",
""
],
[
"Aboutajdine",
"Driss",
""
]
] | TITLE: Image quality assessment measure based on natural image statistics in
the Tetrolet domain
ABSTRACT: This paper deals with a reduced reference (RR) image quality measure based on
natural image statistics modeling. For this purpose, Tetrolet transform is used
since it provides a convenient way to capture local geometric structures. This
transform is applied to both reference and distorted images. Then, Gaussian
Scale Mixture (GSM) is proposed to model subbands in order to take account
statistical dependencies between tetrolet coefficients. In order to quantify
the visual degradation, a measure based on Kullback Leibler Divergence (KLD) is
provided. The proposed measure was tested on the Cornell VCL A-57 dataset and
compared with other measures according to FR-TV1 VQEG framework.
| no_new_dataset | 0.951863 |
1412.2716 | Paul Ginsparg | Daniel T. Citron and Paul Ginsparg | Patterns of Text Reuse in a Scientific Corpus | 6 pages, plus 10 pages of supplementary material. To appear in PNAS
(online 8 Dec 2014) | null | 10.1073/pnas.1415135111 | null | cs.DL physics.soc-ph | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | We consider the incidence of text "reuse" by researchers, via a systematic
pairwise comparison of the text content of all articles deposited to arXiv.org
from 1991--2012. We measure the global frequencies of three classes of text
reuse, and measure how chronic text reuse is distributed among authors in the
dataset. We infer a baseline for accepted practice, perhaps surprisingly
permissive compared with other societal contexts, and a clearly delineated set
of aberrant authors. We find a negative correlation between the amount of
reused text in an article and its influence, as measured by subsequent
citations. Finally, we consider the distribution of countries of origin of
articles containing large amounts of reused text.
| [
{
"version": "v1",
"created": "Mon, 8 Dec 2014 20:01:17 GMT"
}
] | 2014-12-09T00:00:00 | [
[
"Citron",
"Daniel T.",
""
],
[
"Ginsparg",
"Paul",
""
]
] | TITLE: Patterns of Text Reuse in a Scientific Corpus
ABSTRACT: We consider the incidence of text "reuse" by researchers, via a systematic
pairwise comparison of the text content of all articles deposited to arXiv.org
from 1991--2012. We measure the global frequencies of three classes of text
reuse, and measure how chronic text reuse is distributed among authors in the
dataset. We infer a baseline for accepted practice, perhaps surprisingly
permissive compared with other societal contexts, and a clearly delineated set
of aberrant authors. We find a negative correlation between the amount of
reused text in an article and its influence, as measured by subsequent
citations. Finally, we consider the distribution of countries of origin of
articles containing large amounts of reused text.
| no_new_dataset | 0.9462 |
1410.7454 | Neeraj Dhungel | Neeraj Dhungel, Gustavo Carneiro, Andrew P. Bradley | Deep Structured learning for mass segmentation from Mammograms | 4 pages, 2 figures | null | null | null | cs.CV | http://creativecommons.org/licenses/by/3.0/ | In this paper, we present a novel method for the segmentation of breast
masses from mammograms exploring structured and deep learning. Specifically,
using structured support vector machine (SSVM), we formulate a model that
combines different types of potential functions, including one that classifies
image regions using deep learning. Our main goal with this work is to show the
accuracy and efficiency improvements that these relatively new techniques can
provide for the segmentation of breast masses from mammograms. We also propose
an easily reproducible quantitative analysis to as- sess the performance of
breast mass segmentation methodologies based on widely accepted accuracy and
running time measurements on public datasets, which will facilitate further
comparisons for this segmentation problem. In particular, we use two publicly
available datasets (DDSM-BCRP and INbreast) and propose the computa- tion of
the running time taken for the methodology to produce a mass segmentation given
an input image and the use of the Dice index to quantitatively measure the
segmentation accuracy. For both databases, we show that our proposed
methodology produces competitive results in terms of accuracy and running time.
| [
{
"version": "v1",
"created": "Mon, 27 Oct 2014 22:44:26 GMT"
},
{
"version": "v2",
"created": "Fri, 5 Dec 2014 01:21:39 GMT"
}
] | 2014-12-08T00:00:00 | [
[
"Dhungel",
"Neeraj",
""
],
[
"Carneiro",
"Gustavo",
""
],
[
"Bradley",
"Andrew P.",
""
]
] | TITLE: Deep Structured learning for mass segmentation from Mammograms
ABSTRACT: In this paper, we present a novel method for the segmentation of breast
masses from mammograms exploring structured and deep learning. Specifically,
using structured support vector machine (SSVM), we formulate a model that
combines different types of potential functions, including one that classifies
image regions using deep learning. Our main goal with this work is to show the
accuracy and efficiency improvements that these relatively new techniques can
provide for the segmentation of breast masses from mammograms. We also propose
an easily reproducible quantitative analysis to as- sess the performance of
breast mass segmentation methodologies based on widely accepted accuracy and
running time measurements on public datasets, which will facilitate further
comparisons for this segmentation problem. In particular, we use two publicly
available datasets (DDSM-BCRP and INbreast) and propose the computa- tion of
the running time taken for the methodology to produce a mass segmentation given
an input image and the use of the Dice index to quantitatively measure the
segmentation accuracy. For both databases, we show that our proposed
methodology produces competitive results in terms of accuracy and running time.
| no_new_dataset | 0.951594 |
1412.1842 | Max Jaderberg | Max Jaderberg, Karen Simonyan, Andrea Vedaldi, Andrew Zisserman | Reading Text in the Wild with Convolutional Neural Networks | null | null | null | null | cs.CV | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | In this work we present an end-to-end system for text spotting -- localising
and recognising text in natural scene images -- and text based image retrieval.
This system is based on a region proposal mechanism for detection and deep
convolutional neural networks for recognition. Our pipeline uses a novel
combination of complementary proposal generation techniques to ensure high
recall, and a fast subsequent filtering stage for improving precision. For the
recognition and ranking of proposals, we train very large convolutional neural
networks to perform word recognition on the whole proposal region at the same
time, departing from the character classifier based systems of the past. These
networks are trained solely on data produced by a synthetic text generation
engine, requiring no human labelled data.
Analysing the stages of our pipeline, we show state-of-the-art performance
throughout. We perform rigorous experiments across a number of standard
end-to-end text spotting benchmarks and text-based image retrieval datasets,
showing a large improvement over all previous methods. Finally, we demonstrate
a real-world application of our text spotting system to allow thousands of
hours of news footage to be instantly searchable via a text query.
| [
{
"version": "v1",
"created": "Thu, 4 Dec 2014 21:14:59 GMT"
}
] | 2014-12-08T00:00:00 | [
[
"Jaderberg",
"Max",
""
],
[
"Simonyan",
"Karen",
""
],
[
"Vedaldi",
"Andrea",
""
],
[
"Zisserman",
"Andrew",
""
]
] | TITLE: Reading Text in the Wild with Convolutional Neural Networks
ABSTRACT: In this work we present an end-to-end system for text spotting -- localising
and recognising text in natural scene images -- and text based image retrieval.
This system is based on a region proposal mechanism for detection and deep
convolutional neural networks for recognition. Our pipeline uses a novel
combination of complementary proposal generation techniques to ensure high
recall, and a fast subsequent filtering stage for improving precision. For the
recognition and ranking of proposals, we train very large convolutional neural
networks to perform word recognition on the whole proposal region at the same
time, departing from the character classifier based systems of the past. These
networks are trained solely on data produced by a synthetic text generation
engine, requiring no human labelled data.
Analysing the stages of our pipeline, we show state-of-the-art performance
throughout. We perform rigorous experiments across a number of standard
end-to-end text spotting benchmarks and text-based image retrieval datasets,
showing a large improvement over all previous methods. Finally, we demonstrate
a real-world application of our text spotting system to allow thousands of
hours of news footage to be instantly searchable via a text query.
| no_new_dataset | 0.95018 |
1412.1888 | Rafi Muhammad | Muhammad Rafi, Farnaz Amin, Mohammad Shahid Shaikh | Document clustering using graph based document representation with
constraints | null | null | null | null | cs.IR | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | Document clustering is an unsupervised approach in which a large collection
of documents (corpus) is subdivided into smaller, meaningful, identifiable, and
verifiable sub-groups (clusters). Meaningful representation of documents and
implicitly identifying the patterns, on which this separation is performed, is
the challenging part of document clustering. We have proposed a document
clustering technique using graph based document representation with
constraints. A graph data structure can easily capture the non-linear
relationships of nodes, document contains various feature terms that can be
non-linearly connected hence a graph can easily represents this information.
Constrains, are explicit conditions for document clustering where background
knowledge is use to set the direction for Linking or Not-Linking a set of
documents for a target clusters, thus guiding the clustering process. We deemed
clustering is an ill-define problem, there can be many clustering results.
Background knowledge can be used to drive the clustering algorithm in the right
direction. We have proposed three different types of constraints, Instance
level, corpus level and cluster level constraints. A new algorithm Constrained
HAC is also proposed which will incorporate Instance level constraints as prior
knowledge; it will guide the clustering process leading to better results.
Extensive set of experiments have been performed on both synthetic and standard
document clustering datasets, results are compared on standard clustering
measures like: purity, entropy and F-measure. Results clearly establish that
our proposed approach leads to improvement in cluster quality.
| [
{
"version": "v1",
"created": "Fri, 5 Dec 2014 03:47:34 GMT"
}
] | 2014-12-08T00:00:00 | [
[
"Rafi",
"Muhammad",
""
],
[
"Amin",
"Farnaz",
""
],
[
"Shaikh",
"Mohammad Shahid",
""
]
] | TITLE: Document clustering using graph based document representation with
constraints
ABSTRACT: Document clustering is an unsupervised approach in which a large collection
of documents (corpus) is subdivided into smaller, meaningful, identifiable, and
verifiable sub-groups (clusters). Meaningful representation of documents and
implicitly identifying the patterns, on which this separation is performed, is
the challenging part of document clustering. We have proposed a document
clustering technique using graph based document representation with
constraints. A graph data structure can easily capture the non-linear
relationships of nodes, document contains various feature terms that can be
non-linearly connected hence a graph can easily represents this information.
Constrains, are explicit conditions for document clustering where background
knowledge is use to set the direction for Linking or Not-Linking a set of
documents for a target clusters, thus guiding the clustering process. We deemed
clustering is an ill-define problem, there can be many clustering results.
Background knowledge can be used to drive the clustering algorithm in the right
direction. We have proposed three different types of constraints, Instance
level, corpus level and cluster level constraints. A new algorithm Constrained
HAC is also proposed which will incorporate Instance level constraints as prior
knowledge; it will guide the clustering process leading to better results.
Extensive set of experiments have been performed on both synthetic and standard
document clustering datasets, results are compared on standard clustering
measures like: purity, entropy and F-measure. Results clearly establish that
our proposed approach leads to improvement in cluster quality.
| no_new_dataset | 0.952442 |
1412.1908 | Rui Zhao | Rui Zhao, Wanli Ouyang, Xiaogang Wang | Person Re-identification by Saliency Learning | This manuscript has 14 pages with 25 figures, and a preliminary
version was published in ICCV 2013 | null | null | null | cs.CV | http://creativecommons.org/licenses/by-nc-sa/3.0/ | Human eyes can recognize person identities based on small salient regions,
i.e. human saliency is distinctive and reliable in pedestrian matching across
disjoint camera views. However, such valuable information is often hidden when
computing similarities of pedestrian images with existing approaches. Inspired
by our user study result of human perception on human saliency, we propose a
novel perspective for person re-identification based on learning human saliency
and matching saliency distribution. The proposed saliency learning and matching
framework consists of four steps: (1) To handle misalignment caused by drastic
viewpoint change and pose variations, we apply adjacency constrained patch
matching to build dense correspondence between image pairs. (2) We propose two
alternative methods, i.e. K-Nearest Neighbors and One-class SVM, to estimate a
saliency score for each image patch, through which distinctive features stand
out without using identity labels in the training procedure. (3) saliency
matching is proposed based on patch matching. Matching patches with
inconsistent saliency brings penalty, and images of the same identity are
recognized by minimizing the saliency matching cost. (4) Furthermore, saliency
matching is tightly integrated with patch matching in a unified structural
RankSVM learning framework. The effectiveness of our approach is validated on
the VIPeR dataset and the CUHK01 dataset. Our approach outperforms the
state-of-the-art person re-identification methods on both datasets.
| [
{
"version": "v1",
"created": "Fri, 5 Dec 2014 07:33:48 GMT"
}
] | 2014-12-08T00:00:00 | [
[
"Zhao",
"Rui",
""
],
[
"Ouyang",
"Wanli",
""
],
[
"Wang",
"Xiaogang",
""
]
] | TITLE: Person Re-identification by Saliency Learning
ABSTRACT: Human eyes can recognize person identities based on small salient regions,
i.e. human saliency is distinctive and reliable in pedestrian matching across
disjoint camera views. However, such valuable information is often hidden when
computing similarities of pedestrian images with existing approaches. Inspired
by our user study result of human perception on human saliency, we propose a
novel perspective for person re-identification based on learning human saliency
and matching saliency distribution. The proposed saliency learning and matching
framework consists of four steps: (1) To handle misalignment caused by drastic
viewpoint change and pose variations, we apply adjacency constrained patch
matching to build dense correspondence between image pairs. (2) We propose two
alternative methods, i.e. K-Nearest Neighbors and One-class SVM, to estimate a
saliency score for each image patch, through which distinctive features stand
out without using identity labels in the training procedure. (3) saliency
matching is proposed based on patch matching. Matching patches with
inconsistent saliency brings penalty, and images of the same identity are
recognized by minimizing the saliency matching cost. (4) Furthermore, saliency
matching is tightly integrated with patch matching in a unified structural
RankSVM learning framework. The effectiveness of our approach is validated on
the VIPeR dataset and the CUHK01 dataset. Our approach outperforms the
state-of-the-art person re-identification methods on both datasets.
| no_new_dataset | 0.950778 |
1412.1947 | Aditya AV Sastry Mr. | Aditya AV Sastry and Kalyan Netti | A parallel sampling based clustering | null | null | null | null | cs.LG | http://creativecommons.org/licenses/publicdomain/ | The problem of automatically clustering data is an age old problem. People
have created numerous algorithms to tackle this problem. The execution time of
any of this algorithm grows with the number of input points and the number of
cluster centers required. To reduce the number of input points we could average
the points locally and use the means or the local centers as the input for
clustering. However since the required number of local centers is very high,
running the clustering algorithm on the entire dataset to obtain these
representational points is very time consuming. To remedy this problem, in this
paper we are proposing two subclustering schemes where by we subdivide the
dataset into smaller sets and run the clustering algorithm on the smaller
datasets to obtain the required number of datapoints to run our clustering
algorithm with. As we are subdividing the given dataset, we could run
clustering algorithm on each smaller piece of the dataset in parallel. We found
that both parallel and serial execution of this method to be much faster than
the original clustering algorithm and error in running the clustering algorithm
on a reduced set to be very less.
| [
{
"version": "v1",
"created": "Fri, 5 Dec 2014 10:50:31 GMT"
}
] | 2014-12-08T00:00:00 | [
[
"Sastry",
"Aditya AV",
""
],
[
"Netti",
"Kalyan",
""
]
] | TITLE: A parallel sampling based clustering
ABSTRACT: The problem of automatically clustering data is an age old problem. People
have created numerous algorithms to tackle this problem. The execution time of
any of this algorithm grows with the number of input points and the number of
cluster centers required. To reduce the number of input points we could average
the points locally and use the means or the local centers as the input for
clustering. However since the required number of local centers is very high,
running the clustering algorithm on the entire dataset to obtain these
representational points is very time consuming. To remedy this problem, in this
paper we are proposing two subclustering schemes where by we subdivide the
dataset into smaller sets and run the clustering algorithm on the smaller
datasets to obtain the required number of datapoints to run our clustering
algorithm with. As we are subdividing the given dataset, we could run
clustering algorithm on each smaller piece of the dataset in parallel. We found
that both parallel and serial execution of this method to be much faster than
the original clustering algorithm and error in running the clustering algorithm
on a reduced set to be very less.
| no_new_dataset | 0.95388 |
1409.1801 | Stephen Plaza | Stephen M. Plaza, Toufiq Parag, Gary B. Huang, Donald J. Olbris,
Mathew A. Saunders, Patricia K. Rivlin | Annotating Synapses in Large EM Datasets | null | null | null | null | q-bio.QM cs.CV q-bio.NC | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | Reconstructing neuronal circuits at the level of synapses is a central
problem in neuroscience and becoming a focus of the emerging field of
connectomics. To date, electron microscopy (EM) is the most proven technique
for identifying and quantifying synaptic connections. As advances in EM make
acquiring larger datasets possible, subsequent manual synapse identification
({\em i.e.}, proofreading) for deciphering a connectome becomes a major time
bottleneck. Here we introduce a large-scale, high-throughput, and
semi-automated methodology to efficiently identify synapses. We successfully
applied our methodology to the Drosophila medulla optic lobe, annotating many
more synapses than previous connectome efforts. Our approaches are extensible
and will make the often complicated process of synapse identification
accessible to a wider-community of potential proofreaders.
| [
{
"version": "v1",
"created": "Fri, 5 Sep 2014 13:52:47 GMT"
},
{
"version": "v2",
"created": "Thu, 4 Dec 2014 16:18:01 GMT"
}
] | 2014-12-05T00:00:00 | [
[
"Plaza",
"Stephen M.",
""
],
[
"Parag",
"Toufiq",
""
],
[
"Huang",
"Gary B.",
""
],
[
"Olbris",
"Donald J.",
""
],
[
"Saunders",
"Mathew A.",
""
],
[
"Rivlin",
"Patricia K.",
""
]
] | TITLE: Annotating Synapses in Large EM Datasets
ABSTRACT: Reconstructing neuronal circuits at the level of synapses is a central
problem in neuroscience and becoming a focus of the emerging field of
connectomics. To date, electron microscopy (EM) is the most proven technique
for identifying and quantifying synaptic connections. As advances in EM make
acquiring larger datasets possible, subsequent manual synapse identification
({\em i.e.}, proofreading) for deciphering a connectome becomes a major time
bottleneck. Here we introduce a large-scale, high-throughput, and
semi-automated methodology to efficiently identify synapses. We successfully
applied our methodology to the Drosophila medulla optic lobe, annotating many
more synapses than previous connectome efforts. Our approaches are extensible
and will make the often complicated process of synapse identification
accessible to a wider-community of potential proofreaders.
| no_new_dataset | 0.951953 |
1412.1602 | Jan Chorowski | Jan Chorowski, Dzmitry Bahdanau, Kyunghyun Cho, Yoshua Bengio | End-to-end Continuous Speech Recognition using Attention-based Recurrent
NN: First Results | As accepted to: Deep Learning and Representation Learning Workshop,
NIPS 2014 | null | null | null | cs.NE cs.LG stat.ML | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | We replace the Hidden Markov Model (HMM) which is traditionally used in in
continuous speech recognition with a bi-directional recurrent neural network
encoder coupled to a recurrent neural network decoder that directly emits a
stream of phonemes. The alignment between the input and output sequences is
established using an attention mechanism: the decoder emits each symbol based
on a context created with a subset of input symbols elected by the attention
mechanism. We report initial results demonstrating that this new approach
achieves phoneme error rates that are comparable to the state-of-the-art
HMM-based decoders, on the TIMIT dataset.
| [
{
"version": "v1",
"created": "Thu, 4 Dec 2014 10:00:19 GMT"
}
] | 2014-12-05T00:00:00 | [
[
"Chorowski",
"Jan",
""
],
[
"Bahdanau",
"Dzmitry",
""
],
[
"Cho",
"Kyunghyun",
""
],
[
"Bengio",
"Yoshua",
""
]
] | TITLE: End-to-end Continuous Speech Recognition using Attention-based Recurrent
NN: First Results
ABSTRACT: We replace the Hidden Markov Model (HMM) which is traditionally used in in
continuous speech recognition with a bi-directional recurrent neural network
encoder coupled to a recurrent neural network decoder that directly emits a
stream of phonemes. The alignment between the input and output sequences is
established using an attention mechanism: the decoder emits each symbol based
on a context created with a subset of input symbols elected by the attention
mechanism. We report initial results demonstrating that this new approach
achieves phoneme error rates that are comparable to the state-of-the-art
HMM-based decoders, on the TIMIT dataset.
| no_new_dataset | 0.953188 |
1412.1632 | Lei Yu | Lei Yu, Karl Moritz Hermann, Phil Blunsom and Stephen Pulman | Deep Learning for Answer Sentence Selection | 9 pages, accepted by NIPS deep learning workshop | null | null | null | cs.CL | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | Answer sentence selection is the task of identifying sentences that contain
the answer to a given question. This is an important problem in its own right
as well as in the larger context of open domain question answering. We propose
a novel approach to solving this task via means of distributed representations,
and learn to match questions with answers by considering their semantic
encoding. This contrasts prior work on this task, which typically relies on
classifiers with large numbers of hand-crafted syntactic and semantic features
and various external resources. Our approach does not require any feature
engineering nor does it involve specialist linguistic data, making this model
easily applicable to a wide range of domains and languages. Experimental
results on a standard benchmark dataset from TREC demonstrate that---despite
its simplicity---our model matches state of the art performance on the answer
sentence selection task.
| [
{
"version": "v1",
"created": "Thu, 4 Dec 2014 11:53:02 GMT"
}
] | 2014-12-05T00:00:00 | [
[
"Yu",
"Lei",
""
],
[
"Hermann",
"Karl Moritz",
""
],
[
"Blunsom",
"Phil",
""
],
[
"Pulman",
"Stephen",
""
]
] | TITLE: Deep Learning for Answer Sentence Selection
ABSTRACT: Answer sentence selection is the task of identifying sentences that contain
the answer to a given question. This is an important problem in its own right
as well as in the larger context of open domain question answering. We propose
a novel approach to solving this task via means of distributed representations,
and learn to match questions with answers by considering their semantic
encoding. This contrasts prior work on this task, which typically relies on
classifiers with large numbers of hand-crafted syntactic and semantic features
and various external resources. Our approach does not require any feature
engineering nor does it involve specialist linguistic data, making this model
easily applicable to a wide range of domains and languages. Experimental
results on a standard benchmark dataset from TREC demonstrate that---despite
its simplicity---our model matches state of the art performance on the answer
sentence selection task.
| no_new_dataset | 0.946547 |
1412.1710 | Kaiming He | Kaiming He, Jian Sun | Convolutional Neural Networks at Constrained Time Cost | 8-page technical report | null | null | null | cs.CV | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | Though recent advanced convolutional neural networks (CNNs) have been
improving the image recognition accuracy, the models are getting more complex
and time-consuming. For real-world applications in industrial and commercial
scenarios, engineers and developers are often faced with the requirement of
constrained time budget. In this paper, we investigate the accuracy of CNNs
under constrained time cost. Under this constraint, the designs of the network
architectures should exhibit as trade-offs among the factors like depth,
numbers of filters, filter sizes, etc. With a series of controlled comparisons,
we progressively modify a baseline model while preserving its time complexity.
This is also helpful for understanding the importance of the factors in network
designs. We present an architecture that achieves very competitive accuracy in
the ImageNet dataset (11.8% top-5 error, 10-view test), yet is 20% faster than
"AlexNet" (16.0% top-5 error, 10-view test).
| [
{
"version": "v1",
"created": "Thu, 4 Dec 2014 16:00:47 GMT"
}
] | 2014-12-05T00:00:00 | [
[
"He",
"Kaiming",
""
],
[
"Sun",
"Jian",
""
]
] | TITLE: Convolutional Neural Networks at Constrained Time Cost
ABSTRACT: Though recent advanced convolutional neural networks (CNNs) have been
improving the image recognition accuracy, the models are getting more complex
and time-consuming. For real-world applications in industrial and commercial
scenarios, engineers and developers are often faced with the requirement of
constrained time budget. In this paper, we investigate the accuracy of CNNs
under constrained time cost. Under this constraint, the designs of the network
architectures should exhibit as trade-offs among the factors like depth,
numbers of filters, filter sizes, etc. With a series of controlled comparisons,
we progressively modify a baseline model while preserving its time complexity.
This is also helpful for understanding the importance of the factors in network
designs. We present an architecture that achieves very competitive accuracy in
the ImageNet dataset (11.8% top-5 error, 10-view test), yet is 20% faster than
"AlexNet" (16.0% top-5 error, 10-view test).
| no_new_dataset | 0.951863 |
1412.1788 | Francis Bach | Felipe Yanez (LIENS, INRIA Paris - Rocquencourt), Francis Bach (LIENS,
INRIA Paris - Rocquencourt) | Primal-Dual Algorithms for Non-negative Matrix Factorization with the
Kullback-Leibler Divergence | null | null | null | null | cs.LG math.OC | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | Non-negative matrix factorization (NMF) approximates a given matrix as a
product of two non-negative matrices. Multiplicative algorithms deliver
reliable results, but they show slow convergence for high-dimensional data and
may be stuck away from local minima. Gradient descent methods have better
behavior, but only apply to smooth losses such as the least-squares loss. In
this article, we propose a first-order primal-dual algorithm for non-negative
decomposition problems (where one factor is fixed) with the KL divergence,
based on the Chambolle-Pock algorithm. All required computations may be
obtained in closed form and we provide an efficient heuristic way to select
step-sizes. By using alternating optimization, our algorithm readily extends to
NMF and, on synthetic examples, face recognition or music source separation
datasets, it is either faster than existing algorithms, or leads to improved
local optima, or both.
| [
{
"version": "v1",
"created": "Thu, 4 Dec 2014 19:56:02 GMT"
}
] | 2014-12-05T00:00:00 | [
[
"Yanez",
"Felipe",
"",
"LIENS, INRIA Paris - Rocquencourt"
],
[
"Bach",
"Francis",
"",
"LIENS,\n INRIA Paris - Rocquencourt"
]
] | TITLE: Primal-Dual Algorithms for Non-negative Matrix Factorization with the
Kullback-Leibler Divergence
ABSTRACT: Non-negative matrix factorization (NMF) approximates a given matrix as a
product of two non-negative matrices. Multiplicative algorithms deliver
reliable results, but they show slow convergence for high-dimensional data and
may be stuck away from local minima. Gradient descent methods have better
behavior, but only apply to smooth losses such as the least-squares loss. In
this article, we propose a first-order primal-dual algorithm for non-negative
decomposition problems (where one factor is fixed) with the KL divergence,
based on the Chambolle-Pock algorithm. All required computations may be
obtained in closed form and we provide an efficient heuristic way to select
step-sizes. By using alternating optimization, our algorithm readily extends to
NMF and, on synthetic examples, face recognition or music source separation
datasets, it is either faster than existing algorithms, or leads to improved
local optima, or both.
| no_new_dataset | 0.942771 |
1309.5643 | Veronika Cheplygina | Veronika Cheplygina, David M. J. Tax, and Marco Loog | Multiple Instance Learning with Bag Dissimilarities | Pattern Recognition, in press | Pattern Recognition 48.1 (2015): 264-275 | 10.1016/j.patcog.2014.07.022 | null | stat.ML cs.LG | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | Multiple instance learning (MIL) is concerned with learning from sets (bags)
of objects (instances), where the individual instance labels are ambiguous. In
this setting, supervised learning cannot be applied directly. Often,
specialized MIL methods learn by making additional assumptions about the
relationship of the bag labels and instance labels. Such assumptions may fit a
particular dataset, but do not generalize to the whole range of MIL problems.
Other MIL methods shift the focus of assumptions from the labels to the overall
(dis)similarity of bags, and therefore learn from bags directly. We propose to
represent each bag by a vector of its dissimilarities to other bags in the
training set, and treat these dissimilarities as a feature representation. We
show several alternatives to define a dissimilarity between bags and discuss
which definitions are more suitable for particular MIL problems. The
experimental results show that the proposed approach is computationally
inexpensive, yet very competitive with state-of-the-art algorithms on a wide
range of MIL datasets.
| [
{
"version": "v1",
"created": "Sun, 22 Sep 2013 20:24:50 GMT"
},
{
"version": "v2",
"created": "Thu, 6 Feb 2014 13:13:11 GMT"
},
{
"version": "v3",
"created": "Tue, 12 Aug 2014 09:04:32 GMT"
}
] | 2014-12-04T00:00:00 | [
[
"Cheplygina",
"Veronika",
""
],
[
"Tax",
"David M. J.",
""
],
[
"Loog",
"Marco",
""
]
] | TITLE: Multiple Instance Learning with Bag Dissimilarities
ABSTRACT: Multiple instance learning (MIL) is concerned with learning from sets (bags)
of objects (instances), where the individual instance labels are ambiguous. In
this setting, supervised learning cannot be applied directly. Often,
specialized MIL methods learn by making additional assumptions about the
relationship of the bag labels and instance labels. Such assumptions may fit a
particular dataset, but do not generalize to the whole range of MIL problems.
Other MIL methods shift the focus of assumptions from the labels to the overall
(dis)similarity of bags, and therefore learn from bags directly. We propose to
represent each bag by a vector of its dissimilarities to other bags in the
training set, and treat these dissimilarities as a feature representation. We
show several alternatives to define a dissimilarity between bags and discuss
which definitions are more suitable for particular MIL problems. The
experimental results show that the proposed approach is computationally
inexpensive, yet very competitive with state-of-the-art algorithms on a wide
range of MIL datasets.
| no_new_dataset | 0.947624 |
1412.1138 | Ben Fulcher | B. D. Fulcher, A. E. Georgieva, C. W. G. Redman, Nick S. Jones | Highly comparative fetal heart rate analysis | 7 pages, 4 figures | Fulcher, B. D., Georgieva, A., Redman, C. W., & Jones, N. S.
(2012). Highly comparative fetal heart rate analysis (pp. 3135-3138).
Presented at the 34th Annual International Conference of the IEEE EMBS, San
Diego, CA, USA | 10.1109/EMBC.2012.6346629 | null | cs.LG cs.AI q-bio.QM | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | A database of fetal heart rate (FHR) time series measured from 7221 patients
during labor is analyzed with the aim of learning the types of features of
these recordings that are informative of low cord pH. Our 'highly comparative'
analysis involves extracting over 9000 time-series analysis features from each
FHR time series, including measures of autocorrelation, entropy, distribution,
and various model fits. This diverse collection of features was developed in
previous work, and is publicly available. We describe five features that most
accurately classify a balanced training set of 59 'low pH' and 59 'normal pH'
FHR recordings. We then describe five of the features with the strongest linear
correlation to cord pH across the full dataset of FHR time series. The features
identified in this work may be used as part of a system for guiding
intervention during labor in future. This work successfully demonstrates the
utility of comparing across a large, interdisciplinary literature on
time-series analysis to automatically contribute new scientific results for
specific biomedical signal processing challenges.
| [
{
"version": "v1",
"created": "Wed, 3 Dec 2014 00:00:42 GMT"
}
] | 2014-12-04T00:00:00 | [
[
"Fulcher",
"B. D.",
""
],
[
"Georgieva",
"A. E.",
""
],
[
"Redman",
"C. W. G.",
""
],
[
"Jones",
"Nick S.",
""
]
] | TITLE: Highly comparative fetal heart rate analysis
ABSTRACT: A database of fetal heart rate (FHR) time series measured from 7221 patients
during labor is analyzed with the aim of learning the types of features of
these recordings that are informative of low cord pH. Our 'highly comparative'
analysis involves extracting over 9000 time-series analysis features from each
FHR time series, including measures of autocorrelation, entropy, distribution,
and various model fits. This diverse collection of features was developed in
previous work, and is publicly available. We describe five features that most
accurately classify a balanced training set of 59 'low pH' and 59 'normal pH'
FHR recordings. We then describe five of the features with the strongest linear
correlation to cord pH across the full dataset of FHR time series. The features
identified in this work may be used as part of a system for guiding
intervention during labor in future. This work successfully demonstrates the
utility of comparing across a large, interdisciplinary literature on
time-series analysis to automatically contribute new scientific results for
specific biomedical signal processing challenges.
| no_new_dataset | 0.929504 |
1412.1194 | Feng Shi | Feng Shi, Robert Laganiere, Emil Petriu | Gradient Boundary Histograms for Action Recognition | null | null | null | null | cs.CV | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | This paper introduces a high efficient local spatiotemporal descriptor,
called gradient boundary histograms (GBH). The proposed GBH descriptor is built
on simple spatio-temporal gradients, which are fast to compute. We demonstrate
that it can better represent local structure and motion than other
gradient-based descriptors, and significantly outperforms them on large
realistic datasets. A comprehensive evaluation shows that the recognition
accuracy is preserved while the spatial resolution is greatly reduced, which
yields both high efficiency and low memory usage.
| [
{
"version": "v1",
"created": "Wed, 3 Dec 2014 05:23:03 GMT"
}
] | 2014-12-04T00:00:00 | [
[
"Shi",
"Feng",
""
],
[
"Laganiere",
"Robert",
""
],
[
"Petriu",
"Emil",
""
]
] | TITLE: Gradient Boundary Histograms for Action Recognition
ABSTRACT: This paper introduces a high efficient local spatiotemporal descriptor,
called gradient boundary histograms (GBH). The proposed GBH descriptor is built
on simple spatio-temporal gradients, which are fast to compute. We demonstrate
that it can better represent local structure and motion than other
gradient-based descriptors, and significantly outperforms them on large
realistic datasets. A comprehensive evaluation shows that the recognition
accuracy is preserved while the spatial resolution is greatly reduced, which
yields both high efficiency and low memory usage.
| no_new_dataset | 0.949669 |
1412.1442 | Maxwell Collins | Maxwell D. Collins and Pushmeet Kohli | Memory Bounded Deep Convolutional Networks | null | null | null | null | cs.CV | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | In this work, we investigate the use of sparsity-inducing regularizers during
training of Convolution Neural Networks (CNNs). These regularizers encourage
that fewer connections in the convolution and fully connected layers take
non-zero values and in effect result in sparse connectivity between hidden
units in the deep network. This in turn reduces the memory and runtime cost
involved in deploying the learned CNNs. We show that training with such
regularization can still be performed using stochastic gradient descent
implying that it can be used easily in existing codebases. Experimental
evaluation of our approach on MNIST, CIFAR, and ImageNet datasets shows that
our regularizers can result in dramatic reductions in memory requirements. For
instance, when applied on AlexNet, our method can reduce the memory consumption
by a factor of four with minimal loss in accuracy.
| [
{
"version": "v1",
"created": "Wed, 3 Dec 2014 19:08:38 GMT"
}
] | 2014-12-04T00:00:00 | [
[
"Collins",
"Maxwell D.",
""
],
[
"Kohli",
"Pushmeet",
""
]
] | TITLE: Memory Bounded Deep Convolutional Networks
ABSTRACT: In this work, we investigate the use of sparsity-inducing regularizers during
training of Convolution Neural Networks (CNNs). These regularizers encourage
that fewer connections in the convolution and fully connected layers take
non-zero values and in effect result in sparse connectivity between hidden
units in the deep network. This in turn reduces the memory and runtime cost
involved in deploying the learned CNNs. We show that training with such
regularization can still be performed using stochastic gradient descent
implying that it can be used easily in existing codebases. Experimental
evaluation of our approach on MNIST, CIFAR, and ImageNet datasets shows that
our regularizers can result in dramatic reductions in memory requirements. For
instance, when applied on AlexNet, our method can reduce the memory consumption
by a factor of four with minimal loss in accuracy.
| no_new_dataset | 0.949248 |
1409.3809 | Daniel Crankshaw | Daniel Crankshaw, Peter Bailis, Joseph E. Gonzalez, Haoyuan Li, Zhao
Zhang, Michael J. Franklin, Ali Ghodsi, Michael I. Jordan | The Missing Piece in Complex Analytics: Low Latency, Scalable Model
Management and Serving with Velox | null | null | null | null | cs.DB | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | To support complex data-intensive applications such as personalized
recommendations, targeted advertising, and intelligent services, the data
management community has focused heavily on the design of systems to support
training complex models on large datasets. Unfortunately, the design of these
systems largely ignores a critical component of the overall analytics process:
the deployment and serving of models at scale. In this work, we present Velox,
a new component of the Berkeley Data Analytics Stack. Velox is a data
management system for facilitating the next steps in real-world, large-scale
analytics pipelines: online model management, maintenance, and serving. Velox
provides end-user applications and services with a low-latency, intuitive
interface to models, transforming the raw statistical models currently trained
using existing offline large-scale compute frameworks into full-blown,
end-to-end data products capable of recommending products, targeting
advertisements, and personalizing web content. To provide up-to-date results
for these complex models, Velox also facilitates lightweight online model
maintenance and selection (i.e., dynamic weighting). In this paper, we describe
the challenges and architectural considerations required to achieve this
functionality, including the abilities to span online and offline systems, to
adaptively adjust model materialization strategies, and to exploit inherent
statistical properties such as model error tolerance, all while operating at
"Big Data" scale.
| [
{
"version": "v1",
"created": "Fri, 12 Sep 2014 18:12:24 GMT"
},
{
"version": "v2",
"created": "Mon, 1 Dec 2014 23:20:30 GMT"
}
] | 2014-12-03T00:00:00 | [
[
"Crankshaw",
"Daniel",
""
],
[
"Bailis",
"Peter",
""
],
[
"Gonzalez",
"Joseph E.",
""
],
[
"Li",
"Haoyuan",
""
],
[
"Zhang",
"Zhao",
""
],
[
"Franklin",
"Michael J.",
""
],
[
"Ghodsi",
"Ali",
""
],
[
"Jordan",
"Michael I.",
""
]
] | TITLE: The Missing Piece in Complex Analytics: Low Latency, Scalable Model
Management and Serving with Velox
ABSTRACT: To support complex data-intensive applications such as personalized
recommendations, targeted advertising, and intelligent services, the data
management community has focused heavily on the design of systems to support
training complex models on large datasets. Unfortunately, the design of these
systems largely ignores a critical component of the overall analytics process:
the deployment and serving of models at scale. In this work, we present Velox,
a new component of the Berkeley Data Analytics Stack. Velox is a data
management system for facilitating the next steps in real-world, large-scale
analytics pipelines: online model management, maintenance, and serving. Velox
provides end-user applications and services with a low-latency, intuitive
interface to models, transforming the raw statistical models currently trained
using existing offline large-scale compute frameworks into full-blown,
end-to-end data products capable of recommending products, targeting
advertisements, and personalizing web content. To provide up-to-date results
for these complex models, Velox also facilitates lightweight online model
maintenance and selection (i.e., dynamic weighting). In this paper, we describe
the challenges and architectural considerations required to achieve this
functionality, including the abilities to span online and offline systems, to
adaptively adjust model materialization strategies, and to exploit inherent
statistical properties such as model error tolerance, all while operating at
"Big Data" scale.
| no_new_dataset | 0.946597 |
1402.3384 | Weina Wang | Weina Wang, Lei Ying and Junshan Zhang | A Minimax Distortion View of Differentially Private Query Release | null | null | null | null | cs.CR cs.DB cs.IT math.IT | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | We consider the problem of differentially private query release through a
synthetic database approach. Departing from the existing approaches that
require the query set to be specified in advance, we advocate to devise
query-set independent mechanisms, with an ambitious goal of providing accurate
answers, while meeting the privacy constraints, for all queries in a general
query class. Specifically, a differentially private mechanism is constructed to
"encode" rich stochastic structure into the synthetic database, and
"customized" companion estimators are then derived to provide accurate answers
by making use of all available information, including the mechanism (which is
public information) and the query functions. Accordingly, the distortion under
the best of this kind of mechanisms at the worst-case query in a general query
class, so called the minimax distortion, provides a fundamental
characterization of differentially private query release.
For the general class of statistical queries, we prove that with the
squared-error distortion measure, the minimax distortion is $O(1/n)$ by
deriving asymptotically tight upper and lower bounds in the regime that the
database size $n$ goes to infinity. The upper bound is achievable by a
mechanism $\mathcal{E}$ and its corresponding companion estimators, which
points directly to the feasibility of the proposed approach in large databases.
We further evaluate the mechanism $\mathcal{E}$ and the companion estimators
through experiments on real datasets from Netflix and Facebook. Experimental
results show improvement over the state-of-art MWEM algorithm and verify the
scaling behavior $O(1/n)$ of the minimax distortion.
| [
{
"version": "v1",
"created": "Fri, 14 Feb 2014 06:54:49 GMT"
},
{
"version": "v2",
"created": "Mon, 1 Dec 2014 07:01:08 GMT"
}
] | 2014-12-02T00:00:00 | [
[
"Wang",
"Weina",
""
],
[
"Ying",
"Lei",
""
],
[
"Zhang",
"Junshan",
""
]
] | TITLE: A Minimax Distortion View of Differentially Private Query Release
ABSTRACT: We consider the problem of differentially private query release through a
synthetic database approach. Departing from the existing approaches that
require the query set to be specified in advance, we advocate to devise
query-set independent mechanisms, with an ambitious goal of providing accurate
answers, while meeting the privacy constraints, for all queries in a general
query class. Specifically, a differentially private mechanism is constructed to
"encode" rich stochastic structure into the synthetic database, and
"customized" companion estimators are then derived to provide accurate answers
by making use of all available information, including the mechanism (which is
public information) and the query functions. Accordingly, the distortion under
the best of this kind of mechanisms at the worst-case query in a general query
class, so called the minimax distortion, provides a fundamental
characterization of differentially private query release.
For the general class of statistical queries, we prove that with the
squared-error distortion measure, the minimax distortion is $O(1/n)$ by
deriving asymptotically tight upper and lower bounds in the regime that the
database size $n$ goes to infinity. The upper bound is achievable by a
mechanism $\mathcal{E}$ and its corresponding companion estimators, which
points directly to the feasibility of the proposed approach in large databases.
We further evaluate the mechanism $\mathcal{E}$ and the companion estimators
through experiments on real datasets from Netflix and Facebook. Experimental
results show improvement over the state-of-art MWEM algorithm and verify the
scaling behavior $O(1/n)$ of the minimax distortion.
| no_new_dataset | 0.944228 |
1410.3726 | Chee Seng Chan | Chern Hong Lim, Anhar Risnumawan and Chee Seng Chan | Scene Image is Non-Mutually Exclusive - A Fuzzy Qualitative Scene
Understanding | Accepted in IEEE Transactions on Fuzzy Systems | IEEE Transactions on Fuzzy Systems, vol. 22(6), pp. 1541 - 1556,
2014 | 10.1109/TFUZZ.2014.2298233 | null | cs.CV cs.AI cs.IR | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | Ambiguity or uncertainty is a pervasive element of many real world decision
making processes. Variation in decisions is a norm in this situation when the
same problem is posed to different subjects. Psychological and metaphysical
research had proven that decision making by human is subjective. It is
influenced by many factors such as experience, age, background, etc. Scene
understanding is one of the computer vision problems that fall into this
category. Conventional methods relax this problem by assuming scene images are
mutually exclusive; and therefore, focus on developing different approaches to
perform the binary classification tasks. In this paper, we show that scene
images are non-mutually exclusive, and propose the Fuzzy Qualitative Rank
Classifier (FQRC) to tackle the aforementioned problems. The proposed FQRC
provides a ranking interpretation instead of binary decision. Evaluations in
term of qualitative and quantitative using large numbers and challenging public
scene datasets have shown the effectiveness of our proposed method in modeling
the non-mutually exclusive scene images.
| [
{
"version": "v1",
"created": "Tue, 14 Oct 2014 15:19:43 GMT"
}
] | 2014-12-02T00:00:00 | [
[
"Lim",
"Chern Hong",
""
],
[
"Risnumawan",
"Anhar",
""
],
[
"Chan",
"Chee Seng",
""
]
] | TITLE: Scene Image is Non-Mutually Exclusive - A Fuzzy Qualitative Scene
Understanding
ABSTRACT: Ambiguity or uncertainty is a pervasive element of many real world decision
making processes. Variation in decisions is a norm in this situation when the
same problem is posed to different subjects. Psychological and metaphysical
research had proven that decision making by human is subjective. It is
influenced by many factors such as experience, age, background, etc. Scene
understanding is one of the computer vision problems that fall into this
category. Conventional methods relax this problem by assuming scene images are
mutually exclusive; and therefore, focus on developing different approaches to
perform the binary classification tasks. In this paper, we show that scene
images are non-mutually exclusive, and propose the Fuzzy Qualitative Rank
Classifier (FQRC) to tackle the aforementioned problems. The proposed FQRC
provides a ranking interpretation instead of binary decision. Evaluations in
term of qualitative and quantitative using large numbers and challenging public
scene datasets have shown the effectiveness of our proposed method in modeling
the non-mutually exclusive scene images.
| no_new_dataset | 0.957675 |
1412.0008 | Robert Templeman | Mohammed Korayem, Robert Templeman, Dennis Chen, David Crandall, Apu
Kapadia | ScreenAvoider: Protecting Computer Screens from Ubiquitous Cameras | null | null | null | null | cs.CR | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | We live and work in environments that are inundated with cameras embedded in
devices such as phones, tablets, laptops, and monitors. Newer wearable devices
like Google Glass, Narrative Clip, and Autographer offer the ability to quietly
log our lives with cameras from a `first person' perspective. While capturing
several meaningful and interesting moments, a significant number of images
captured by these wearable cameras can contain computer screens. Given the
potentially sensitive information that is visible on our displays, there is a
need to guard computer screens from undesired photography. People need
protection against photography of their screens, whether by other people's
cameras or their own cameras.
We present ScreenAvoider, a framework that controls the collection and
disclosure of images with computer screens and their sensitive content.
ScreenAvoider can detect images with computer screens with high accuracy and
can even go so far as to discriminate amongst screen content. We also introduce
a ScreenTag system that aids in the identification of screen content, flagging
images with highly sensitive content such as messaging applications or email
webpages. We evaluate our concept on realistic lifelogging datasets, showing
that ScreenAvoider provides a practical and useful solution that can help users
manage their privacy.
| [
{
"version": "v1",
"created": "Fri, 28 Nov 2014 01:50:53 GMT"
}
] | 2014-12-02T00:00:00 | [
[
"Korayem",
"Mohammed",
""
],
[
"Templeman",
"Robert",
""
],
[
"Chen",
"Dennis",
""
],
[
"Crandall",
"David",
""
],
[
"Kapadia",
"Apu",
""
]
] | TITLE: ScreenAvoider: Protecting Computer Screens from Ubiquitous Cameras
ABSTRACT: We live and work in environments that are inundated with cameras embedded in
devices such as phones, tablets, laptops, and monitors. Newer wearable devices
like Google Glass, Narrative Clip, and Autographer offer the ability to quietly
log our lives with cameras from a `first person' perspective. While capturing
several meaningful and interesting moments, a significant number of images
captured by these wearable cameras can contain computer screens. Given the
potentially sensitive information that is visible on our displays, there is a
need to guard computer screens from undesired photography. People need
protection against photography of their screens, whether by other people's
cameras or their own cameras.
We present ScreenAvoider, a framework that controls the collection and
disclosure of images with computer screens and their sensitive content.
ScreenAvoider can detect images with computer screens with high accuracy and
can even go so far as to discriminate amongst screen content. We also introduce
a ScreenTag system that aids in the identification of screen content, flagging
images with highly sensitive content such as messaging applications or email
webpages. We evaluate our concept on realistic lifelogging datasets, showing
that ScreenAvoider provides a practical and useful solution that can help users
manage their privacy.
| no_new_dataset | 0.916559 |
1412.0059 | Takashi Sakuragawa | Kent Miyajima and Takashi Sakuragawa | Continuous and robust clustering coefficients for weighted and directed
networks | 29 pages, 14 figures | null | null | null | physics.soc-ph cs.SI | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | We introduce new clustering coefficients for weighted networks. They are
continuous and robust against edge weight changes. Recently, generalized
clustering coefficients for weighted and directed networks have been proposed.
These generalizations have a common property, that their values are not
continuous. They are sensitive with edge weight changes, especially at zero
weight. With these generalizations, if vanishingly low weights of edges are
truncated to weight zero for some reason, the coefficient value may change
significantly from the original value. It is preferable that small changes of
edge weights cause small changes of coefficient value. We call this property
the continuity of generalized clustering coefficients. Our new coefficients
admit this property. In the past, few studies have focused on the continuity of
generalized clustering coefficients. In experiments, we performed comparative
assessments of existing and our generalizations. In the case of a real world
network dataset (C. Elegans Neural network), after adding random edge weight
errors, though the value of one discontinuous generalization was changed about
436%, the value of proposed one was only changed 0.2%.
| [
{
"version": "v1",
"created": "Sat, 29 Nov 2014 01:57:08 GMT"
}
] | 2014-12-02T00:00:00 | [
[
"Miyajima",
"Kent",
""
],
[
"Sakuragawa",
"Takashi",
""
]
] | TITLE: Continuous and robust clustering coefficients for weighted and directed
networks
ABSTRACT: We introduce new clustering coefficients for weighted networks. They are
continuous and robust against edge weight changes. Recently, generalized
clustering coefficients for weighted and directed networks have been proposed.
These generalizations have a common property, that their values are not
continuous. They are sensitive with edge weight changes, especially at zero
weight. With these generalizations, if vanishingly low weights of edges are
truncated to weight zero for some reason, the coefficient value may change
significantly from the original value. It is preferable that small changes of
edge weights cause small changes of coefficient value. We call this property
the continuity of generalized clustering coefficients. Our new coefficients
admit this property. In the past, few studies have focused on the continuity of
generalized clustering coefficients. In experiments, we performed comparative
assessments of existing and our generalizations. In the case of a real world
network dataset (C. Elegans Neural network), after adding random edge weight
errors, though the value of one discontinuous generalization was changed about
436%, the value of proposed one was only changed 0.2%.
| no_new_dataset | 0.949623 |
1412.0065 | Gr\'egory Rogez | Gregory Rogez, James S. Supancic III, Maryam Khademi, Jose Maria
Martinez Montiel, Deva Ramanan | 3D Hand Pose Detection in Egocentric RGB-D Images | 14 pages, 15 figures, extended version of the corresponding ECCV
workshop paper, submitted to International Journal of Computer Vision | null | null | null | cs.CV | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | We focus on the task of everyday hand pose estimation from egocentric
viewpoints. For this task, we show that depth sensors are particularly
informative for extracting near-field interactions of the camera wearer with
his/her environment. Despite the recent advances in full-body pose estimation
using Kinect-like sensors, reliable monocular hand pose estimation in RGB-D
images is still an unsolved problem. The problem is considerably exacerbated
when analyzing hands performing daily activities from a first-person viewpoint,
due to severe occlusions arising from object manipulations and a limited
field-of-view. Our system addresses these difficulties by exploiting strong
priors over viewpoint and pose in a discriminative tracking-by-detection
framework. Our priors are operationalized through a photorealistic synthetic
model of egocentric scenes, which is used to generate training data for
learning depth-based pose classifiers. We evaluate our approach on an annotated
dataset of real egocentric object manipulation scenes and compare to both
commercial and academic approaches. Our method provides state-of-the-art
performance for both hand detection and pose estimation in egocentric RGB-D
images.
| [
{
"version": "v1",
"created": "Sat, 29 Nov 2014 03:19:56 GMT"
}
] | 2014-12-02T00:00:00 | [
[
"Rogez",
"Gregory",
""
],
[
"Supancic",
"James S.",
"III"
],
[
"Khademi",
"Maryam",
""
],
[
"Montiel",
"Jose Maria Martinez",
""
],
[
"Ramanan",
"Deva",
""
]
] | TITLE: 3D Hand Pose Detection in Egocentric RGB-D Images
ABSTRACT: We focus on the task of everyday hand pose estimation from egocentric
viewpoints. For this task, we show that depth sensors are particularly
informative for extracting near-field interactions of the camera wearer with
his/her environment. Despite the recent advances in full-body pose estimation
using Kinect-like sensors, reliable monocular hand pose estimation in RGB-D
images is still an unsolved problem. The problem is considerably exacerbated
when analyzing hands performing daily activities from a first-person viewpoint,
due to severe occlusions arising from object manipulations and a limited
field-of-view. Our system addresses these difficulties by exploiting strong
priors over viewpoint and pose in a discriminative tracking-by-detection
framework. Our priors are operationalized through a photorealistic synthetic
model of egocentric scenes, which is used to generate training data for
learning depth-based pose classifiers. We evaluate our approach on an annotated
dataset of real egocentric object manipulation scenes and compare to both
commercial and academic approaches. Our method provides state-of-the-art
performance for both hand detection and pose estimation in egocentric RGB-D
images.
| no_new_dataset | 0.946941 |
1412.0069 | Yonglong Tian | Yonglong Tian, Ping Luo, Xiaogang Wang, Xiaoou Tang | Pedestrian Detection aided by Deep Learning Semantic Tasks | null | null | null | null | cs.CV | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | Deep learning methods have achieved great success in pedestrian detection,
owing to its ability to learn features from raw pixels. However, they mainly
capture middle-level representations, such as pose of pedestrian, but confuse
positive with hard negative samples, which have large ambiguity, e.g. the shape
and appearance of `tree trunk' or `wire pole' are similar to pedestrian in
certain viewpoint. This ambiguity can be distinguished by high-level
representation. To this end, this work jointly optimizes pedestrian detection
with semantic tasks, including pedestrian attributes (e.g. `carrying backpack')
and scene attributes (e.g. `road', `tree', and `horizontal'). Rather than
expensively annotating scene attributes, we transfer attributes information
from existing scene segmentation datasets to the pedestrian dataset, by
proposing a novel deep model to learn high-level features from multiple tasks
and multiple data sources. Since distinct tasks have distinct convergence rates
and data from different datasets have different distributions, a multi-task
objective function is carefully designed to coordinate tasks and reduce
discrepancies among datasets. The importance coefficients of tasks and network
parameters in this objective function can be iteratively estimated. Extensive
evaluations show that the proposed approach outperforms the state-of-the-art on
the challenging Caltech and ETH datasets, where it reduces the miss rates of
previous deep models by 17 and 5.5 percent, respectively.
| [
{
"version": "v1",
"created": "Sat, 29 Nov 2014 04:34:23 GMT"
}
] | 2014-12-02T00:00:00 | [
[
"Tian",
"Yonglong",
""
],
[
"Luo",
"Ping",
""
],
[
"Wang",
"Xiaogang",
""
],
[
"Tang",
"Xiaoou",
""
]
] | TITLE: Pedestrian Detection aided by Deep Learning Semantic Tasks
ABSTRACT: Deep learning methods have achieved great success in pedestrian detection,
owing to its ability to learn features from raw pixels. However, they mainly
capture middle-level representations, such as pose of pedestrian, but confuse
positive with hard negative samples, which have large ambiguity, e.g. the shape
and appearance of `tree trunk' or `wire pole' are similar to pedestrian in
certain viewpoint. This ambiguity can be distinguished by high-level
representation. To this end, this work jointly optimizes pedestrian detection
with semantic tasks, including pedestrian attributes (e.g. `carrying backpack')
and scene attributes (e.g. `road', `tree', and `horizontal'). Rather than
expensively annotating scene attributes, we transfer attributes information
from existing scene segmentation datasets to the pedestrian dataset, by
proposing a novel deep model to learn high-level features from multiple tasks
and multiple data sources. Since distinct tasks have distinct convergence rates
and data from different datasets have different distributions, a multi-task
objective function is carefully designed to coordinate tasks and reduce
discrepancies among datasets. The importance coefficients of tasks and network
parameters in this objective function can be iteratively estimated. Extensive
evaluations show that the proposed approach outperforms the state-of-the-art on
the challenging Caltech and ETH datasets, where it reduces the miss rates of
previous deep models by 17 and 5.5 percent, respectively.
| no_new_dataset | 0.941385 |
1412.0100 | Stefan Mathe | Stefan Mathe, Cristian Sminchisescu | Multiple Instance Reinforcement Learning for Efficient Weakly-Supervised
Detection in Images | null | null | null | null | cs.CV cs.LG | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | State-of-the-art visual recognition and detection systems increasingly rely
on large amounts of training data and complex classifiers. Therefore it becomes
increasingly expensive both to manually annotate datasets and to keep running
times at levels acceptable for practical applications. In this paper, we
propose two solutions to address these issues. First, we introduce a weakly
supervised, segmentation-based approach to learn accurate detectors and image
classifiers from weak supervisory signals that provide only approximate
constraints on target localization. We illustrate our system on the problem of
action detection in static images (Pascal VOC Actions 2012), using human visual
search patterns as our training signal. Second, inspired from the
saccade-and-fixate operating principle of the human visual system, we use
reinforcement learning techniques to train efficient search models for
detection. Our sequential method is weakly supervised and general (it does not
require eye movements), finds optimal search strategies for any given detection
confidence function and achieves performance similar to exhaustive sliding
window search at a fraction of its computational cost.
| [
{
"version": "v1",
"created": "Sat, 29 Nov 2014 12:18:14 GMT"
}
] | 2014-12-02T00:00:00 | [
[
"Mathe",
"Stefan",
""
],
[
"Sminchisescu",
"Cristian",
""
]
] | TITLE: Multiple Instance Reinforcement Learning for Efficient Weakly-Supervised
Detection in Images
ABSTRACT: State-of-the-art visual recognition and detection systems increasingly rely
on large amounts of training data and complex classifiers. Therefore it becomes
increasingly expensive both to manually annotate datasets and to keep running
times at levels acceptable for practical applications. In this paper, we
propose two solutions to address these issues. First, we introduce a weakly
supervised, segmentation-based approach to learn accurate detectors and image
classifiers from weak supervisory signals that provide only approximate
constraints on target localization. We illustrate our system on the problem of
action detection in static images (Pascal VOC Actions 2012), using human visual
search patterns as our training signal. Second, inspired from the
saccade-and-fixate operating principle of the human visual system, we use
reinforcement learning techniques to train efficient search models for
detection. Our sequential method is weakly supervised and general (it does not
require eye movements), finds optimal search strategies for any given detection
confidence function and achieves performance similar to exhaustive sliding
window search at a fraction of its computational cost.
| no_new_dataset | 0.946646 |
1412.0296 | George Papandreou | George Papandreou and Iasonas Kokkinos and Pierre-Andr\'e Savalle | Untangling Local and Global Deformations in Deep Convolutional Networks
for Image Classification and Sliding Window Detection | 13 pages, 7 figures, 5 tables. arXiv admin note: substantial text
overlap with arXiv:1406.2732 | null | null | null | cs.CV | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | Deep Convolutional Neural Networks (DCNNs) commonly use generic `max-pooling'
(MP) layers to extract deformation-invariant features, but we argue in favor of
a more refined treatment. First, we introduce epitomic convolution as a
building block alternative to the common convolution-MP cascade of DCNNs; while
having identical complexity to MP, Epitomic Convolution allows for parameter
sharing across different filters, resulting in faster convergence and better
generalization. Second, we introduce a Multiple Instance Learning approach to
explicitly accommodate global translation and scaling when training a DCNN
exclusively with class labels. For this we rely on a `patchwork' data structure
that efficiently lays out all image scales and positions as candidates to a
DCNN. Factoring global and local deformations allows a DCNN to `focus its
resources' on the treatment of non-rigid deformations and yields a substantial
classification accuracy improvement. Third, further pursuing this idea, we
develop an efficient DCNN sliding window object detector that employs explicit
search over position, scale, and aspect ratio. We provide competitive image
classification and localization results on the ImageNet dataset and object
detection results on the Pascal VOC 2007 benchmark.
| [
{
"version": "v1",
"created": "Sun, 30 Nov 2014 22:20:17 GMT"
}
] | 2014-12-02T00:00:00 | [
[
"Papandreou",
"George",
""
],
[
"Kokkinos",
"Iasonas",
""
],
[
"Savalle",
"Pierre-André",
""
]
] | TITLE: Untangling Local and Global Deformations in Deep Convolutional Networks
for Image Classification and Sliding Window Detection
ABSTRACT: Deep Convolutional Neural Networks (DCNNs) commonly use generic `max-pooling'
(MP) layers to extract deformation-invariant features, but we argue in favor of
a more refined treatment. First, we introduce epitomic convolution as a
building block alternative to the common convolution-MP cascade of DCNNs; while
having identical complexity to MP, Epitomic Convolution allows for parameter
sharing across different filters, resulting in faster convergence and better
generalization. Second, we introduce a Multiple Instance Learning approach to
explicitly accommodate global translation and scaling when training a DCNN
exclusively with class labels. For this we rely on a `patchwork' data structure
that efficiently lays out all image scales and positions as candidates to a
DCNN. Factoring global and local deformations allows a DCNN to `focus its
resources' on the treatment of non-rigid deformations and yields a substantial
classification accuracy improvement. Third, further pursuing this idea, we
develop an efficient DCNN sliding window object detector that employs explicit
search over position, scale, and aspect ratio. We provide competitive image
classification and localization results on the ImageNet dataset and object
detection results on the Pascal VOC 2007 benchmark.
| no_new_dataset | 0.949716 |
1412.0327 | Jiajun Liu | Jiajun Liu, Kun Zhao, Saeed Khan, Mark Cameron, Raja Jurdak | Multi-scale Population and Mobility Estimation with Geo-tagged Tweets | 1st International Workshop on Big Data Analytics for Biosecurity
(BioBAD2015), 4 pages | null | null | null | cs.SI cs.CY physics.soc-ph | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | Recent outbreaks of Ebola and Dengue viruses have again elevated the
significance of the capability to quickly predict disease spread in an emergent
situation. However, existing approaches usually rely heavily on the
time-consuming census processes, or the privacy-sensitive call logs, leading to
their unresponsive nature when facing the abruptly changing dynamics in the
event of an outbreak. In this paper we study the feasibility of using
large-scale Twitter data as a proxy of human mobility to model and predict
disease spread. We report that for Australia, Twitter users' distribution
correlates well the census-based population distribution, and that the Twitter
users' travel patterns appear to loosely follow the gravity law at multiple
scales of geographic distances, i.e. national level, state level and
metropolitan level. The radiation model is also evaluated on this dataset
though it has shown inferior fitness as a result of Australia's sparse
population and large landmass. The outcomes of the study form the cornerstones
for future work towards a model-based, responsive prediction method from
Twitter data for disease spread.
| [
{
"version": "v1",
"created": "Mon, 1 Dec 2014 01:48:53 GMT"
}
] | 2014-12-02T00:00:00 | [
[
"Liu",
"Jiajun",
""
],
[
"Zhao",
"Kun",
""
],
[
"Khan",
"Saeed",
""
],
[
"Cameron",
"Mark",
""
],
[
"Jurdak",
"Raja",
""
]
] | TITLE: Multi-scale Population and Mobility Estimation with Geo-tagged Tweets
ABSTRACT: Recent outbreaks of Ebola and Dengue viruses have again elevated the
significance of the capability to quickly predict disease spread in an emergent
situation. However, existing approaches usually rely heavily on the
time-consuming census processes, or the privacy-sensitive call logs, leading to
their unresponsive nature when facing the abruptly changing dynamics in the
event of an outbreak. In this paper we study the feasibility of using
large-scale Twitter data as a proxy of human mobility to model and predict
disease spread. We report that for Australia, Twitter users' distribution
correlates well the census-based population distribution, and that the Twitter
users' travel patterns appear to loosely follow the gravity law at multiple
scales of geographic distances, i.e. national level, state level and
metropolitan level. The radiation model is also evaluated on this dataset
though it has shown inferior fitness as a result of Australia's sparse
population and large landmass. The outcomes of the study form the cornerstones
for future work towards a model-based, responsive prediction method from
Twitter data for disease spread.
| no_new_dataset | 0.943608 |
1412.0488 | Pavel Tomancak | Tobias Pietzsch and Stephan Saalfeld and Stephan Preibisch and Pavel
Tomancak | BigDataViewer: Interactive Visualization and Image Processing for
Terabyte Data Sets | 38 pages, 1 main figure, 27 supplementary figures, under review at
Nature Methods | null | null | null | q-bio.QM cs.GR | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | The increasingly popular light sheet microscopy techniques generate very
large 3D time-lapse recordings of living biological specimen. The necessity to
make large volumetric datasets available for interactive visualization and
analysis has been widely recognized. However, existing solutions build on
dedicated servers to generate virtual slices that are transferred to the client
applications, practically leading to insufficient frame rates (less than 10
frames per second) for truly interactive experience. An easily accessible open
source solution for interactive arbitrary virtual re-slicing of very large
volumes and time series of volumes has yet been missing. We fill this gap with
BigDataViewer, a Fiji plugin to interactively navigate and visualize large
image sequences from both local and remote data sources.
| [
{
"version": "v1",
"created": "Mon, 1 Dec 2014 14:24:44 GMT"
}
] | 2014-12-02T00:00:00 | [
[
"Pietzsch",
"Tobias",
""
],
[
"Saalfeld",
"Stephan",
""
],
[
"Preibisch",
"Stephan",
""
],
[
"Tomancak",
"Pavel",
""
]
] | TITLE: BigDataViewer: Interactive Visualization and Image Processing for
Terabyte Data Sets
ABSTRACT: The increasingly popular light sheet microscopy techniques generate very
large 3D time-lapse recordings of living biological specimen. The necessity to
make large volumetric datasets available for interactive visualization and
analysis has been widely recognized. However, existing solutions build on
dedicated servers to generate virtual slices that are transferred to the client
applications, practically leading to insufficient frame rates (less than 10
frames per second) for truly interactive experience. An easily accessible open
source solution for interactive arbitrary virtual re-slicing of very large
volumes and time series of volumes has yet been missing. We fill this gap with
BigDataViewer, a Fiji plugin to interactively navigate and visualize large
image sequences from both local and remote data sources.
| no_new_dataset | 0.93744 |
1412.0630 | Sean Anderson | Sean Anderson, Timothy D. Barfoot, Chi Hay Tong, Simo S\"arkk\"a | Batch Nonlinear Continuous-Time Trajectory Estimation as Exactly Sparse
Gaussian Process Regression | Submitted to Autonomous Robots on 20 November 2014, manuscript #
AURO-D-14-00185, 16 pages, 7 figures | null | null | null | cs.RO | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | In this paper, we revisit batch state estimation through the lens of Gaussian
process (GP) regression. We consider continuous-discrete estimation problems
wherein a trajectory is viewed as a one-dimensional GP, with time as the
independent variable. Our continuous-time prior can be defined by any
nonlinear, time-varying stochastic differential equation driven by white noise;
this allows the possibility of smoothing our trajectory estimates using a
variety of vehicle dynamics models (e.g., `constant-velocity'). We show that
this class of prior results in an inverse kernel matrix (i.e., covariance
matrix between all pairs of measurement times) that is exactly sparse
(block-tridiagonal) and that this can be exploited to carry out GP regression
(and interpolation) very efficiently. When the prior is based on a linear,
time-varying stochastic differential equation and the measurement model is also
linear, this GP approach is equivalent to classical, discrete-time smoothing
(at the measurement times); when a nonlinearity is present, we iterate over the
whole trajectory to maximize accuracy. We test the approach experimentally on a
simultaneous trajectory estimation and mapping problem using a mobile robot
dataset.
| [
{
"version": "v1",
"created": "Mon, 1 Dec 2014 20:24:08 GMT"
}
] | 2014-12-02T00:00:00 | [
[
"Anderson",
"Sean",
""
],
[
"Barfoot",
"Timothy D.",
""
],
[
"Tong",
"Chi Hay",
""
],
[
"Särkkä",
"Simo",
""
]
] | TITLE: Batch Nonlinear Continuous-Time Trajectory Estimation as Exactly Sparse
Gaussian Process Regression
ABSTRACT: In this paper, we revisit batch state estimation through the lens of Gaussian
process (GP) regression. We consider continuous-discrete estimation problems
wherein a trajectory is viewed as a one-dimensional GP, with time as the
independent variable. Our continuous-time prior can be defined by any
nonlinear, time-varying stochastic differential equation driven by white noise;
this allows the possibility of smoothing our trajectory estimates using a
variety of vehicle dynamics models (e.g., `constant-velocity'). We show that
this class of prior results in an inverse kernel matrix (i.e., covariance
matrix between all pairs of measurement times) that is exactly sparse
(block-tridiagonal) and that this can be exploited to carry out GP regression
(and interpolation) very efficiently. When the prior is based on a linear,
time-varying stochastic differential equation and the measurement model is also
linear, this GP approach is equivalent to classical, discrete-time smoothing
(at the measurement times); when a nonlinearity is present, we iterate over the
whole trajectory to maximize accuracy. We test the approach experimentally on a
simultaneous trajectory estimation and mapping problem using a mobile robot
dataset.
| no_new_dataset | 0.94887 |
1411.7441 | Stefano Ermon | Stefano Ermon, Ronan Le Bras, Santosh K. Suram, John M. Gregoire,
Carla Gomes, Bart Selman, Robert B. van Dover | Pattern Decomposition with Complex Combinatorial Constraints:
Application to Materials Discovery | null | null | null | null | cs.AI cs.LG stat.ML | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | Identifying important components or factors in large amounts of noisy data is
a key problem in machine learning and data mining. Motivated by a pattern
decomposition problem in materials discovery, aimed at discovering new
materials for renewable energy, e.g. for fuel and solar cells, we introduce
CombiFD, a framework for factor based pattern decomposition that allows the
incorporation of a-priori knowledge as constraints, including complex
combinatorial constraints. In addition, we propose a new pattern decomposition
algorithm, called AMIQO, based on solving a sequence of (mixed-integer)
quadratic programs. Our approach considerably outperforms the state of the art
on the materials discovery problem, scaling to larger datasets and recovering
more precise and physically meaningful decompositions. We also show the
effectiveness of our approach for enforcing background knowledge on other
application domains.
| [
{
"version": "v1",
"created": "Thu, 27 Nov 2014 02:31:41 GMT"
}
] | 2014-12-01T00:00:00 | [
[
"Ermon",
"Stefano",
""
],
[
"Bras",
"Ronan Le",
""
],
[
"Suram",
"Santosh K.",
""
],
[
"Gregoire",
"John M.",
""
],
[
"Gomes",
"Carla",
""
],
[
"Selman",
"Bart",
""
],
[
"van Dover",
"Robert B.",
""
]
] | TITLE: Pattern Decomposition with Complex Combinatorial Constraints:
Application to Materials Discovery
ABSTRACT: Identifying important components or factors in large amounts of noisy data is
a key problem in machine learning and data mining. Motivated by a pattern
decomposition problem in materials discovery, aimed at discovering new
materials for renewable energy, e.g. for fuel and solar cells, we introduce
CombiFD, a framework for factor based pattern decomposition that allows the
incorporation of a-priori knowledge as constraints, including complex
combinatorial constraints. In addition, we propose a new pattern decomposition
algorithm, called AMIQO, based on solving a sequence of (mixed-integer)
quadratic programs. Our approach considerably outperforms the state of the art
on the materials discovery problem, scaling to larger datasets and recovering
more precise and physically meaningful decompositions. We also show the
effectiveness of our approach for enforcing background knowledge on other
application domains.
| no_new_dataset | 0.942665 |
1411.7445 | Chao Xu | Tao Han, Chao Xu, Ryan Loxton, Lei Xie | Bi-objective Optimization for Robust RGB-D Visual Odometry | null | null | null | null | cs.RO cs.CV | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | This paper considers a new bi-objective optimization formulation for robust
RGB-D visual odometry. We investigate two methods for solving the proposed
bi-objective optimization problem: the weighted sum method (in which the
objective functions are combined into a single objective function) and the
bounded objective method (in which one of the objective functions is optimized
and the value of the other objective function is bounded via a constraint). Our
experimental results for the open source TUM RGB-D dataset show that the new
bi-objective optimization formulation is superior to several existing RGB-D
odometry methods. In particular, the new formulation yields more accurate
motion estimates and is more robust when textural or structural features in the
image sequence are lacking.
| [
{
"version": "v1",
"created": "Thu, 27 Nov 2014 02:37:41 GMT"
}
] | 2014-12-01T00:00:00 | [
[
"Han",
"Tao",
""
],
[
"Xu",
"Chao",
""
],
[
"Loxton",
"Ryan",
""
],
[
"Xie",
"Lei",
""
]
] | TITLE: Bi-objective Optimization for Robust RGB-D Visual Odometry
ABSTRACT: This paper considers a new bi-objective optimization formulation for robust
RGB-D visual odometry. We investigate two methods for solving the proposed
bi-objective optimization problem: the weighted sum method (in which the
objective functions are combined into a single objective function) and the
bounded objective method (in which one of the objective functions is optimized
and the value of the other objective function is bounded via a constraint). Our
experimental results for the open source TUM RGB-D dataset show that the new
bi-objective optimization formulation is superior to several existing RGB-D
odometry methods. In particular, the new formulation yields more accurate
motion estimates and is more robust when textural or structural features in the
image sequence are lacking.
| no_new_dataset | 0.951188 |
1411.7466 | Chunhua Shen | Lingqiao Liu, Chunhua Shen, Anton van den Hengel | The Treasure beneath Convolutional Layers: Cross-convolutional-layer
Pooling for Image Classification | null | null | null | null | cs.CV | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | A number of recent studies have shown that a Deep Convolutional Neural
Network (DCNN) pretrained on a large dataset can be adopted as a universal
image description which leads to astounding performance in many visual
classification tasks. Most of these studies, if not all, adopt activations of
the fully-connected layer of a DCNN as the image or region representation and
it is believed that convolutional layer activations are less discriminative.
This paper, however, advocates that if used appropriately convolutional layer
activations can be turned into a powerful image representation which enjoys
many advantages over fully-connected layer activations. This is achieved by
adopting a new technique proposed in this paper called
cross-convolutional-layer pooling. More specifically, it extracts subarrays of
feature maps of one convolutional layer as local features and pools the
extracted features with the guidance of feature maps of the successive
convolutional layer. Compared with exising methods that apply DCNNs in the
local feature setting, the proposed method is significantly faster since it
requires much fewer times of DCNN forward computation. Moreover, it avoids the
domain mismatch issue which is usually encountered when applying fully
connected layer activations to describe local regions. By applying our method
to four popular visual classification tasks, it is demonstrated that the
proposed method can achieve comparable or in some cases significantly better
performance than existing fully-connected layer based image representations
while incurring much lower computational cost.
| [
{
"version": "v1",
"created": "Thu, 27 Nov 2014 04:12:57 GMT"
}
] | 2014-12-01T00:00:00 | [
[
"Liu",
"Lingqiao",
""
],
[
"Shen",
"Chunhua",
""
],
[
"Hengel",
"Anton van den",
""
]
] | TITLE: The Treasure beneath Convolutional Layers: Cross-convolutional-layer
Pooling for Image Classification
ABSTRACT: A number of recent studies have shown that a Deep Convolutional Neural
Network (DCNN) pretrained on a large dataset can be adopted as a universal
image description which leads to astounding performance in many visual
classification tasks. Most of these studies, if not all, adopt activations of
the fully-connected layer of a DCNN as the image or region representation and
it is believed that convolutional layer activations are less discriminative.
This paper, however, advocates that if used appropriately convolutional layer
activations can be turned into a powerful image representation which enjoys
many advantages over fully-connected layer activations. This is achieved by
adopting a new technique proposed in this paper called
cross-convolutional-layer pooling. More specifically, it extracts subarrays of
feature maps of one convolutional layer as local features and pools the
extracted features with the guidance of feature maps of the successive
convolutional layer. Compared with exising methods that apply DCNNs in the
local feature setting, the proposed method is significantly faster since it
requires much fewer times of DCNN forward computation. Moreover, it avoids the
domain mismatch issue which is usually encountered when applying fully
connected layer activations to describe local regions. By applying our method
to four popular visual classification tasks, it is demonstrated that the
proposed method can achieve comparable or in some cases significantly better
performance than existing fully-connected layer based image representations
while incurring much lower computational cost.
| no_new_dataset | 0.950365 |
1411.7469 | Sanjay Chakraborty | Lopamudra Dey and Sanjay Chakraborty | Canonical PSO Based k-Means Clustering Approach for Real Datasets | null | null | null | null | cs.DB | http://creativecommons.org/licenses/by/3.0/ | "Clustering" the significance and application of this technique is spread
over various fields. Clustering is an unsupervised process in data mining, that
is why the proper evaluation of the results and measuring the compactness and
separability of the clusters are important issues.The procedure of evaluating
the results of a clustering algorithm is known as cluster validity measure.
Different types of indexes are used to solve different types of problems and
indices selection depends on the kind of available data.This paper first
proposes Canonical PSO based K-means clustering algorithm and also analyses
some important clustering indices (intercluster, intracluster) and then
evaluates the effects of those indices on real-time air pollution
database,wholesale customer, wine, and vehicle datasets using typical K-means,
Canonical PSO based K-means, simple PSO based K-means,DBSCAN, and Hierarchical
clustering algorithms.This paper also describes the nature of the clusters and
finally compares the performances of these clustering algorithms according to
the validity assessment. It also defines which algorithm will be more desirable
among all these algorithms to make proper compact clusters on this particular
real life datasets. It actually deals with the behaviour of these clustering
algorithms with respect to validation indexes and represents their results of
evaluation in terms of mathematical and graphical forms.
| [
{
"version": "v1",
"created": "Thu, 27 Nov 2014 04:50:30 GMT"
}
] | 2014-12-01T00:00:00 | [
[
"Dey",
"Lopamudra",
""
],
[
"Chakraborty",
"Sanjay",
""
]
] | TITLE: Canonical PSO Based k-Means Clustering Approach for Real Datasets
ABSTRACT: "Clustering" the significance and application of this technique is spread
over various fields. Clustering is an unsupervised process in data mining, that
is why the proper evaluation of the results and measuring the compactness and
separability of the clusters are important issues.The procedure of evaluating
the results of a clustering algorithm is known as cluster validity measure.
Different types of indexes are used to solve different types of problems and
indices selection depends on the kind of available data.This paper first
proposes Canonical PSO based K-means clustering algorithm and also analyses
some important clustering indices (intercluster, intracluster) and then
evaluates the effects of those indices on real-time air pollution
database,wholesale customer, wine, and vehicle datasets using typical K-means,
Canonical PSO based K-means, simple PSO based K-means,DBSCAN, and Hierarchical
clustering algorithms.This paper also describes the nature of the clusters and
finally compares the performances of these clustering algorithms according to
the validity assessment. It also defines which algorithm will be more desirable
among all these algorithms to make proper compact clusters on this particular
real life datasets. It actually deals with the behaviour of these clustering
algorithms with respect to validation indexes and represents their results of
evaluation in terms of mathematical and graphical forms.
| no_new_dataset | 0.952618 |
1411.7474 | Deepinder Kaur Er. | Deepinder Kaur | A Comparative Study of Various Distance Measures for Software fault
prediction | 4 pages,2 figures,"Published with International Journal of Computer
Trends and Technology (IJCTT)" | International Journal of Computer Trends and Technology (IJCTT),
17(3): 117-120, Nov 2014 | 10.14445/22312803/IJCTT-V17P122 | null | cs.SE | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | Different distance measures have been used for efficiently predicting
software faults at early stages of software development. One stereotyped
approach for software fault prediction due to its computational efficiency is
K-means clustering, which partitions the dataset into K number of clusters
using any distance measure. Distance measures by using some metrics are used to
extract similar data objects which help in developing efficient algorithms for
clustering and classification. In this paper, we study K-means clustering with
three different distance measures Euclidean, Sorensen and Canberra by using
datasets that have been collected from NASA MDP (metrics data program) .Results
are displayed with the help of ROC curve. The experimental results shows that
K-means clustering with Sorensen distance is better than Euclidean distance and
Canberra distance.
| [
{
"version": "v1",
"created": "Thu, 27 Nov 2014 06:07:46 GMT"
}
] | 2014-12-01T00:00:00 | [
[
"Kaur",
"Deepinder",
""
]
] | TITLE: A Comparative Study of Various Distance Measures for Software fault
prediction
ABSTRACT: Different distance measures have been used for efficiently predicting
software faults at early stages of software development. One stereotyped
approach for software fault prediction due to its computational efficiency is
K-means clustering, which partitions the dataset into K number of clusters
using any distance measure. Distance measures by using some metrics are used to
extract similar data objects which help in developing efficient algorithms for
clustering and classification. In this paper, we study K-means clustering with
three different distance measures Euclidean, Sorensen and Canberra by using
datasets that have been collected from NASA MDP (metrics data program) .Results
are displayed with the help of ROC curve. The experimental results shows that
K-means clustering with Sorensen distance is better than Euclidean distance and
Canberra distance.
| no_new_dataset | 0.949763 |
1411.7923 | Dong Yi | Dong Yi, Zhen Lei, Shengcai Liao, and Stan Z. Li | Learning Face Representation from Scratch | null | null | null | null | cs.CV | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | Pushing by big data and deep convolutional neural network (CNN), the
performance of face recognition is becoming comparable to human. Using private
large scale training datasets, several groups achieve very high performance on
LFW, i.e., 97% to 99%. While there are many open source implementations of CNN,
none of large scale face dataset is publicly available. The current situation
in the field of face recognition is that data is more important than algorithm.
To solve this problem, this paper proposes a semi-automatical way to collect
face images from Internet and builds a large scale dataset containing about
10,000 subjects and 500,000 images, called CASIAWebFace. Based on the database,
we use a 11-layer CNN to learn discriminative representation and obtain
state-of-theart accuracy on LFW and YTF. The publication of CASIAWebFace will
attract more research groups entering this field and accelerate the development
of face recognition in the wild.
| [
{
"version": "v1",
"created": "Fri, 28 Nov 2014 16:05:18 GMT"
}
] | 2014-12-01T00:00:00 | [
[
"Yi",
"Dong",
""
],
[
"Lei",
"Zhen",
""
],
[
"Liao",
"Shengcai",
""
],
[
"Li",
"Stan Z.",
""
]
] | TITLE: Learning Face Representation from Scratch
ABSTRACT: Pushing by big data and deep convolutional neural network (CNN), the
performance of face recognition is becoming comparable to human. Using private
large scale training datasets, several groups achieve very high performance on
LFW, i.e., 97% to 99%. While there are many open source implementations of CNN,
none of large scale face dataset is publicly available. The current situation
in the field of face recognition is that data is more important than algorithm.
To solve this problem, this paper proposes a semi-automatical way to collect
face images from Internet and builds a large scale dataset containing about
10,000 subjects and 500,000 images, called CASIAWebFace. Based on the database,
we use a 11-layer CNN to learn discriminative representation and obtain
state-of-theart accuracy on LFW and YTF. The publication of CASIAWebFace will
attract more research groups entering this field and accelerate the development
of face recognition in the wild.
| new_dataset | 0.96525 |
1404.4114 | Matthew D. Hoffman | Matthew D. Hoffman and David M. Blei | Structured Stochastic Variational Inference | null | null | null | null | cs.LG | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | Stochastic variational inference makes it possible to approximate posterior
distributions induced by large datasets quickly using stochastic optimization.
The algorithm relies on the use of fully factorized variational distributions.
However, this "mean-field" independence approximation limits the fidelity of
the posterior approximation, and introduces local optima. We show how to relax
the mean-field approximation to allow arbitrary dependencies between global
parameters and local hidden variables, producing better parameter estimates by
reducing bias, sensitivity to local optima, and sensitivity to hyperparameters.
| [
{
"version": "v1",
"created": "Wed, 16 Apr 2014 00:12:03 GMT"
},
{
"version": "v2",
"created": "Tue, 25 Nov 2014 18:56:38 GMT"
},
{
"version": "v3",
"created": "Wed, 26 Nov 2014 04:14:16 GMT"
}
] | 2014-11-27T00:00:00 | [
[
"Hoffman",
"Matthew D.",
""
],
[
"Blei",
"David M.",
""
]
] | TITLE: Structured Stochastic Variational Inference
ABSTRACT: Stochastic variational inference makes it possible to approximate posterior
distributions induced by large datasets quickly using stochastic optimization.
The algorithm relies on the use of fully factorized variational distributions.
However, this "mean-field" independence approximation limits the fidelity of
the posterior approximation, and introduces local optima. We show how to relax
the mean-field approximation to allow arbitrary dependencies between global
parameters and local hidden variables, producing better parameter estimates by
reducing bias, sensitivity to local optima, and sensitivity to hyperparameters.
| no_new_dataset | 0.953188 |
1411.7336 | Mohammed alzaidi | Mohammed A. Talab, Siti Norul Huda Sheikh Abdullah, Bilal Bataineh | Edge direction matrixes-based local binar patterns descriptor for shape
pattern recognition | null | null | null | null | cs.CV cs.IR | http://creativecommons.org/licenses/by/3.0/ | Shapes and texture image recognition usage is an essential branch of pattern
recognition. It is made up of techniques that aim at extracting information
from images via human knowledge and works. Local Binary Pattern (LBP) ensures
encoding global and local information and scaling invariance by introducing a
look-up table to reflect the uniformity structure of an object. However, edge
direction matrixes (EDMS) only apply global invariant descriptor which employs
first and secondary order relationships. The main idea behind this methodology
is the need of improved recognition capabilities, a goal achieved by the
combinative use of these descriptors. This collaboration aims to make use of
the major advantages each one presents, by simultaneously complementing each
other, in order to elevate their weak points. By using multiple classifier
approaches such as random forest and multi-layer perceptron neural network, the
proposed combinative descriptor are compared with the state of the art
combinative methods based on Gray-Level Co-occurrence matrix (GLCM with EDMS),
LBP and moment invariant on four benchmark dataset MPEG-7 CE-Shape-1, KTH-TIPS
image, Enghlishfnt and Arabic calligraphy . The experiments have shown the
superiority of the introduced descriptor over the GLCM with EDMS, LBP and
moment invariants and other well-known descriptor such as Scale Invariant
Feature Transform from the literature.
| [
{
"version": "v1",
"created": "Wed, 26 Nov 2014 19:12:33 GMT"
}
] | 2014-11-27T00:00:00 | [
[
"Talab",
"Mohammed A.",
""
],
[
"Abdullah",
"Siti Norul Huda Sheikh",
""
],
[
"Bataineh",
"Bilal",
""
]
] | TITLE: Edge direction matrixes-based local binar patterns descriptor for shape
pattern recognition
ABSTRACT: Shapes and texture image recognition usage is an essential branch of pattern
recognition. It is made up of techniques that aim at extracting information
from images via human knowledge and works. Local Binary Pattern (LBP) ensures
encoding global and local information and scaling invariance by introducing a
look-up table to reflect the uniformity structure of an object. However, edge
direction matrixes (EDMS) only apply global invariant descriptor which employs
first and secondary order relationships. The main idea behind this methodology
is the need of improved recognition capabilities, a goal achieved by the
combinative use of these descriptors. This collaboration aims to make use of
the major advantages each one presents, by simultaneously complementing each
other, in order to elevate their weak points. By using multiple classifier
approaches such as random forest and multi-layer perceptron neural network, the
proposed combinative descriptor are compared with the state of the art
combinative methods based on Gray-Level Co-occurrence matrix (GLCM with EDMS),
LBP and moment invariant on four benchmark dataset MPEG-7 CE-Shape-1, KTH-TIPS
image, Enghlishfnt and Arabic calligraphy . The experiments have shown the
superiority of the introduced descriptor over the GLCM with EDMS, LBP and
moment invariants and other well-known descriptor such as Scale Invariant
Feature Transform from the literature.
| no_new_dataset | 0.949482 |
1406.5549 | Piotr Doll\'ar | Piotr Doll\'ar and C. Lawrence Zitnick | Fast Edge Detection Using Structured Forests | update corresponding to acceptance to PAMI | null | null | null | cs.CV | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | Edge detection is a critical component of many vision systems, including
object detectors and image segmentation algorithms. Patches of edges exhibit
well-known forms of local structure, such as straight lines or T-junctions. In
this paper we take advantage of the structure present in local image patches to
learn both an accurate and computationally efficient edge detector. We
formulate the problem of predicting local edge masks in a structured learning
framework applied to random decision forests. Our novel approach to learning
decision trees robustly maps the structured labels to a discrete space on which
standard information gain measures may be evaluated. The result is an approach
that obtains realtime performance that is orders of magnitude faster than many
competing state-of-the-art approaches, while also achieving state-of-the-art
edge detection results on the BSDS500 Segmentation dataset and NYU Depth
dataset. Finally, we show the potential of our approach as a general purpose
edge detector by showing our learned edge models generalize well across
datasets.
| [
{
"version": "v1",
"created": "Fri, 20 Jun 2014 22:28:29 GMT"
},
{
"version": "v2",
"created": "Tue, 25 Nov 2014 02:49:28 GMT"
}
] | 2014-11-26T00:00:00 | [
[
"Dollár",
"Piotr",
""
],
[
"Zitnick",
"C. Lawrence",
""
]
] | TITLE: Fast Edge Detection Using Structured Forests
ABSTRACT: Edge detection is a critical component of many vision systems, including
object detectors and image segmentation algorithms. Patches of edges exhibit
well-known forms of local structure, such as straight lines or T-junctions. In
this paper we take advantage of the structure present in local image patches to
learn both an accurate and computationally efficient edge detector. We
formulate the problem of predicting local edge masks in a structured learning
framework applied to random decision forests. Our novel approach to learning
decision trees robustly maps the structured labels to a discrete space on which
standard information gain measures may be evaluated. The result is an approach
that obtains realtime performance that is orders of magnitude faster than many
competing state-of-the-art approaches, while also achieving state-of-the-art
edge detection results on the BSDS500 Segmentation dataset and NYU Depth
dataset. Finally, we show the potential of our approach as a general purpose
edge detector by showing our learned edge models generalize well across
datasets.
| no_new_dataset | 0.952838 |
1410.7182 | Leon Derczynski | Leon Derczynski, Diana Maynard, Giuseppe Rizzo, Marieke van Erp,
Genevieve Gorrell, Rapha\"el Troncy, Johann Petrak, Kalina Bontcheva | Analysis of Named Entity Recognition and Linking for Tweets | 35 pages, accepted to journal Information Processing and Management | Information Processing & Management 51 (2), 32-49, 2014 | 10.1016/j.ipm.2014.10.006 | null | cs.CL | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | Applying natural language processing for mining and intelligent information
access to tweets (a form of microblog) is a challenging, emerging research
area. Unlike carefully authored news text and other longer content, tweets pose
a number of new challenges, due to their short, noisy, context-dependent, and
dynamic nature. Information extraction from tweets is typically performed in a
pipeline, comprising consecutive stages of language identification,
tokenisation, part-of-speech tagging, named entity recognition and entity
disambiguation (e.g. with respect to DBpedia). In this work, we describe a new
Twitter entity disambiguation dataset, and conduct an empirical analysis of
named entity recognition and disambiguation, investigating how robust a number
of state-of-the-art systems are on such noisy texts, what the main sources of
error are, and which problems should be further investigated to improve the
state of the art.
| [
{
"version": "v1",
"created": "Mon, 27 Oct 2014 11:09:36 GMT"
}
] | 2014-11-26T00:00:00 | [
[
"Derczynski",
"Leon",
""
],
[
"Maynard",
"Diana",
""
],
[
"Rizzo",
"Giuseppe",
""
],
[
"van Erp",
"Marieke",
""
],
[
"Gorrell",
"Genevieve",
""
],
[
"Troncy",
"Raphaël",
""
],
[
"Petrak",
"Johann",
""
],
[
"Bontcheva",
"Kalina",
""
]
] | TITLE: Analysis of Named Entity Recognition and Linking for Tweets
ABSTRACT: Applying natural language processing for mining and intelligent information
access to tweets (a form of microblog) is a challenging, emerging research
area. Unlike carefully authored news text and other longer content, tweets pose
a number of new challenges, due to their short, noisy, context-dependent, and
dynamic nature. Information extraction from tweets is typically performed in a
pipeline, comprising consecutive stages of language identification,
tokenisation, part-of-speech tagging, named entity recognition and entity
disambiguation (e.g. with respect to DBpedia). In this work, we describe a new
Twitter entity disambiguation dataset, and conduct an empirical analysis of
named entity recognition and disambiguation, investigating how robust a number
of state-of-the-art systems are on such noisy texts, what the main sources of
error are, and which problems should be further investigated to improve the
state of the art.
| new_dataset | 0.959039 |
1411.6777 | Debajyoti Mukhopadhyay Prof. | Laxmi Lahoti, Chaitali Chandankhede, Debajyoti Mukhopadhyay | Modified Apriori Approach for Evade Network Intrusion Detection System | 5 pages, 3 figures | null | null | null | cs.CR | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | Intrusion Detection System or IDS is a software or hardware tool that
repeatedly scans and monitors events that took place in a computer or a
network. A set of rules are used by Signature based Network Intrusion Detection
Systems or NIDS to detect hostile traffic in network segments or packets, which
are so important in detecting malicious and anomalous behaviour over the
network like known attacks that hackers look for new techniques to go unseen.
Sometime, a single failure at any layer will cause the NIDS to miss that
attack. To overcome this problem, a technique is used that will trigger a
failure in that layer. Such technique is known as Evasive technique. An Evasion
can be defined as any technique that modifies a visible attack into any other
form in order to stay away from being detect. The proposed system is used for
detecting attacks which are going on the network and also gives actual
categorization of attacks. The proposed system has advantage of getting low
false alarm rate and high detection rate. So that leads into decrease in
complexity and overhead on the system. The paper presents the Evasion technique
for customized apriori algorithm. The paper aims to make a new functional
structure to evade NIDS. This framework can be used to audit NIDS. This
framework shows that a proof of concept showing how to evade a self built NIDS
considering two publicly available datasets.
| [
{
"version": "v1",
"created": "Tue, 25 Nov 2014 09:16:01 GMT"
}
] | 2014-11-26T00:00:00 | [
[
"Lahoti",
"Laxmi",
""
],
[
"Chandankhede",
"Chaitali",
""
],
[
"Mukhopadhyay",
"Debajyoti",
""
]
] | TITLE: Modified Apriori Approach for Evade Network Intrusion Detection System
ABSTRACT: Intrusion Detection System or IDS is a software or hardware tool that
repeatedly scans and monitors events that took place in a computer or a
network. A set of rules are used by Signature based Network Intrusion Detection
Systems or NIDS to detect hostile traffic in network segments or packets, which
are so important in detecting malicious and anomalous behaviour over the
network like known attacks that hackers look for new techniques to go unseen.
Sometime, a single failure at any layer will cause the NIDS to miss that
attack. To overcome this problem, a technique is used that will trigger a
failure in that layer. Such technique is known as Evasive technique. An Evasion
can be defined as any technique that modifies a visible attack into any other
form in order to stay away from being detect. The proposed system is used for
detecting attacks which are going on the network and also gives actual
categorization of attacks. The proposed system has advantage of getting low
false alarm rate and high detection rate. So that leads into decrease in
complexity and overhead on the system. The paper presents the Evasion technique
for customized apriori algorithm. The paper aims to make a new functional
structure to evade NIDS. This framework can be used to audit NIDS. This
framework shows that a proof of concept showing how to evade a self built NIDS
considering two publicly available datasets.
| no_new_dataset | 0.943243 |
1411.6850 | Khalid Jebari hassani | Amina Dik, Khalid Jebari, Abdelaziz Bouroumi and Aziz Ettouhami | Similarity- based approach for outlier detection | International Journal of Computer Science Issues 2014 | null | null | null | cs.CV | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | This paper presents a new approach for detecting outliers by introducing the
notion of object's proximity. The main idea is that normal point has similar
characteristics with several neighbors. So the point in not an outlier if it
has a high degree of proximity and its neighbors are several. The performance
of this approach is illustrated through real datasets
| [
{
"version": "v1",
"created": "Tue, 25 Nov 2014 13:13:47 GMT"
}
] | 2014-11-26T00:00:00 | [
[
"Dik",
"Amina",
""
],
[
"Jebari",
"Khalid",
""
],
[
"Bouroumi",
"Abdelaziz",
""
],
[
"Ettouhami",
"Aziz",
""
]
] | TITLE: Similarity- based approach for outlier detection
ABSTRACT: This paper presents a new approach for detecting outliers by introducing the
notion of object's proximity. The main idea is that normal point has similar
characteristics with several neighbors. So the point in not an outlier if it
has a high degree of proximity and its neighbors are several. The performance
of this approach is illustrated through real datasets
| no_new_dataset | 0.955319 |
1411.6909 | Aaron Hertzmann | Hamid Izadinia, Ali Farhadi, Aaron Hertzmann, Matthew D. Hoffman | Image Classification and Retrieval from User-Supplied Tags | null | null | null | null | cs.CV | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | This paper proposes direct learning of image classification from
user-supplied tags, without filtering. Each tag is supplied by the user who
shared the image online. Enormous numbers of these tags are freely available
online, and they give insight about the image categories important to users and
to image classification. Our approach is complementary to the conventional
approach of manual annotation, which is extremely costly. We analyze of the
Flickr 100 Million Image dataset, making several useful observations about the
statistics of these tags. We introduce a large-scale robust classification
algorithm, in order to handle the inherent noise in these tags, and a
calibration procedure to better predict objective annotations. We show that
freely available, user-supplied tags can obtain similar or superior results to
large databases of costly manual annotations.
| [
{
"version": "v1",
"created": "Tue, 25 Nov 2014 16:17:09 GMT"
}
] | 2014-11-26T00:00:00 | [
[
"Izadinia",
"Hamid",
""
],
[
"Farhadi",
"Ali",
""
],
[
"Hertzmann",
"Aaron",
""
],
[
"Hoffman",
"Matthew D.",
""
]
] | TITLE: Image Classification and Retrieval from User-Supplied Tags
ABSTRACT: This paper proposes direct learning of image classification from
user-supplied tags, without filtering. Each tag is supplied by the user who
shared the image online. Enormous numbers of these tags are freely available
online, and they give insight about the image categories important to users and
to image classification. Our approach is complementary to the conventional
approach of manual annotation, which is extremely costly. We analyze of the
Flickr 100 Million Image dataset, making several useful observations about the
statistics of these tags. We introduce a large-scale robust classification
algorithm, in order to handle the inherent noise in these tags, and a
calibration procedure to better predict objective annotations. We show that
freely available, user-supplied tags can obtain similar or superior results to
large databases of costly manual annotations.
| no_new_dataset | 0.953449 |
1210.5268 | Jacob Eisenstein | Jacob Eisenstein, Brendan O'Connor, Noah A. Smith, Eric P. Xing | Diffusion of Lexical Change in Social Media | preprint of PLOS-ONE paper from November 2014; PLoS ONE 9(11) e113114 | null | 10.1371/journal.pone.0113114 | null | cs.CL cs.SI physics.soc-ph | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | Computer-mediated communication is driving fundamental changes in the nature
of written language. We investigate these changes by statistical analysis of a
dataset comprising 107 million Twitter messages (authored by 2.7 million unique
user accounts). Using a latent vector autoregressive model to aggregate across
thousands of words, we identify high-level patterns in diffusion of linguistic
change over the United States. Our model is robust to unpredictable changes in
Twitter's sampling rate, and provides a probabilistic characterization of the
relationship of macro-scale linguistic influence to a set of demographic and
geographic predictors. The results of this analysis offer support for prior
arguments that focus on geographical proximity and population size. However,
demographic similarity -- especially with regard to race -- plays an even more
central role, as cities with similar racial demographics are far more likely to
share linguistic influence. Rather than moving towards a single unified
"netspeak" dialect, language evolution in computer-mediated communication
reproduces existing fault lines in spoken American English.
| [
{
"version": "v1",
"created": "Thu, 18 Oct 2012 21:46:09 GMT"
},
{
"version": "v2",
"created": "Mon, 22 Oct 2012 01:40:54 GMT"
},
{
"version": "v3",
"created": "Tue, 23 Oct 2012 21:15:56 GMT"
},
{
"version": "v4",
"created": "Mon, 24 Nov 2014 03:34:24 GMT"
}
] | 2014-11-25T00:00:00 | [
[
"Eisenstein",
"Jacob",
""
],
[
"O'Connor",
"Brendan",
""
],
[
"Smith",
"Noah A.",
""
],
[
"Xing",
"Eric P.",
""
]
] | TITLE: Diffusion of Lexical Change in Social Media
ABSTRACT: Computer-mediated communication is driving fundamental changes in the nature
of written language. We investigate these changes by statistical analysis of a
dataset comprising 107 million Twitter messages (authored by 2.7 million unique
user accounts). Using a latent vector autoregressive model to aggregate across
thousands of words, we identify high-level patterns in diffusion of linguistic
change over the United States. Our model is robust to unpredictable changes in
Twitter's sampling rate, and provides a probabilistic characterization of the
relationship of macro-scale linguistic influence to a set of demographic and
geographic predictors. The results of this analysis offer support for prior
arguments that focus on geographical proximity and population size. However,
demographic similarity -- especially with regard to race -- plays an even more
central role, as cities with similar racial demographics are far more likely to
share linguistic influence. Rather than moving towards a single unified
"netspeak" dialect, language evolution in computer-mediated communication
reproduces existing fault lines in spoken American English.
| no_new_dataset | 0.874185 |
1411.6235 | Xiaojun Chang | Xiaojun Chang, Feiping Nie, Zhigang Ma, and Yi Yang | Balanced k-Means and Min-Cut Clustering | null | null | null | null | cs.LG | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | Clustering is an effective technique in data mining to generate groups that
are the matter of interest. Among various clustering approaches, the family of
k-means algorithms and min-cut algorithms gain most popularity due to their
simplicity and efficacy. The classical k-means algorithm partitions a number of
data points into several subsets by iteratively updating the clustering centers
and the associated data points. By contrast, a weighted undirected graph is
constructed in min-cut algorithms which partition the vertices of the graph
into two sets. However, existing clustering algorithms tend to cluster minority
of data points into a subset, which shall be avoided when the target dataset is
balanced. To achieve more accurate clustering for balanced dataset, we propose
to leverage exclusive lasso on k-means and min-cut to regulate the balance
degree of the clustering results. By optimizing our objective functions that
build atop the exclusive lasso, we can make the clustering result as much
balanced as possible. Extensive experiments on several large-scale datasets
validate the advantage of the proposed algorithms compared to the
state-of-the-art clustering algorithms.
| [
{
"version": "v1",
"created": "Sun, 23 Nov 2014 13:16:25 GMT"
}
] | 2014-11-25T00:00:00 | [
[
"Chang",
"Xiaojun",
""
],
[
"Nie",
"Feiping",
""
],
[
"Ma",
"Zhigang",
""
],
[
"Yang",
"Yi",
""
]
] | TITLE: Balanced k-Means and Min-Cut Clustering
ABSTRACT: Clustering is an effective technique in data mining to generate groups that
are the matter of interest. Among various clustering approaches, the family of
k-means algorithms and min-cut algorithms gain most popularity due to their
simplicity and efficacy. The classical k-means algorithm partitions a number of
data points into several subsets by iteratively updating the clustering centers
and the associated data points. By contrast, a weighted undirected graph is
constructed in min-cut algorithms which partition the vertices of the graph
into two sets. However, existing clustering algorithms tend to cluster minority
of data points into a subset, which shall be avoided when the target dataset is
balanced. To achieve more accurate clustering for balanced dataset, we propose
to leverage exclusive lasso on k-means and min-cut to regulate the balance
degree of the clustering results. By optimizing our objective functions that
build atop the exclusive lasso, we can make the clustering result as much
balanced as possible. Extensive experiments on several large-scale datasets
validate the advantage of the proposed algorithms compared to the
state-of-the-art clustering algorithms.
| no_new_dataset | 0.951414 |
1411.6308 | Xiaojun Chang | Xiaojun Chang, Feiping Nie, Zhigang Ma, Yi Yang and Xiaofang Zhou | A Convex Formulation for Spectral Shrunk Clustering | AAAI2015 | null | null | null | cs.LG | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | Spectral clustering is a fundamental technique in the field of data mining
and information processing. Most existing spectral clustering algorithms
integrate dimensionality reduction into the clustering process assisted by
manifold learning in the original space. However, the manifold in
reduced-dimensional subspace is likely to exhibit altered properties in
contrast with the original space. Thus, applying manifold information obtained
from the original space to the clustering process in a low-dimensional subspace
is prone to inferior performance. Aiming to address this issue, we propose a
novel convex algorithm that mines the manifold structure in the low-dimensional
subspace. In addition, our unified learning process makes the manifold learning
particularly tailored for the clustering. Compared with other related methods,
the proposed algorithm results in more structured clustering result. To
validate the efficacy of the proposed algorithm, we perform extensive
experiments on several benchmark datasets in comparison with some
state-of-the-art clustering approaches. The experimental results demonstrate
that the proposed algorithm has quite promising clustering performance.
| [
{
"version": "v1",
"created": "Sun, 23 Nov 2014 22:12:52 GMT"
}
] | 2014-11-25T00:00:00 | [
[
"Chang",
"Xiaojun",
""
],
[
"Nie",
"Feiping",
""
],
[
"Ma",
"Zhigang",
""
],
[
"Yang",
"Yi",
""
],
[
"Zhou",
"Xiaofang",
""
]
] | TITLE: A Convex Formulation for Spectral Shrunk Clustering
ABSTRACT: Spectral clustering is a fundamental technique in the field of data mining
and information processing. Most existing spectral clustering algorithms
integrate dimensionality reduction into the clustering process assisted by
manifold learning in the original space. However, the manifold in
reduced-dimensional subspace is likely to exhibit altered properties in
contrast with the original space. Thus, applying manifold information obtained
from the original space to the clustering process in a low-dimensional subspace
is prone to inferior performance. Aiming to address this issue, we propose a
novel convex algorithm that mines the manifold structure in the low-dimensional
subspace. In addition, our unified learning process makes the manifold learning
particularly tailored for the clustering. Compared with other related methods,
the proposed algorithm results in more structured clustering result. To
validate the efficacy of the proposed algorithm, we perform extensive
experiments on several benchmark datasets in comparison with some
state-of-the-art clustering approaches. The experimental results demonstrate
that the proposed algorithm has quite promising clustering performance.
| no_new_dataset | 0.952838 |
1411.6447 | Tianjun Xiao | Tianjun Xiao, Yichong Xu, Kuiyuan Yang, Jiaxing Zhang, Yuxin Peng,
Zheng Zhang | The Application of Two-level Attention Models in Deep Convolutional
Neural Network for Fine-grained Image Classification | null | null | null | null | cs.CV | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | Fine-grained classification is challenging because categories can only be
discriminated by subtle and local differences. Variances in the pose, scale or
rotation usually make the problem more difficult. Most fine-grained
classification systems follow the pipeline of finding foreground object or
object parts (where) to extract discriminative features (what).
In this paper, we propose to apply visual attention to fine-grained
classification task using deep neural network. Our pipeline integrates three
types of attention: the bottom-up attention that propose candidate patches, the
object-level top-down attention that selects relevant patches to a certain
object, and the part-level top-down attention that localizes discriminative
parts. We combine these attentions to train domain-specific deep nets, then use
it to improve both the what and where aspects. Importantly, we avoid using
expensive annotations like bounding box or part information from end-to-end.
The weak supervision constraint makes our work easier to generalize.
We have verified the effectiveness of the method on the subsets of ILSVRC2012
dataset and CUB200_2011 dataset. Our pipeline delivered significant
improvements and achieved the best accuracy under the weakest supervision
condition. The performance is competitive against other methods that rely on
additional annotations.
| [
{
"version": "v1",
"created": "Mon, 24 Nov 2014 13:30:07 GMT"
}
] | 2014-11-25T00:00:00 | [
[
"Xiao",
"Tianjun",
""
],
[
"Xu",
"Yichong",
""
],
[
"Yang",
"Kuiyuan",
""
],
[
"Zhang",
"Jiaxing",
""
],
[
"Peng",
"Yuxin",
""
],
[
"Zhang",
"Zheng",
""
]
] | TITLE: The Application of Two-level Attention Models in Deep Convolutional
Neural Network for Fine-grained Image Classification
ABSTRACT: Fine-grained classification is challenging because categories can only be
discriminated by subtle and local differences. Variances in the pose, scale or
rotation usually make the problem more difficult. Most fine-grained
classification systems follow the pipeline of finding foreground object or
object parts (where) to extract discriminative features (what).
In this paper, we propose to apply visual attention to fine-grained
classification task using deep neural network. Our pipeline integrates three
types of attention: the bottom-up attention that propose candidate patches, the
object-level top-down attention that selects relevant patches to a certain
object, and the part-level top-down attention that localizes discriminative
parts. We combine these attentions to train domain-specific deep nets, then use
it to improve both the what and where aspects. Importantly, we avoid using
expensive annotations like bounding box or part information from end-to-end.
The weak supervision constraint makes our work easier to generalize.
We have verified the effectiveness of the method on the subsets of ILSVRC2012
dataset and CUB200_2011 dataset. Our pipeline delivered significant
improvements and achieved the best accuracy under the weakest supervision
condition. The performance is competitive against other methods that rely on
additional annotations.
| no_new_dataset | 0.948394 |
1411.6562 | Manas Joglekar | Manas Joglekar, Hector Garcia-Molina, Aditya Parameswaran | Evaluating the Crowd with Confidence | null | null | null | null | cs.DB | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | Worker quality control is a crucial aspect of crowdsourcing systems;
typically occupying a large fraction of the time and money invested on
crowdsourcing. In this work, we devise techniques to generate confidence
intervals for worker error rate estimates, thereby enabling a better evaluation
of worker quality. We show that our techniques generate correct confidence
intervals on a range of real-world datasets, and demonstrate wide applicability
by using them to evict poorly performing workers, and provide confidence
intervals on the accuracy of the answers.
| [
{
"version": "v1",
"created": "Wed, 12 Nov 2014 23:50:43 GMT"
}
] | 2014-11-25T00:00:00 | [
[
"Joglekar",
"Manas",
""
],
[
"Garcia-Molina",
"Hector",
""
],
[
"Parameswaran",
"Aditya",
""
]
] | TITLE: Evaluating the Crowd with Confidence
ABSTRACT: Worker quality control is a crucial aspect of crowdsourcing systems;
typically occupying a large fraction of the time and money invested on
crowdsourcing. In this work, we devise techniques to generate confidence
intervals for worker error rate estimates, thereby enabling a better evaluation
of worker quality. We show that our techniques generate correct confidence
intervals on a range of real-world datasets, and demonstrate wide applicability
by using them to evict poorly performing workers, and provide confidence
intervals on the accuracy of the answers.
| no_new_dataset | 0.959116 |
1405.0538 | Rados{\l}aw Michalski | Rados{\l}aw Michalski, Tomasz Kajdanowicz, Piotr Br\'odka,
Przemys{\l}aw Kazienko | Seed Selection for Spread of Influence in Social Networks: Temporal vs.
Static Approach | null | New Generation Computing, Vol. 32, Issue 3-4, pp. 213-235, 2014 | 10.1007/s00354-014-0402-9 | null | cs.SI physics.soc-ph | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | The problem of finding optimal set of users for influencing others in the
social network has been widely studied. Because it is NP-hard, some heuristics
were proposed to find sub-optimal solutions. Still, one of the commonly used
assumption is the one that seeds are chosen on the static network, not the
dynamic one. This static approach is in fact far from the real-world networks,
where new nodes may appear and old ones dynamically disappear in course of
time.
The main purpose of this paper is to analyse how the results of one of the
typical models for spread of influence - linear threshold - differ depending on
the strategy of building the social network used later for choosing seeds. To
show the impact of network creation strategy on the final number of influenced
nodes - outcome of spread of influence, the results for three approaches were
studied: one static and two temporal with different granularities, i.e. various
number of time windows. Social networks for each time window encapsulated
dynamic changes in the network structure. Calculation of various node
structural measures like degree or betweenness respected these changes by means
of forgetting mechanism - more recent data had greater influence on node
measure values. These measures were, in turn, used for node ranking and their
selection for seeding.
All concepts were applied to experimental verification on five real datasets.
The results revealed that temporal approach is always better than static and
the higher granularity in the temporal social network while seeding, the more
finally influenced nodes. Additionally, outdegree measure with exponential
forgetting typically outperformed other time-dependent structural measures, if
used for seed candidate ranking.
| [
{
"version": "v1",
"created": "Fri, 2 May 2014 23:32:04 GMT"
},
{
"version": "v2",
"created": "Fri, 21 Nov 2014 13:20:36 GMT"
}
] | 2014-11-24T00:00:00 | [
[
"Michalski",
"Radosław",
""
],
[
"Kajdanowicz",
"Tomasz",
""
],
[
"Bródka",
"Piotr",
""
],
[
"Kazienko",
"Przemysław",
""
]
] | TITLE: Seed Selection for Spread of Influence in Social Networks: Temporal vs.
Static Approach
ABSTRACT: The problem of finding optimal set of users for influencing others in the
social network has been widely studied. Because it is NP-hard, some heuristics
were proposed to find sub-optimal solutions. Still, one of the commonly used
assumption is the one that seeds are chosen on the static network, not the
dynamic one. This static approach is in fact far from the real-world networks,
where new nodes may appear and old ones dynamically disappear in course of
time.
The main purpose of this paper is to analyse how the results of one of the
typical models for spread of influence - linear threshold - differ depending on
the strategy of building the social network used later for choosing seeds. To
show the impact of network creation strategy on the final number of influenced
nodes - outcome of spread of influence, the results for three approaches were
studied: one static and two temporal with different granularities, i.e. various
number of time windows. Social networks for each time window encapsulated
dynamic changes in the network structure. Calculation of various node
structural measures like degree or betweenness respected these changes by means
of forgetting mechanism - more recent data had greater influence on node
measure values. These measures were, in turn, used for node ranking and their
selection for seeding.
All concepts were applied to experimental verification on five real datasets.
The results revealed that temporal approach is always better than static and
the higher granularity in the temporal social network while seeding, the more
finally influenced nodes. Additionally, outdegree measure with exponential
forgetting typically outperformed other time-dependent structural measures, if
used for seed candidate ranking.
| no_new_dataset | 0.958148 |
1410.4510 | Finale Doshi-Velez | Finale Doshi-Velez and Byron Wallace and Ryan Adams | Graph-Sparse LDA: A Topic Model with Structured Sparsity | null | null | null | null | stat.ML cs.CL cs.LG | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | Originally designed to model text, topic modeling has become a powerful tool
for uncovering latent structure in domains including medicine, finance, and
vision. The goals for the model vary depending on the application: in some
cases, the discovered topics may be used for prediction or some other
downstream task. In other cases, the content of the topic itself may be of
intrinsic scientific interest.
Unfortunately, even using modern sparse techniques, the discovered topics are
often difficult to interpret due to the high dimensionality of the underlying
space. To improve topic interpretability, we introduce Graph-Sparse LDA, a
hierarchical topic model that leverages knowledge of relationships between
words (e.g., as encoded by an ontology). In our model, topics are summarized by
a few latent concept-words from the underlying graph that explain the observed
words. Graph-Sparse LDA recovers sparse, interpretable summaries on two
real-world biomedical datasets while matching state-of-the-art prediction
performance.
| [
{
"version": "v1",
"created": "Thu, 16 Oct 2014 17:35:31 GMT"
},
{
"version": "v2",
"created": "Fri, 21 Nov 2014 16:38:59 GMT"
}
] | 2014-11-24T00:00:00 | [
[
"Doshi-Velez",
"Finale",
""
],
[
"Wallace",
"Byron",
""
],
[
"Adams",
"Ryan",
""
]
] | TITLE: Graph-Sparse LDA: A Topic Model with Structured Sparsity
ABSTRACT: Originally designed to model text, topic modeling has become a powerful tool
for uncovering latent structure in domains including medicine, finance, and
vision. The goals for the model vary depending on the application: in some
cases, the discovered topics may be used for prediction or some other
downstream task. In other cases, the content of the topic itself may be of
intrinsic scientific interest.
Unfortunately, even using modern sparse techniques, the discovered topics are
often difficult to interpret due to the high dimensionality of the underlying
space. To improve topic interpretability, we introduce Graph-Sparse LDA, a
hierarchical topic model that leverages knowledge of relationships between
words (e.g., as encoded by an ontology). In our model, topics are summarized by
a few latent concept-words from the underlying graph that explain the observed
words. Graph-Sparse LDA recovers sparse, interpretable summaries on two
real-world biomedical datasets while matching state-of-the-art prediction
performance.
| no_new_dataset | 0.952086 |
1411.5428 | Ben Stoddard | Ben Stoddard and Yan Chen and Ashwin Machanavajjhala | Differentially Private Algorithms for Empirical Machine Learning | null | null | null | null | cs.LG | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | An important use of private data is to build machine learning classifiers.
While there is a burgeoning literature on differentially private classification
algorithms, we find that they are not practical in real applications due to two
reasons. First, existing differentially private classifiers provide poor
accuracy on real world datasets. Second, there is no known differentially
private algorithm for empirically evaluating the private classifier on a
private test dataset.
In this paper, we develop differentially private algorithms that mirror real
world empirical machine learning workflows. We consider the private classifier
training algorithm as a blackbox. We present private algorithms for selecting
features that are input to the classifier. Though adding a preprocessing step
takes away some of the privacy budget from the actual classification process
(thus potentially making it noisier and less accurate), we show that our novel
preprocessing techniques significantly increase classifier accuracy on three
real-world datasets. We also present the first private algorithms for
empirically constructing receiver operating characteristic (ROC) curves on a
private test set.
| [
{
"version": "v1",
"created": "Thu, 20 Nov 2014 03:10:47 GMT"
},
{
"version": "v2",
"created": "Fri, 21 Nov 2014 20:41:04 GMT"
}
] | 2014-11-24T00:00:00 | [
[
"Stoddard",
"Ben",
""
],
[
"Chen",
"Yan",
""
],
[
"Machanavajjhala",
"Ashwin",
""
]
] | TITLE: Differentially Private Algorithms for Empirical Machine Learning
ABSTRACT: An important use of private data is to build machine learning classifiers.
While there is a burgeoning literature on differentially private classification
algorithms, we find that they are not practical in real applications due to two
reasons. First, existing differentially private classifiers provide poor
accuracy on real world datasets. Second, there is no known differentially
private algorithm for empirically evaluating the private classifier on a
private test dataset.
In this paper, we develop differentially private algorithms that mirror real
world empirical machine learning workflows. We consider the private classifier
training algorithm as a blackbox. We present private algorithms for selecting
features that are input to the classifier. Though adding a preprocessing step
takes away some of the privacy budget from the actual classification process
(thus potentially making it noisier and less accurate), we show that our novel
preprocessing techniques significantly increase classifier accuracy on three
real-world datasets. We also present the first private algorithms for
empirically constructing receiver operating characteristic (ROC) curves on a
private test set.
| no_new_dataset | 0.948394 |
1411.5731 | Suleyman Cetintas | Can Xu, Suleyman Cetintas, Kuang-Chih Lee, Li-Jia Li | Visual Sentiment Prediction with Deep Convolutional Neural Networks | null | null | null | null | cs.CV cs.NE stat.ML | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | Images have become one of the most popular types of media through which users
convey their emotions within online social networks. Although vast amount of
research is devoted to sentiment analysis of textual data, there has been very
limited work that focuses on analyzing sentiment of image data. In this work,
we propose a novel visual sentiment prediction framework that performs image
understanding with Deep Convolutional Neural Networks (CNN). Specifically, the
proposed sentiment prediction framework performs transfer learning from a CNN
with millions of parameters, which is pre-trained on large-scale data for
object recognition. Experiments conducted on two real-world datasets from
Twitter and Tumblr demonstrate the effectiveness of the proposed visual
sentiment analysis framework.
| [
{
"version": "v1",
"created": "Fri, 21 Nov 2014 00:39:43 GMT"
}
] | 2014-11-24T00:00:00 | [
[
"Xu",
"Can",
""
],
[
"Cetintas",
"Suleyman",
""
],
[
"Lee",
"Kuang-Chih",
""
],
[
"Li",
"Li-Jia",
""
]
] | TITLE: Visual Sentiment Prediction with Deep Convolutional Neural Networks
ABSTRACT: Images have become one of the most popular types of media through which users
convey their emotions within online social networks. Although vast amount of
research is devoted to sentiment analysis of textual data, there has been very
limited work that focuses on analyzing sentiment of image data. In this work,
we propose a novel visual sentiment prediction framework that performs image
understanding with Deep Convolutional Neural Networks (CNN). Specifically, the
proposed sentiment prediction framework performs transfer learning from a CNN
with millions of parameters, which is pre-trained on large-scale data for
object recognition. Experiments conducted on two real-world datasets from
Twitter and Tumblr demonstrate the effectiveness of the proposed visual
sentiment analysis framework.
| no_new_dataset | 0.952309 |
1411.5935 | M. Zeeshan Zia | M.Zeeshan Zia, Michael Stark, Konrad Schindler | Towards Scene Understanding with Detailed 3D Object Representations | International Journal of Computer Vision (appeared online on 4
November 2014). Online version:
http://link.springer.com/article/10.1007/s11263-014-0780-y | null | 10.1007/s11263-014-0780-y | null | cs.CV | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | Current approaches to semantic image and scene understanding typically employ
rather simple object representations such as 2D or 3D bounding boxes. While
such coarse models are robust and allow for reliable object detection, they
discard much of the information about objects' 3D shape and pose, and thus do
not lend themselves well to higher-level reasoning. Here, we propose to base
scene understanding on a high-resolution object representation. An object class
- in our case cars - is modeled as a deformable 3D wireframe, which enables
fine-grained modeling at the level of individual vertices and faces. We augment
that model to explicitly include vertex-level occlusion, and embed all
instances in a common coordinate frame, in order to infer and exploit
object-object interactions. Specifically, from a single view we jointly
estimate the shapes and poses of multiple objects in a common 3D frame. A
ground plane in that frame is estimated by consensus among different objects,
which significantly stabilizes monocular 3D pose estimation. The fine-grained
model, in conjunction with the explicit 3D scene model, further allows one to
infer part-level occlusions between the modeled objects, as well as occlusions
by other, unmodeled scene elements. To demonstrate the benefits of such
detailed object class models in the context of scene understanding we
systematically evaluate our approach on the challenging KITTI street scene
dataset. The experiments show that the model's ability to utilize image
evidence at the level of individual parts improves monocular 3D pose estimation
w.r.t. both location and (continuous) viewpoint.
| [
{
"version": "v1",
"created": "Tue, 18 Nov 2014 15:07:19 GMT"
}
] | 2014-11-24T00:00:00 | [
[
"Zia",
"M. Zeeshan",
""
],
[
"Stark",
"Michael",
""
],
[
"Schindler",
"Konrad",
""
]
] | TITLE: Towards Scene Understanding with Detailed 3D Object Representations
ABSTRACT: Current approaches to semantic image and scene understanding typically employ
rather simple object representations such as 2D or 3D bounding boxes. While
such coarse models are robust and allow for reliable object detection, they
discard much of the information about objects' 3D shape and pose, and thus do
not lend themselves well to higher-level reasoning. Here, we propose to base
scene understanding on a high-resolution object representation. An object class
- in our case cars - is modeled as a deformable 3D wireframe, which enables
fine-grained modeling at the level of individual vertices and faces. We augment
that model to explicitly include vertex-level occlusion, and embed all
instances in a common coordinate frame, in order to infer and exploit
object-object interactions. Specifically, from a single view we jointly
estimate the shapes and poses of multiple objects in a common 3D frame. A
ground plane in that frame is estimated by consensus among different objects,
which significantly stabilizes monocular 3D pose estimation. The fine-grained
model, in conjunction with the explicit 3D scene model, further allows one to
infer part-level occlusions between the modeled objects, as well as occlusions
by other, unmodeled scene elements. To demonstrate the benefits of such
detailed object class models in the context of scene understanding we
systematically evaluate our approach on the challenging KITTI street scene
dataset. The experiments show that the model's ability to utilize image
evidence at the level of individual parts improves monocular 3D pose estimation
w.r.t. both location and (continuous) viewpoint.
| no_new_dataset | 0.944536 |
1411.5995 | George Ovchinnikov | G.V. Ovchinnikov, D.A. Kolesnikov, I.V. Oseledets | Algebraic reputation model RepRank and its application to spambot
detection | null | null | null | null | cs.SI | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | Due to popularity surge social networks became lucrative targets for spammers
and guerilla marketers, who are trying to game ranking systems and broadcast
their messages at little to none cost. Ranking systems, for example Twitter's
Trends, can be gamed by scripted users also called bots, who are automatically
or semi-automatically twitting essentially the same message. Judging by the
prices and abundance of supply from PR firms this is an easy to implement and
widely used tactic, at least in Russian blogosphere. Aggregative analysis of
social networks should at best mark those messages as spam or at least
correctly downplay their importance as they represent opinions only of a few,
if dedicated, users. Hence bot detection plays a crucial role in social network
mining and analysis. In this paper we propose technique called RepRank which
could be viewed as Markov chain based model for reputation propagation on
graphs utilizing simultaneous trust and anti-trust propagation and provide
effective numerical approach for its computation. Comparison with another
models such as TrustRank and some of its modifications on sample of 320000
Russian speaking Twitter users is presented. The dataset is presented as well.
| [
{
"version": "v1",
"created": "Thu, 20 Nov 2014 13:50:39 GMT"
}
] | 2014-11-24T00:00:00 | [
[
"Ovchinnikov",
"G. V.",
""
],
[
"Kolesnikov",
"D. A.",
""
],
[
"Oseledets",
"I. V.",
""
]
] | TITLE: Algebraic reputation model RepRank and its application to spambot
detection
ABSTRACT: Due to popularity surge social networks became lucrative targets for spammers
and guerilla marketers, who are trying to game ranking systems and broadcast
their messages at little to none cost. Ranking systems, for example Twitter's
Trends, can be gamed by scripted users also called bots, who are automatically
or semi-automatically twitting essentially the same message. Judging by the
prices and abundance of supply from PR firms this is an easy to implement and
widely used tactic, at least in Russian blogosphere. Aggregative analysis of
social networks should at best mark those messages as spam or at least
correctly downplay their importance as they represent opinions only of a few,
if dedicated, users. Hence bot detection plays a crucial role in social network
mining and analysis. In this paper we propose technique called RepRank which
could be viewed as Markov chain based model for reputation propagation on
graphs utilizing simultaneous trust and anti-trust propagation and provide
effective numerical approach for its computation. Comparison with another
models such as TrustRank and some of its modifications on sample of 320000
Russian speaking Twitter users is presented. The dataset is presented as well.
| new_dataset | 0.666049 |
1307.3176 | L.A. Prashanth | Nathaniel Korda, Prashanth L.A. and R\'emi Munos | Fast gradient descent for drifting least squares regression, with
application to bandits | null | null | null | null | cs.LG stat.ML | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | Online learning algorithms require to often recompute least squares
regression estimates of parameters. We study improving the computational
complexity of such algorithms by using stochastic gradient descent (SGD) type
schemes in place of classic regression solvers. We show that SGD schemes
efficiently track the true solutions of the regression problems, even in the
presence of a drift. This finding coupled with an $O(d)$ improvement in
complexity, where $d$ is the dimension of the data, make them attractive for
implementation in the big data settings. In the case when strong convexity in
the regression problem is guaranteed, we provide bounds on the error both in
expectation and high probability (the latter is often needed to provide
theoretical guarantees for higher level algorithms), despite the drifting least
squares solution. As an example of this case we prove that the regret
performance of an SGD version of the PEGE linear bandit algorithm
[Rusmevichientong and Tsitsiklis 2010] is worse that that of PEGE itself only
by a factor of $O(\log^4 n)$. When strong convexity of the regression problem
cannot be guaranteed, we investigate using an adaptive regularisation. We make
an empirical study of an adaptively regularised, SGD version of LinUCB [Li et
al. 2010] in a news article recommendation application, which uses the large
scale news recommendation dataset from Yahoo! front page. These experiments
show a large gain in computational complexity, with a consistently low tracking
error and click-through-rate (CTR) performance that is $75\%$ close.
| [
{
"version": "v1",
"created": "Thu, 11 Jul 2013 16:36:29 GMT"
},
{
"version": "v2",
"created": "Wed, 19 Feb 2014 00:27:18 GMT"
},
{
"version": "v3",
"created": "Thu, 24 Jul 2014 14:29:52 GMT"
},
{
"version": "v4",
"created": "Thu, 20 Nov 2014 12:40:48 GMT"
}
] | 2014-11-21T00:00:00 | [
[
"Korda",
"Nathaniel",
""
],
[
"A.",
"Prashanth L.",
""
],
[
"Munos",
"Rémi",
""
]
] | TITLE: Fast gradient descent for drifting least squares regression, with
application to bandits
ABSTRACT: Online learning algorithms require to often recompute least squares
regression estimates of parameters. We study improving the computational
complexity of such algorithms by using stochastic gradient descent (SGD) type
schemes in place of classic regression solvers. We show that SGD schemes
efficiently track the true solutions of the regression problems, even in the
presence of a drift. This finding coupled with an $O(d)$ improvement in
complexity, where $d$ is the dimension of the data, make them attractive for
implementation in the big data settings. In the case when strong convexity in
the regression problem is guaranteed, we provide bounds on the error both in
expectation and high probability (the latter is often needed to provide
theoretical guarantees for higher level algorithms), despite the drifting least
squares solution. As an example of this case we prove that the regret
performance of an SGD version of the PEGE linear bandit algorithm
[Rusmevichientong and Tsitsiklis 2010] is worse that that of PEGE itself only
by a factor of $O(\log^4 n)$. When strong convexity of the regression problem
cannot be guaranteed, we investigate using an adaptive regularisation. We make
an empirical study of an adaptively regularised, SGD version of LinUCB [Li et
al. 2010] in a news article recommendation application, which uses the large
scale news recommendation dataset from Yahoo! front page. These experiments
show a large gain in computational complexity, with a consistently low tracking
error and click-through-rate (CTR) performance that is $75\%$ close.
| no_new_dataset | 0.937783 |
1004.3499 | Lee Samuel Finn | Lee Samuel Finn, Andrea N. Lommen | Detection, Localization and Characterization of Gravitational Wave
Bursts in a Pulsar Timing Array | 43 pages, 13 figures, submitted to ApJ. | Astrophys.J.718:1400-1415,2010 | 10.1088/0004-637X/718/2/1400 | null | astro-ph.IM gr-qc physics.data-an | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | Efforts to detect gravitational waves by timing an array of pulsars have
focused traditionally on stationary gravitational waves: e.g., stochastic or
periodic signals. Gravitational wave bursts --- signals whose duration is much
shorter than the observation period --- will also arise in the pulsar timing
array waveband. Sources that give rise to detectable bursts include the
formation or coalescence of supermassive black holes (SMBHs), the periapsis
passage of compact objects in highly elliptic or unbound orbits about a SMBH,
or cusps on cosmic strings. Here we describe how pulsar timing array data may
be analyzed to detect and characterize these bursts. Our analysis addresses, in
a mutually consistent manner, a hierarchy of three questions: \emph{i}) What
are the odds that a dataset includes the signal from a gravitational wave
burst? \emph{ii}) Assuming the presence of a burst, what is the direction to
its source? and \emph{iii}) Assuming the burst propagation direction, what is
the burst waveform's time dependence in each of its polarization states?
Applying our analysis to synthetic data sets we find that we can \emph{detect}
gravitational waves even when the radiation is too weak to either localize the
source of infer the waveform, and \emph{detect} and \emph{localize} sources
even when the radiation amplitude is too weak to permit the waveform to be
determined. While the context of our discussion is gravitational wave detection
via pulsar timing arrays, the analysis itself is directly applicable to
gravitational wave detection using either ground or space-based detector data.
| [
{
"version": "v1",
"created": "Tue, 20 Apr 2010 16:40:35 GMT"
}
] | 2014-11-20T00:00:00 | [
[
"Finn",
"Lee Samuel",
""
],
[
"Lommen",
"Andrea N.",
""
]
] | TITLE: Detection, Localization and Characterization of Gravitational Wave
Bursts in a Pulsar Timing Array
ABSTRACT: Efforts to detect gravitational waves by timing an array of pulsars have
focused traditionally on stationary gravitational waves: e.g., stochastic or
periodic signals. Gravitational wave bursts --- signals whose duration is much
shorter than the observation period --- will also arise in the pulsar timing
array waveband. Sources that give rise to detectable bursts include the
formation or coalescence of supermassive black holes (SMBHs), the periapsis
passage of compact objects in highly elliptic or unbound orbits about a SMBH,
or cusps on cosmic strings. Here we describe how pulsar timing array data may
be analyzed to detect and characterize these bursts. Our analysis addresses, in
a mutually consistent manner, a hierarchy of three questions: \emph{i}) What
are the odds that a dataset includes the signal from a gravitational wave
burst? \emph{ii}) Assuming the presence of a burst, what is the direction to
its source? and \emph{iii}) Assuming the burst propagation direction, what is
the burst waveform's time dependence in each of its polarization states?
Applying our analysis to synthetic data sets we find that we can \emph{detect}
gravitational waves even when the radiation is too weak to either localize the
source of infer the waveform, and \emph{detect} and \emph{localize} sources
even when the radiation amplitude is too weak to permit the waveform to be
determined. While the context of our discussion is gravitational wave detection
via pulsar timing arrays, the analysis itself is directly applicable to
gravitational wave detection using either ground or space-based detector data.
| no_new_dataset | 0.94743 |
1406.6312 | Ahmed El-Kishky | Ahmed El-Kishky, Yanglei Song, Chi Wang, Clare Voss, Jiawei Han | Scalable Topical Phrase Mining from Text Corpora | null | Proceedings of the VLDB Endowment, Vol. 8(3), pp. 305 - 316, 2014 | null | null | cs.CL cs.IR cs.LG | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | While most topic modeling algorithms model text corpora with unigrams, human
interpretation often relies on inherent grouping of terms into phrases. As
such, we consider the problem of discovering topical phrases of mixed lengths.
Existing work either performs post processing to the inference results of
unigram-based topic models, or utilizes complex n-gram-discovery topic models.
These methods generally produce low-quality topical phrases or suffer from poor
scalability on even moderately-sized datasets. We propose a different approach
that is both computationally efficient and effective. Our solution combines a
novel phrase mining framework to segment a document into single and multi-word
phrases, and a new topic model that operates on the induced document partition.
Our approach discovers high quality topical phrases with negligible extra cost
to the bag-of-words topic model in a variety of datasets including research
publication titles, abstracts, reviews, and news articles.
| [
{
"version": "v1",
"created": "Tue, 24 Jun 2014 17:10:29 GMT"
},
{
"version": "v2",
"created": "Wed, 19 Nov 2014 00:18:06 GMT"
}
] | 2014-11-20T00:00:00 | [
[
"El-Kishky",
"Ahmed",
""
],
[
"Song",
"Yanglei",
""
],
[
"Wang",
"Chi",
""
],
[
"Voss",
"Clare",
""
],
[
"Han",
"Jiawei",
""
]
] | TITLE: Scalable Topical Phrase Mining from Text Corpora
ABSTRACT: While most topic modeling algorithms model text corpora with unigrams, human
interpretation often relies on inherent grouping of terms into phrases. As
such, we consider the problem of discovering topical phrases of mixed lengths.
Existing work either performs post processing to the inference results of
unigram-based topic models, or utilizes complex n-gram-discovery topic models.
These methods generally produce low-quality topical phrases or suffer from poor
scalability on even moderately-sized datasets. We propose a different approach
that is both computationally efficient and effective. Our solution combines a
novel phrase mining framework to segment a document into single and multi-word
phrases, and a new topic model that operates on the induced document partition.
Our approach discovers high quality topical phrases with negligible extra cost
to the bag-of-words topic model in a variety of datasets including research
publication titles, abstracts, reviews, and news articles.
| no_new_dataset | 0.950549 |
1407.7094 | Bruno Gon\c{c}alves | Bruno Gon\c{c}alves and David S\'anchez | Crowdsourcing Dialect Characterization through Twitter | 10 pages, 5 figures | PLoS One 9, E112074 (2014) | 10.1371/journal.pone.0112074 | null | physics.soc-ph cs.CL cs.SI stat.ML | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | We perform a large-scale analysis of language diatopic variation using
geotagged microblogging datasets. By collecting all Twitter messages written in
Spanish over more than two years, we build a corpus from which a carefully
selected list of concepts allows us to characterize Spanish varieties on a
global scale. A cluster analysis proves the existence of well defined
macroregions sharing common lexical properties. Remarkably enough, we find that
Spanish language is split into two superdialects, namely, an urban speech used
across major American and Spanish citites and a diverse form that encompasses
rural areas and small towns. The latter can be further clustered into smaller
varieties with a stronger regional character.
| [
{
"version": "v1",
"created": "Sat, 26 Jul 2014 04:16:31 GMT"
}
] | 2014-11-20T00:00:00 | [
[
"Gonçalves",
"Bruno",
""
],
[
"Sánchez",
"David",
""
]
] | TITLE: Crowdsourcing Dialect Characterization through Twitter
ABSTRACT: We perform a large-scale analysis of language diatopic variation using
geotagged microblogging datasets. By collecting all Twitter messages written in
Spanish over more than two years, we build a corpus from which a carefully
selected list of concepts allows us to characterize Spanish varieties on a
global scale. A cluster analysis proves the existence of well defined
macroregions sharing common lexical properties. Remarkably enough, we find that
Spanish language is split into two superdialects, namely, an urban speech used
across major American and Spanish citites and a diverse form that encompasses
rural areas and small towns. The latter can be further clustered into smaller
varieties with a stronger regional character.
| no_new_dataset | 0.799364 |
1410.0745 | Robinson Piramuthu Robinson Piramuthu | Qiaosong Wang, Vignesh Jagadeesh, Bryan Ressler, Robinson Piramuthu | Im2Fit: Fast 3D Model Fitting and Anthropometrics using Single Consumer
Depth Camera and Synthetic Data | null | null | null | null | cs.CV | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | Recent advances in consumer depth sensors have created many opportunities for
human body measurement and modeling. Estimation of 3D body shape is
particularly useful for fashion e-commerce applications such as virtual try-on
or fit personalization. In this paper, we propose a method for capturing
accurate human body shape and anthropometrics from a single consumer grade
depth sensor. We first generate a large dataset of synthetic 3D human body
models using real-world body size distributions. Next, we estimate key body
measurements from a single monocular depth image. We combine body measurement
estimates with local geometry features around key joint positions to form a
robust multi-dimensional feature vector. This allows us to conduct a fast
nearest-neighbor search to every sample in the dataset and return the closest
one. Compared to existing methods, our approach is able to predict accurate
full body parameters from a partial view using measurement parameters learned
from the synthetic dataset. Furthermore, our system is capable of generating 3D
human mesh models in real-time, which is significantly faster than methods
which attempt to model shape and pose deformations. To validate the efficiency
and applicability of our system, we collected a dataset that contains frontal
and back scans of 83 clothed people with ground truth height and weight.
Experiments on real-world dataset show that the proposed method can achieve
real-time performance with competing results achieving an average error of 1.9
cm in estimated measurements.
| [
{
"version": "v1",
"created": "Fri, 3 Oct 2014 02:33:08 GMT"
},
{
"version": "v2",
"created": "Wed, 19 Nov 2014 20:30:32 GMT"
}
] | 2014-11-20T00:00:00 | [
[
"Wang",
"Qiaosong",
""
],
[
"Jagadeesh",
"Vignesh",
""
],
[
"Ressler",
"Bryan",
""
],
[
"Piramuthu",
"Robinson",
""
]
] | TITLE: Im2Fit: Fast 3D Model Fitting and Anthropometrics using Single Consumer
Depth Camera and Synthetic Data
ABSTRACT: Recent advances in consumer depth sensors have created many opportunities for
human body measurement and modeling. Estimation of 3D body shape is
particularly useful for fashion e-commerce applications such as virtual try-on
or fit personalization. In this paper, we propose a method for capturing
accurate human body shape and anthropometrics from a single consumer grade
depth sensor. We first generate a large dataset of synthetic 3D human body
models using real-world body size distributions. Next, we estimate key body
measurements from a single monocular depth image. We combine body measurement
estimates with local geometry features around key joint positions to form a
robust multi-dimensional feature vector. This allows us to conduct a fast
nearest-neighbor search to every sample in the dataset and return the closest
one. Compared to existing methods, our approach is able to predict accurate
full body parameters from a partial view using measurement parameters learned
from the synthetic dataset. Furthermore, our system is capable of generating 3D
human mesh models in real-time, which is significantly faster than methods
which attempt to model shape and pose deformations. To validate the efficiency
and applicability of our system, we collected a dataset that contains frontal
and back scans of 83 clothed people with ground truth height and weight.
Experiments on real-world dataset show that the proposed method can achieve
real-time performance with competing results achieving an average error of 1.9
cm in estimated measurements.
| new_dataset | 0.958538 |
1411.5140 | Qian Wang | Qian Wang, Jiaxing Zhang, Sen Song, Zheng Zhang | Attentional Neural Network: Feature Selection Using Cognitive Feedback | Poster in Neural Information Processing Systems (NIPS) 2014 | null | null | null | cs.CV cs.NE | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | Attentional Neural Network is a new framework that integrates top-down
cognitive bias and bottom-up feature extraction in one coherent architecture.
The top-down influence is especially effective when dealing with high noise or
difficult segmentation problems. Our system is modular and extensible. It is
also easy to train and cheap to run, and yet can accommodate complex behaviors.
We obtain classification accuracy better than or competitive with state of art
results on the MNIST variation dataset, and successfully disentangle overlaid
digits with high success rates. We view such a general purpose framework as an
essential foundation for a larger system emulating the cognitive abilities of
the whole brain.
| [
{
"version": "v1",
"created": "Wed, 19 Nov 2014 08:33:28 GMT"
}
] | 2014-11-20T00:00:00 | [
[
"Wang",
"Qian",
""
],
[
"Zhang",
"Jiaxing",
""
],
[
"Song",
"Sen",
""
],
[
"Zhang",
"Zheng",
""
]
] | TITLE: Attentional Neural Network: Feature Selection Using Cognitive Feedback
ABSTRACT: Attentional Neural Network is a new framework that integrates top-down
cognitive bias and bottom-up feature extraction in one coherent architecture.
The top-down influence is especially effective when dealing with high noise or
difficult segmentation problems. Our system is modular and extensible. It is
also easy to train and cheap to run, and yet can accommodate complex behaviors.
We obtain classification accuracy better than or competitive with state of art
results on the MNIST variation dataset, and successfully disentangle overlaid
digits with high success rates. We view such a general purpose framework as an
essential foundation for a larger system emulating the cognitive abilities of
the whole brain.
| no_new_dataset | 0.9462 |
1411.5204 | Diego Saez-Trumper | Alessandro Venerandi, Giovanni Quattrone, Licia Capra, Daniele
Quercia, Diego Saez-Trumper | Measuring Urban Deprivation from User Generated Content | CSCW'15, March 14 - 18 2015, Vancouver, BC, Canada | null | 10.1145/2675133.2675233 | null | cs.SI | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | Measuring socioeconomic deprivation of cities in an accurate and timely
fashion has become a priority for governments around the world, as the massive
urbanization process we are witnessing is causing high levels of inequalities
which require intervention. Traditionally, deprivation indexes have been
derived from census data, which is however very expensive to obtain, and thus
acquired only every few years. Alternative computational methods have been
proposed in recent years to automatically extract proxies of deprivation at a
fine spatio-temporal level of granularity; however, they usually require access
to datasets (e.g., call details records) that are not publicly available to
governments and agencies.
To remedy this, we propose a new method to automatically mine deprivation at
a fine level of spatio-temporal granularity that only requires access to freely
available user-generated content. More precisely, the method needs access to
datasets describing what urban elements are present in the physical
environment; examples of such datasets are Foursquare and OpenStreetMap. Using
these datasets, we quantitatively describe neighborhoods by means of a metric,
called {\em Offering Advantage}, that reflects which urban elements are
distinctive features of each neighborhood. We then use that metric to {\em (i)}
build accurate classifiers of urban deprivation and {\em (ii)} interpret the
outcomes through thematic analysis. We apply the method to three UK urban areas
of different scale and elaborate on the results in terms of precision and
recall.
| [
{
"version": "v1",
"created": "Wed, 19 Nov 2014 12:44:12 GMT"
}
] | 2014-11-20T00:00:00 | [
[
"Venerandi",
"Alessandro",
""
],
[
"Quattrone",
"Giovanni",
""
],
[
"Capra",
"Licia",
""
],
[
"Quercia",
"Daniele",
""
],
[
"Saez-Trumper",
"Diego",
""
]
] | TITLE: Measuring Urban Deprivation from User Generated Content
ABSTRACT: Measuring socioeconomic deprivation of cities in an accurate and timely
fashion has become a priority for governments around the world, as the massive
urbanization process we are witnessing is causing high levels of inequalities
which require intervention. Traditionally, deprivation indexes have been
derived from census data, which is however very expensive to obtain, and thus
acquired only every few years. Alternative computational methods have been
proposed in recent years to automatically extract proxies of deprivation at a
fine spatio-temporal level of granularity; however, they usually require access
to datasets (e.g., call details records) that are not publicly available to
governments and agencies.
To remedy this, we propose a new method to automatically mine deprivation at
a fine level of spatio-temporal granularity that only requires access to freely
available user-generated content. More precisely, the method needs access to
datasets describing what urban elements are present in the physical
environment; examples of such datasets are Foursquare and OpenStreetMap. Using
these datasets, we quantitatively describe neighborhoods by means of a metric,
called {\em Offering Advantage}, that reflects which urban elements are
distinctive features of each neighborhood. We then use that metric to {\em (i)}
build accurate classifiers of urban deprivation and {\em (ii)} interpret the
outcomes through thematic analysis. We apply the method to three UK urban areas
of different scale and elaborate on the results in terms of precision and
recall.
| no_new_dataset | 0.950134 |
1411.5260 | Patrick Kimes | Patrick K. Kimes, D. Neil Hayes, J. S. Marron and Yufeng Liu | Large-Margin Classification with Multiple Decision Rules | null | null | null | null | stat.ML cs.LG | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | Binary classification is a common statistical learning problem in which a
model is estimated on a set of covariates for some outcome indicating the
membership of one of two classes. In the literature, there exists a distinction
between hard and soft classification. In soft classification, the conditional
class probability is modeled as a function of the covariates. In contrast, hard
classification methods only target the optimal prediction boundary. While hard
and soft classification methods have been studied extensively, not much work
has been done to compare the actual tasks of hard and soft classification. In
this paper we propose a spectrum of statistical learning problems which span
the hard and soft classification tasks based on fitting multiple decision rules
to the data. By doing so, we reveal a novel collection of learning tasks of
increasing complexity. We study the problems using the framework of
large-margin classifiers and a class of piecewise linear convex surrogates, for
which we derive statistical properties and a corresponding sub-gradient descent
algorithm. We conclude by applying our approach to simulation settings and a
magnetic resonance imaging (MRI) dataset from the Alzheimer's Disease
Neuroimaging Initiative (ADNI) study.
| [
{
"version": "v1",
"created": "Wed, 19 Nov 2014 15:45:54 GMT"
}
] | 2014-11-20T00:00:00 | [
[
"Kimes",
"Patrick K.",
""
],
[
"Hayes",
"D. Neil",
""
],
[
"Marron",
"J. S.",
""
],
[
"Liu",
"Yufeng",
""
]
] | TITLE: Large-Margin Classification with Multiple Decision Rules
ABSTRACT: Binary classification is a common statistical learning problem in which a
model is estimated on a set of covariates for some outcome indicating the
membership of one of two classes. In the literature, there exists a distinction
between hard and soft classification. In soft classification, the conditional
class probability is modeled as a function of the covariates. In contrast, hard
classification methods only target the optimal prediction boundary. While hard
and soft classification methods have been studied extensively, not much work
has been done to compare the actual tasks of hard and soft classification. In
this paper we propose a spectrum of statistical learning problems which span
the hard and soft classification tasks based on fitting multiple decision rules
to the data. By doing so, we reveal a novel collection of learning tasks of
increasing complexity. We study the problems using the framework of
large-margin classifiers and a class of piecewise linear convex surrogates, for
which we derive statistical properties and a corresponding sub-gradient descent
algorithm. We conclude by applying our approach to simulation settings and a
magnetic resonance imaging (MRI) dataset from the Alzheimer's Disease
Neuroimaging Initiative (ADNI) study.
| no_new_dataset | 0.942401 |
1411.5283 | Zaid Alyasseri | Zaid Abdi Alkareem Alyasseri, Kadhim Al-Attar, Mazin Nasser | Parallelize Bubble and Merge Sort Algorithms Using Message Passing
Interface (MPI) | 5 pages, 5 figures. arXiv admin note: substantial text overlap with
arXiv:1407.6603 | null | null | null | cs.DC | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | Sorting has been a profound area for the algorithmic researchers and many
resources are invested to suggest more works for sorting algorithms. For this
purpose, many existing sorting algorithms were observed in terms of the
efficiency of the algorithmic complexity. In this paper we implemented the
bubble and merge sort algorithms using Message Passing Interface (MPI)
approach. The proposed work tested on two standard datasets (text file) with
different size. The main idea of the proposed algorithm is distributing the
elements of the input datasets into many additional temporary sub-arrays
according to a number of characters in each word. The sizes of each of these
sub-arrays are decided depending on a number of elements with the same number
of characters in the input array. We implemented MPI using Intel core i7-3610QM
,(8 CPUs),using two approaches (vectors of string and array 3D) . Finally, we
get the data structure effects on the performance of the algorithm for that we
choice the second approach.
| [
{
"version": "v1",
"created": "Wed, 19 Nov 2014 16:35:16 GMT"
}
] | 2014-11-20T00:00:00 | [
[
"Alyasseri",
"Zaid Abdi Alkareem",
""
],
[
"Al-Attar",
"Kadhim",
""
],
[
"Nasser",
"Mazin",
""
]
] | TITLE: Parallelize Bubble and Merge Sort Algorithms Using Message Passing
Interface (MPI)
ABSTRACT: Sorting has been a profound area for the algorithmic researchers and many
resources are invested to suggest more works for sorting algorithms. For this
purpose, many existing sorting algorithms were observed in terms of the
efficiency of the algorithmic complexity. In this paper we implemented the
bubble and merge sort algorithms using Message Passing Interface (MPI)
approach. The proposed work tested on two standard datasets (text file) with
different size. The main idea of the proposed algorithm is distributing the
elements of the input datasets into many additional temporary sub-arrays
according to a number of characters in each word. The sizes of each of these
sub-arrays are decided depending on a number of elements with the same number
of characters in the input array. We implemented MPI using Intel core i7-3610QM
,(8 CPUs),using two approaches (vectors of string and array 3D) . Finally, we
get the data structure effects on the performance of the algorithm for that we
choice the second approach.
| no_new_dataset | 0.949106 |
1411.5307 | Robinson Piramuthu Robinson Piramuthu | Kevin Shih, Wei Di, Vignesh Jagadeesh, Robinson Piramuthu | Efficient Media Retrieval from Non-Cooperative Queries | 8 pages, 9 figures, 1 table | null | null | null | cs.IR cs.CV | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | Text is ubiquitous in the artificial world and easily attainable when it
comes to book title and author names. Using the images from the book cover set
from the Stanford Mobile Visual Search dataset and additional book covers and
metadata from openlibrary.org, we construct a large scale book cover retrieval
dataset, complete with 100K distractor covers and title and author strings for
each. Because our query images are poorly conditioned for clean text
extraction, we propose a method for extracting a matching noisy and erroneous
OCR readings and matching it against clean author and book title strings in a
standard document look-up problem setup. Finally, we demonstrate how to use
this text-matching as a feature in conjunction with popular retrieval features
such as VLAD using a simple learning setup to achieve significant improvements
in retrieval accuracy over that of either VLAD or the text alone.
| [
{
"version": "v1",
"created": "Wed, 19 Nov 2014 18:34:28 GMT"
}
] | 2014-11-20T00:00:00 | [
[
"Shih",
"Kevin",
""
],
[
"Di",
"Wei",
""
],
[
"Jagadeesh",
"Vignesh",
""
],
[
"Piramuthu",
"Robinson",
""
]
] | TITLE: Efficient Media Retrieval from Non-Cooperative Queries
ABSTRACT: Text is ubiquitous in the artificial world and easily attainable when it
comes to book title and author names. Using the images from the book cover set
from the Stanford Mobile Visual Search dataset and additional book covers and
metadata from openlibrary.org, we construct a large scale book cover retrieval
dataset, complete with 100K distractor covers and title and author strings for
each. Because our query images are poorly conditioned for clean text
extraction, we propose a method for extracting a matching noisy and erroneous
OCR readings and matching it against clean author and book title strings in a
standard document look-up problem setup. Finally, we demonstrate how to use
this text-matching as a feature in conjunction with popular retrieval features
such as VLAD using a simple learning setup to achieve significant improvements
in retrieval accuracy over that of either VLAD or the text alone.
| new_dataset | 0.959078 |
1411.5309 | David Eigen | Li Wan and David Eigen and Rob Fergus | End-to-End Integration of a Convolutional Network, Deformable Parts
Model and Non-Maximum Suppression | null | null | null | null | cs.CV | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | Deformable Parts Models and Convolutional Networks each have achieved notable
performance in object detection. Yet these two approaches find their strengths
in complementary areas: DPMs are well-versed in object composition, modeling
fine-grained spatial relationships between parts; likewise, ConvNets are adept
at producing powerful image features, having been discriminatively trained
directly on the pixels. In this paper, we propose a new model that combines
these two approaches, obtaining the advantages of each. We train this model
using a new structured loss function that considers all bounding boxes within
an image, rather than isolated object instances. This enables the non-maximal
suppression (NMS) operation, previously treated as a separate post-processing
stage, to be integrated into the model. This allows for discriminative training
of our combined Convnet + DPM + NMS model in end-to-end fashion. We evaluate
our system on PASCAL VOC 2007 and 2011 datasets, achieving competitive results
on both benchmarks.
| [
{
"version": "v1",
"created": "Wed, 19 Nov 2014 18:36:09 GMT"
}
] | 2014-11-20T00:00:00 | [
[
"Wan",
"Li",
""
],
[
"Eigen",
"David",
""
],
[
"Fergus",
"Rob",
""
]
] | TITLE: End-to-End Integration of a Convolutional Network, Deformable Parts
Model and Non-Maximum Suppression
ABSTRACT: Deformable Parts Models and Convolutional Networks each have achieved notable
performance in object detection. Yet these two approaches find their strengths
in complementary areas: DPMs are well-versed in object composition, modeling
fine-grained spatial relationships between parts; likewise, ConvNets are adept
at producing powerful image features, having been discriminatively trained
directly on the pixels. In this paper, we propose a new model that combines
these two approaches, obtaining the advantages of each. We train this model
using a new structured loss function that considers all bounding boxes within
an image, rather than isolated object instances. This enables the non-maximal
suppression (NMS) operation, previously treated as a separate post-processing
stage, to be integrated into the model. This allows for discriminative training
of our combined Convnet + DPM + NMS model in end-to-end fashion. We evaluate
our system on PASCAL VOC 2007 and 2011 datasets, achieving competitive results
on both benchmarks.
| no_new_dataset | 0.949059 |
1402.5138 | Dieter Pfoser | Mahmuda Ahmed and Sophia Karagiorgou and Dieter Pfoser and Carola Wenk | A Comparison and Evaluation of Map Construction Algorithms | null | null | 10.1007/s10707-014-0222-6 | null | cs.CG | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | Map construction methods automatically produce and/or update road network
datasets using vehicle tracking data. Enabled by the ubiquitous generation of
georeferenced tracking data, there has been a recent surge in map construction
algorithms coming from different computer science domains. A cross-comparison
of the various algorithms is still very rare, since (i) algorithms and
constructed maps are generally not publicly available and (ii) there is no
standard approach to assess the result quality, given the lack of benchmark
data and quantitative evaluation methods. This work represents a first
comprehensive attempt to benchmark map construction algorithms. We provide an
evaluation and comparison of seven algorithms using four datasets and four
different evaluation measures. In addition to this comprehensive comparison, we
make our datasets, source code of map construction algorithms and evaluation
measures publicly available on mapconstruction.org. This site has been
established as a repository for map con- struction data and algorithms and we
invite other researchers to contribute by uploading code and benchmark data
supporting their contributions to map construction algorithms.
| [
{
"version": "v1",
"created": "Wed, 19 Feb 2014 21:50:39 GMT"
},
{
"version": "v2",
"created": "Thu, 12 Jun 2014 14:51:46 GMT"
}
] | 2014-11-19T00:00:00 | [
[
"Ahmed",
"Mahmuda",
""
],
[
"Karagiorgou",
"Sophia",
""
],
[
"Pfoser",
"Dieter",
""
],
[
"Wenk",
"Carola",
""
]
] | TITLE: A Comparison and Evaluation of Map Construction Algorithms
ABSTRACT: Map construction methods automatically produce and/or update road network
datasets using vehicle tracking data. Enabled by the ubiquitous generation of
georeferenced tracking data, there has been a recent surge in map construction
algorithms coming from different computer science domains. A cross-comparison
of the various algorithms is still very rare, since (i) algorithms and
constructed maps are generally not publicly available and (ii) there is no
standard approach to assess the result quality, given the lack of benchmark
data and quantitative evaluation methods. This work represents a first
comprehensive attempt to benchmark map construction algorithms. We provide an
evaluation and comparison of seven algorithms using four datasets and four
different evaluation measures. In addition to this comprehensive comparison, we
make our datasets, source code of map construction algorithms and evaluation
measures publicly available on mapconstruction.org. This site has been
established as a repository for map con- struction data and algorithms and we
invite other researchers to contribute by uploading code and benchmark data
supporting their contributions to map construction algorithms.
| no_new_dataset | 0.784195 |
1409.6911 | Zhiqiang Shen | Zhiqiang Shen and Xiangyang Xue | Do More Dropouts in Pool5 Feature Maps for Better Object Detection | 9 pages, 7 figures | null | null | null | cs.CV | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | Deep Convolutional Neural Networks (CNNs) have gained great success in image
classification and object detection. In these fields, the outputs of all layers
of CNNs are usually considered as a high dimensional feature vector extracted
from an input image and the correspondence between finer level feature vectors
and concepts that the input image contains is all-important. However, fewer
studies focus on this deserving issue. On considering the correspondence, we
propose a novel approach which generates an edited version for each original
CNN feature vector by applying the maximum entropy principle to abandon
particular vectors. These selected vectors correspond to the unfriendly
concepts in each image category. The classifier trained from merged feature
sets can significantly improve model generalization of individual categories
when training data is limited. The experimental results for
classification-based object detection on canonical datasets including VOC 2007
(60.1%), 2010 (56.4%) and 2012 (56.3%) show obvious improvement in mean average
precision (mAP) with simple linear support vector machines.
| [
{
"version": "v1",
"created": "Wed, 24 Sep 2014 11:50:48 GMT"
},
{
"version": "v2",
"created": "Mon, 17 Nov 2014 12:27:23 GMT"
},
{
"version": "v3",
"created": "Tue, 18 Nov 2014 17:34:22 GMT"
}
] | 2014-11-19T00:00:00 | [
[
"Shen",
"Zhiqiang",
""
],
[
"Xue",
"Xiangyang",
""
]
] | TITLE: Do More Dropouts in Pool5 Feature Maps for Better Object Detection
ABSTRACT: Deep Convolutional Neural Networks (CNNs) have gained great success in image
classification and object detection. In these fields, the outputs of all layers
of CNNs are usually considered as a high dimensional feature vector extracted
from an input image and the correspondence between finer level feature vectors
and concepts that the input image contains is all-important. However, fewer
studies focus on this deserving issue. On considering the correspondence, we
propose a novel approach which generates an edited version for each original
CNN feature vector by applying the maximum entropy principle to abandon
particular vectors. These selected vectors correspond to the unfriendly
concepts in each image category. The classifier trained from merged feature
sets can significantly improve model generalization of individual categories
when training data is limited. The experimental results for
classification-based object detection on canonical datasets including VOC 2007
(60.1%), 2010 (56.4%) and 2012 (56.3%) show obvious improvement in mean average
precision (mAP) with simple linear support vector machines.
| no_new_dataset | 0.950549 |
1411.4670 | Mohamed Hussein | Mohamed E. Hussein and Marwan Torki and Ahmed Elsallamy and Mahmoud
Fayyaz | AlexU-Word: A New Dataset for Isolated-Word Closed-Vocabulary Offline
Arabic Handwriting Recognition | 6 pages, 8 figure, and 6 tables | null | null | null | cs.CV | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | In this paper, we introduce the first phase of a new dataset for offline
Arabic handwriting recognition. The aim is to collect a very large dataset of
isolated Arabic words that covers all letters of the alphabet in all possible
shapes using a small number of simple words. The end goal is to collect a very
large dataset of segmented letter images, which can be used to build and
evaluate Arabic handwriting recognition systems that are based on segmented
letter recognition. The current version of the dataset contains $25114$ samples
of $109$ unique Arabic words that cover all possible shapes of all alphabet
letters. The samples were collected from $907$ writers. In its current form,
the dataset can be used for the problem of closed-vocabulary word recognition.
We evaluated a number of window-based descriptors and classifiers on this task
and obtained an accuracy of $92.16\%$ using a SIFT-based descriptor and ANN.
| [
{
"version": "v1",
"created": "Mon, 17 Nov 2014 21:23:26 GMT"
}
] | 2014-11-19T00:00:00 | [
[
"Hussein",
"Mohamed E.",
""
],
[
"Torki",
"Marwan",
""
],
[
"Elsallamy",
"Ahmed",
""
],
[
"Fayyaz",
"Mahmoud",
""
]
] | TITLE: AlexU-Word: A New Dataset for Isolated-Word Closed-Vocabulary Offline
Arabic Handwriting Recognition
ABSTRACT: In this paper, we introduce the first phase of a new dataset for offline
Arabic handwriting recognition. The aim is to collect a very large dataset of
isolated Arabic words that covers all letters of the alphabet in all possible
shapes using a small number of simple words. The end goal is to collect a very
large dataset of segmented letter images, which can be used to build and
evaluate Arabic handwriting recognition systems that are based on segmented
letter recognition. The current version of the dataset contains $25114$ samples
of $109$ unique Arabic words that cover all possible shapes of all alphabet
letters. The samples were collected from $907$ writers. In its current form,
the dataset can be used for the problem of closed-vocabulary word recognition.
We evaluated a number of window-based descriptors and classifiers on this task
and obtained an accuracy of $92.16\%$ using a SIFT-based descriptor and ANN.
| new_dataset | 0.956227 |
1411.4958 | Xiaolong Wang | Xiaolong Wang, David F. Fouhey, Abhinav Gupta | Designing Deep Networks for Surface Normal Estimation | null | null | null | null | cs.CV | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | In the past few years, convolutional neural nets (CNN) have shown incredible
promise for learning visual representations. In this paper, we use CNNs for the
task of predicting surface normals from a single image. But what is the right
architecture we should use? We propose to build upon the decades of hard work
in 3D scene understanding, to design new CNN architecture for the task of
surface normal estimation. We show by incorporating several constraints
(man-made, manhattan world) and meaningful intermediate representations (room
layout, edge labels) in the architecture leads to state of the art performance
on surface normal estimation. We also show that our network is quite robust and
show state of the art results on other datasets as well without any
fine-tuning.
| [
{
"version": "v1",
"created": "Tue, 18 Nov 2014 18:39:48 GMT"
}
] | 2014-11-19T00:00:00 | [
[
"Wang",
"Xiaolong",
""
],
[
"Fouhey",
"David F.",
""
],
[
"Gupta",
"Abhinav",
""
]
] | TITLE: Designing Deep Networks for Surface Normal Estimation
ABSTRACT: In the past few years, convolutional neural nets (CNN) have shown incredible
promise for learning visual representations. In this paper, we use CNNs for the
task of predicting surface normals from a single image. But what is the right
architecture we should use? We propose to build upon the decades of hard work
in 3D scene understanding, to design new CNN architecture for the task of
surface normal estimation. We show by incorporating several constraints
(man-made, manhattan world) and meaningful intermediate representations (room
layout, edge labels) in the architecture leads to state of the art performance
on surface normal estimation. We also show that our network is quite robust and
show state of the art results on other datasets as well without any
fine-tuning.
| no_new_dataset | 0.955026 |
1411.4960 | Ana Me\v{s}trovi\'c | Hana Rizvi\'c, Sanda Martin\v{c}i\'c-Ip\v{s}i\'c, Ana Me\v{s}trovi\'c | Network Motifs Analysis of Croatian Literature | null | null | null | null | cs.CL | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | In this paper we analyse network motifs in the co-occurrence directed
networks constructed from five different texts (four books and one portal) in
the Croatian language. After preparing the data and network construction, we
perform the network motif analysis. We analyse the motif frequencies and
Z-scores in the five networks. We present the triad significance profile for
five datasets. Furthermore, we compare our results with the existing results
for the linguistic networks. Firstly, we show that the triad significance
profile for the Croatian language is very similar with the other languages and
all the networks belong to the same family of networks. However, there are
certain differences between the Croatian language and other analysed languages.
We conclude that this is due to the free word-order of the Croatian language.
| [
{
"version": "v1",
"created": "Tue, 18 Nov 2014 18:46:36 GMT"
}
] | 2014-11-19T00:00:00 | [
[
"Rizvić",
"Hana",
""
],
[
"Martinčić-Ipšić",
"Sanda",
""
],
[
"Meštrović",
"Ana",
""
]
] | TITLE: Network Motifs Analysis of Croatian Literature
ABSTRACT: In this paper we analyse network motifs in the co-occurrence directed
networks constructed from five different texts (four books and one portal) in
the Croatian language. After preparing the data and network construction, we
perform the network motif analysis. We analyse the motif frequencies and
Z-scores in the five networks. We present the triad significance profile for
five datasets. Furthermore, we compare our results with the existing results
for the linguistic networks. Firstly, we show that the triad significance
profile for the Croatian language is very similar with the other languages and
all the networks belong to the same family of networks. However, there are
certain differences between the Croatian language and other analysed languages.
We conclude that this is due to the free word-order of the Croatian language.
| no_new_dataset | 0.946745 |
1403.3155 | Fei-Yun Zhu | Feiyun Zhu, Ying Wang, Bin Fan, Gaofeng Meng, Shiming Xiang and
Chunhong Pan | Spectral Unmixing via Data-guided Sparsity | null | null | 10.1109/TIP.2014.2363423 | null | cs.CV | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | Hyperspectral unmixing, the process of estimating a common set of spectral
bases and their corresponding composite percentages at each pixel, is an
important task for hyperspectral analysis, visualization and understanding.
From an unsupervised learning perspective, this problem is very
challenging---both the spectral bases and their composite percentages are
unknown, making the solution space too large. To reduce the solution space,
many approaches have been proposed by exploiting various priors. In practice,
these priors would easily lead to some unsuitable solution. This is because
they are achieved by applying an identical strength of constraints to all the
factors, which does not hold in practice. To overcome this limitation, we
propose a novel sparsity based method by learning a data-guided map to describe
the individual mixed level of each pixel. Through this data-guided map, the
$\ell_{p}(0<p<1)$ constraint is applied in an adaptive manner. Such
implementation not only meets the practical situation, but also guides the
spectral bases toward the pixels under highly sparse constraint. What's more,
an elegant optimization scheme as well as its convergence proof have been
provided in this paper. Extensive experiments on several datasets also
demonstrate that the data-guided map is feasible, and high quality unmixing
results could be obtained by our method.
| [
{
"version": "v1",
"created": "Thu, 13 Mar 2014 03:29:22 GMT"
},
{
"version": "v2",
"created": "Mon, 14 Jul 2014 13:49:14 GMT"
},
{
"version": "v3",
"created": "Fri, 19 Sep 2014 02:59:51 GMT"
},
{
"version": "v4",
"created": "Mon, 17 Nov 2014 15:18:15 GMT"
}
] | 2014-11-18T00:00:00 | [
[
"Zhu",
"Feiyun",
""
],
[
"Wang",
"Ying",
""
],
[
"Fan",
"Bin",
""
],
[
"Meng",
"Gaofeng",
""
],
[
"Xiang",
"Shiming",
""
],
[
"Pan",
"Chunhong",
""
]
] | TITLE: Spectral Unmixing via Data-guided Sparsity
ABSTRACT: Hyperspectral unmixing, the process of estimating a common set of spectral
bases and their corresponding composite percentages at each pixel, is an
important task for hyperspectral analysis, visualization and understanding.
From an unsupervised learning perspective, this problem is very
challenging---both the spectral bases and their composite percentages are
unknown, making the solution space too large. To reduce the solution space,
many approaches have been proposed by exploiting various priors. In practice,
these priors would easily lead to some unsuitable solution. This is because
they are achieved by applying an identical strength of constraints to all the
factors, which does not hold in practice. To overcome this limitation, we
propose a novel sparsity based method by learning a data-guided map to describe
the individual mixed level of each pixel. Through this data-guided map, the
$\ell_{p}(0<p<1)$ constraint is applied in an adaptive manner. Such
implementation not only meets the practical situation, but also guides the
spectral bases toward the pixels under highly sparse constraint. What's more,
an elegant optimization scheme as well as its convergence proof have been
provided in this paper. Extensive experiments on several datasets also
demonstrate that the data-guided map is feasible, and high quality unmixing
results could be obtained by our method.
| no_new_dataset | 0.948298 |
1407.4764 | Ken Chatfield | Ken Chatfield, Karen Simonyan and Andrew Zisserman | Efficient On-the-fly Category Retrieval using ConvNets and GPUs | Published in proceedings of ACCV 2014 | null | null | null | cs.CV cs.LG cs.NE | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | We investigate the gains in precision and speed, that can be obtained by
using Convolutional Networks (ConvNets) for on-the-fly retrieval - where
classifiers are learnt at run time for a textual query from downloaded images,
and used to rank large image or video datasets.
We make three contributions: (i) we present an evaluation of state-of-the-art
image representations for object category retrieval over standard benchmark
datasets containing 1M+ images; (ii) we show that ConvNets can be used to
obtain features which are incredibly performant, and yet much lower dimensional
than previous state-of-the-art image representations, and that their
dimensionality can be reduced further without loss in performance by
compression using product quantization or binarization. Consequently, features
with the state-of-the-art performance on large-scale datasets of millions of
images can fit in the memory of even a commodity GPU card; (iii) we show that
an SVM classifier can be learnt within a ConvNet framework on a GPU in parallel
with downloading the new training images, allowing for a continuous refinement
of the model as more images become available, and simultaneous training and
ranking. The outcome is an on-the-fly system that significantly outperforms its
predecessors in terms of: precision of retrieval, memory requirements, and
speed, facilitating accurate on-the-fly learning and ranking in under a second
on a single GPU.
| [
{
"version": "v1",
"created": "Thu, 17 Jul 2014 18:29:38 GMT"
},
{
"version": "v2",
"created": "Wed, 5 Nov 2014 08:27:23 GMT"
},
{
"version": "v3",
"created": "Mon, 17 Nov 2014 12:10:23 GMT"
}
] | 2014-11-18T00:00:00 | [
[
"Chatfield",
"Ken",
""
],
[
"Simonyan",
"Karen",
""
],
[
"Zisserman",
"Andrew",
""
]
] | TITLE: Efficient On-the-fly Category Retrieval using ConvNets and GPUs
ABSTRACT: We investigate the gains in precision and speed, that can be obtained by
using Convolutional Networks (ConvNets) for on-the-fly retrieval - where
classifiers are learnt at run time for a textual query from downloaded images,
and used to rank large image or video datasets.
We make three contributions: (i) we present an evaluation of state-of-the-art
image representations for object category retrieval over standard benchmark
datasets containing 1M+ images; (ii) we show that ConvNets can be used to
obtain features which are incredibly performant, and yet much lower dimensional
than previous state-of-the-art image representations, and that their
dimensionality can be reduced further without loss in performance by
compression using product quantization or binarization. Consequently, features
with the state-of-the-art performance on large-scale datasets of millions of
images can fit in the memory of even a commodity GPU card; (iii) we show that
an SVM classifier can be learnt within a ConvNet framework on a GPU in parallel
with downloading the new training images, allowing for a continuous refinement
of the model as more images become available, and simultaneous training and
ranking. The outcome is an on-the-fly system that significantly outperforms its
predecessors in terms of: precision of retrieval, memory requirements, and
speed, facilitating accurate on-the-fly learning and ranking in under a second
on a single GPU.
| no_new_dataset | 0.948394 |
1409.3660 | Fei-Yun Zhu | Feiyun Zhu, Bin Fan, Xinliang Zhu, Ying Wang, Shiming Xiang and
Chunhong Pan | 10,000+ Times Accelerated Robust Subset Selection (ARSS) | null | null | null | null | cs.LG cs.CV stat.ML | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | Subset selection from massive data with noised information is increasingly
popular for various applications. This problem is still highly challenging as
current methods are generally slow in speed and sensitive to outliers. To
address the above two issues, we propose an accelerated robust subset selection
(ARSS) method. Specifically in the subset selection area, this is the first
attempt to employ the $\ell_{p}(0<p\leq1)$-norm based measure for the
representation loss, preventing large errors from dominating our objective. As
a result, the robustness against outlier elements is greatly enhanced.
Actually, data size is generally much larger than feature length, i.e. $N\gg
L$. Based on this observation, we propose a speedup solver (via ALM and
equivalent derivations) to highly reduce the computational cost, theoretically
from $O(N^{4})$ to $O(N{}^{2}L)$. Extensive experiments on ten benchmark
datasets verify that our method not only outperforms state of the art methods,
but also runs 10,000+ times faster than the most related method.
| [
{
"version": "v1",
"created": "Fri, 12 Sep 2014 07:18:17 GMT"
},
{
"version": "v2",
"created": "Fri, 19 Sep 2014 02:49:19 GMT"
},
{
"version": "v3",
"created": "Mon, 13 Oct 2014 07:58:57 GMT"
},
{
"version": "v4",
"created": "Mon, 17 Nov 2014 14:39:31 GMT"
}
] | 2014-11-18T00:00:00 | [
[
"Zhu",
"Feiyun",
""
],
[
"Fan",
"Bin",
""
],
[
"Zhu",
"Xinliang",
""
],
[
"Wang",
"Ying",
""
],
[
"Xiang",
"Shiming",
""
],
[
"Pan",
"Chunhong",
""
]
] | TITLE: 10,000+ Times Accelerated Robust Subset Selection (ARSS)
ABSTRACT: Subset selection from massive data with noised information is increasingly
popular for various applications. This problem is still highly challenging as
current methods are generally slow in speed and sensitive to outliers. To
address the above two issues, we propose an accelerated robust subset selection
(ARSS) method. Specifically in the subset selection area, this is the first
attempt to employ the $\ell_{p}(0<p\leq1)$-norm based measure for the
representation loss, preventing large errors from dominating our objective. As
a result, the robustness against outlier elements is greatly enhanced.
Actually, data size is generally much larger than feature length, i.e. $N\gg
L$. Based on this observation, we propose a speedup solver (via ALM and
equivalent derivations) to highly reduce the computational cost, theoretically
from $O(N^{4})$ to $O(N{}^{2}L)$. Extensive experiments on ten benchmark
datasets verify that our method not only outperforms state of the art methods,
but also runs 10,000+ times faster than the most related method.
| no_new_dataset | 0.944125 |
1411.3519 | Mohamed Hussein | Marwan Torki, Mohamed E. Hussein, Ahmed Elsallamy, Mahmoud Fayyaz,
Shehab Yaser | Window-Based Descriptors for Arabic Handwritten Alphabet Recognition: A
Comparative Study on a Novel Dataset | null | null | null | null | cs.CV | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | This paper presents a comparative study for window-based descriptors on the
application of Arabic handwritten alphabet recognition. We show a detailed
experimental evaluation of different descriptors with several classifiers. The
objective of the paper is to evaluate different window-based descriptors on the
problem of Arabic letter recognition. Our experiments clearly show that they
perform very well. Moreover, we introduce a novel spatial pyramid partitioning
scheme that enhances the recognition accuracy for most descriptors. In
addition, we introduce a novel dataset for Arabic handwritten isolated alphabet
letters, which can serve as a benchmark for future research.
| [
{
"version": "v1",
"created": "Thu, 13 Nov 2014 12:22:57 GMT"
},
{
"version": "v2",
"created": "Mon, 17 Nov 2014 17:55:32 GMT"
}
] | 2014-11-18T00:00:00 | [
[
"Torki",
"Marwan",
""
],
[
"Hussein",
"Mohamed E.",
""
],
[
"Elsallamy",
"Ahmed",
""
],
[
"Fayyaz",
"Mahmoud",
""
],
[
"Yaser",
"Shehab",
""
]
] | TITLE: Window-Based Descriptors for Arabic Handwritten Alphabet Recognition: A
Comparative Study on a Novel Dataset
ABSTRACT: This paper presents a comparative study for window-based descriptors on the
application of Arabic handwritten alphabet recognition. We show a detailed
experimental evaluation of different descriptors with several classifiers. The
objective of the paper is to evaluate different window-based descriptors on the
problem of Arabic letter recognition. Our experiments clearly show that they
perform very well. Moreover, we introduce a novel spatial pyramid partitioning
scheme that enhances the recognition accuracy for most descriptors. In
addition, we introduce a novel dataset for Arabic handwritten isolated alphabet
letters, which can serve as a benchmark for future research.
| new_dataset | 0.961929 |
1411.4080 | Miriam Redi | Miriam Redi, Neil O Hare, Rossano Schifanella, Michele Trevisiol,
Alejandro Jaimes | 6 Seconds of Sound and Vision: Creativity in Micro-Videos | 8 pages, 1 figures, conference IEEE CVPR 2014 | null | 10.1109/CVPR.2014.544 | null | cs.MM cs.CV cs.HC | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | The notion of creativity, as opposed to related concepts such as beauty or
interestingness, has not been studied from the perspective of automatic
analysis of multimedia content. Meanwhile, short online videos shared on social
media platforms, or micro-videos, have arisen as a new medium for creative
expression. In this paper we study creative micro-videos in an effort to
understand the features that make a video creative, and to address the problem
of automatic detection of creative content. Defining creative videos as those
that are novel and have aesthetic value, we conduct a crowdsourcing experiment
to create a dataset of over 3,800 micro-videos labelled as creative and
non-creative. We propose a set of computational features that we map to the
components of our definition of creativity, and conduct an analysis to
determine which of these features correlate most with creative video. Finally,
we evaluate a supervised approach to automatically detect creative video, with
promising results, showing that it is necessary to model both aesthetic value
and novelty to achieve optimal classification accuracy.
| [
{
"version": "v1",
"created": "Fri, 14 Nov 2014 23:29:18 GMT"
}
] | 2014-11-18T00:00:00 | [
[
"Redi",
"Miriam",
""
],
[
"Hare",
"Neil O",
""
],
[
"Schifanella",
"Rossano",
""
],
[
"Trevisiol",
"Michele",
""
],
[
"Jaimes",
"Alejandro",
""
]
] | TITLE: 6 Seconds of Sound and Vision: Creativity in Micro-Videos
ABSTRACT: The notion of creativity, as opposed to related concepts such as beauty or
interestingness, has not been studied from the perspective of automatic
analysis of multimedia content. Meanwhile, short online videos shared on social
media platforms, or micro-videos, have arisen as a new medium for creative
expression. In this paper we study creative micro-videos in an effort to
understand the features that make a video creative, and to address the problem
of automatic detection of creative content. Defining creative videos as those
that are novel and have aesthetic value, we conduct a crowdsourcing experiment
to create a dataset of over 3,800 micro-videos labelled as creative and
non-creative. We propose a set of computational features that we map to the
components of our definition of creativity, and conduct an analysis to
determine which of these features correlate most with creative video. Finally,
we evaluate a supervised approach to automatically detect creative video, with
promising results, showing that it is necessary to model both aesthetic value
and novelty to achieve optimal classification accuracy.
| new_dataset | 0.959988 |
1411.4086 | Hongwei Li | Hongwei Li and Bin Yu | Error Rate Bounds and Iterative Weighted Majority Voting for
Crowdsourcing | Journal Submission | null | null | null | stat.ML cs.HC cs.LG math.PR math.ST stat.TH | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | Crowdsourcing has become an effective and popular tool for human-powered
computation to label large datasets. Since the workers can be unreliable, it is
common in crowdsourcing to assign multiple workers to one task, and to
aggregate the labels in order to obtain results of high quality. In this paper,
we provide finite-sample exponential bounds on the error rate (in probability
and in expectation) of general aggregation rules under the Dawid-Skene
crowdsourcing model. The bounds are derived for multi-class labeling, and can
be used to analyze many aggregation methods, including majority voting,
weighted majority voting and the oracle Maximum A Posteriori (MAP) rule. We
show that the oracle MAP rule approximately optimizes our upper bound on the
mean error rate of weighted majority voting in certain setting. We propose an
iterative weighted majority voting (IWMV) method that optimizes the error rate
bound and approximates the oracle MAP rule. Its one step version has a provable
theoretical guarantee on the error rate. The IWMV method is intuitive and
computationally simple. Experimental results on simulated and real data show
that IWMV performs at least on par with the state-of-the-art methods, and it
has a much lower computational cost (around one hundred times faster) than the
state-of-the-art methods.
| [
{
"version": "v1",
"created": "Sat, 15 Nov 2014 00:02:34 GMT"
}
] | 2014-11-18T00:00:00 | [
[
"Li",
"Hongwei",
""
],
[
"Yu",
"Bin",
""
]
] | TITLE: Error Rate Bounds and Iterative Weighted Majority Voting for
Crowdsourcing
ABSTRACT: Crowdsourcing has become an effective and popular tool for human-powered
computation to label large datasets. Since the workers can be unreliable, it is
common in crowdsourcing to assign multiple workers to one task, and to
aggregate the labels in order to obtain results of high quality. In this paper,
we provide finite-sample exponential bounds on the error rate (in probability
and in expectation) of general aggregation rules under the Dawid-Skene
crowdsourcing model. The bounds are derived for multi-class labeling, and can
be used to analyze many aggregation methods, including majority voting,
weighted majority voting and the oracle Maximum A Posteriori (MAP) rule. We
show that the oracle MAP rule approximately optimizes our upper bound on the
mean error rate of weighted majority voting in certain setting. We propose an
iterative weighted majority voting (IWMV) method that optimizes the error rate
bound and approximates the oracle MAP rule. Its one step version has a provable
theoretical guarantee on the error rate. The IWMV method is intuitive and
computationally simple. Experimental results on simulated and real data show
that IWMV performs at least on par with the state-of-the-art methods, and it
has a much lower computational cost (around one hundred times faster) than the
state-of-the-art methods.
| no_new_dataset | 0.946349 |
1411.4101 | Rahul Mohan Mr. | Rahul Mohan | Deep Deconvolutional Networks for Scene Parsing | null | null | null | null | stat.ML cs.CV cs.LG | http://creativecommons.org/licenses/by/3.0/ | Scene parsing is an important and challenging prob- lem in computer vision.
It requires labeling each pixel in an image with the category it belongs to.
Tradition- ally, it has been approached with hand-engineered features from
color information in images. Recently convolutional neural networks (CNNs),
which automatically learn hierar- chies of features, have achieved record
performance on the task. These approaches typically include a post-processing
technique, such as superpixels, to produce the final label- ing. In this paper,
we propose a novel network architecture that combines deep deconvolutional
neural networks with CNNs. Our experiments show that deconvolutional neu- ral
networks are capable of learning higher order image structure beyond edge
primitives in comparison to CNNs. The new network architecture is employed for
multi-patch training, introduced as part of this work. Multi-patch train- ing
makes it possible to effectively learn spatial priors from scenes. The proposed
approach yields state-of-the-art per- formance on four scene parsing datasets,
namely Stanford Background, SIFT Flow, CamVid, and KITTI. In addition, our
system has the added advantage of having a training system that can be
completely automated end-to-end with- out requiring any post-processing.
| [
{
"version": "v1",
"created": "Sat, 15 Nov 2014 02:03:14 GMT"
}
] | 2014-11-18T00:00:00 | [
[
"Mohan",
"Rahul",
""
]
] | TITLE: Deep Deconvolutional Networks for Scene Parsing
ABSTRACT: Scene parsing is an important and challenging prob- lem in computer vision.
It requires labeling each pixel in an image with the category it belongs to.
Tradition- ally, it has been approached with hand-engineered features from
color information in images. Recently convolutional neural networks (CNNs),
which automatically learn hierar- chies of features, have achieved record
performance on the task. These approaches typically include a post-processing
technique, such as superpixels, to produce the final label- ing. In this paper,
we propose a novel network architecture that combines deep deconvolutional
neural networks with CNNs. Our experiments show that deconvolutional neu- ral
networks are capable of learning higher order image structure beyond edge
primitives in comparison to CNNs. The new network architecture is employed for
multi-patch training, introduced as part of this work. Multi-patch train- ing
makes it possible to effectively learn spatial priors from scenes. The proposed
approach yields state-of-the-art per- formance on four scene parsing datasets,
namely Stanford Background, SIFT Flow, CamVid, and KITTI. In addition, our
system has the added advantage of having a training system that can be
completely automated end-to-end with- out requiring any post-processing.
| no_new_dataset | 0.952486 |
1411.4246 | Md Lisul Islam | Md. Lisul Islam, Swakkhar Shatabda and M. Sohel Rahman | GreMuTRRR: A Novel Genetic Algorithm to Solve Distance Geometry Problem
for Protein Structures | Accepted for publication in the 8th International Conference on
Electrical and Computer Engineering (ICECE 2014) | null | null | null | cs.NE cs.CE | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | Nuclear Magnetic Resonance (NMR) Spectroscopy is a widely used technique to
predict the native structure of proteins. However, NMR machines are only able
to report approximate and partial distances between pair of atoms. To build the
protein structure one has to solve the Euclidean distance geometry problem
given the incomplete interval distance data produced by NMR machines. In this
paper, we propose a new genetic algorithm for solving the Euclidean distance
geometry problem for protein structure prediction given sparse NMR data. Our
genetic algorithm uses a greedy mutation operator to intensify the search, a
twin removal technique for diversification in the population and a random
restart method to recover stagnation. On a standard set of benchmark dataset,
our algorithm significantly outperforms standard genetic algorithms.
| [
{
"version": "v1",
"created": "Sun, 16 Nov 2014 11:26:06 GMT"
}
] | 2014-11-18T00:00:00 | [
[
"Islam",
"Md. Lisul",
""
],
[
"Shatabda",
"Swakkhar",
""
],
[
"Rahman",
"M. Sohel",
""
]
] | TITLE: GreMuTRRR: A Novel Genetic Algorithm to Solve Distance Geometry Problem
for Protein Structures
ABSTRACT: Nuclear Magnetic Resonance (NMR) Spectroscopy is a widely used technique to
predict the native structure of proteins. However, NMR machines are only able
to report approximate and partial distances between pair of atoms. To build the
protein structure one has to solve the Euclidean distance geometry problem
given the incomplete interval distance data produced by NMR machines. In this
paper, we propose a new genetic algorithm for solving the Euclidean distance
geometry problem for protein structure prediction given sparse NMR data. Our
genetic algorithm uses a greedy mutation operator to intensify the search, a
twin removal technique for diversification in the population and a random
restart method to recover stagnation. On a standard set of benchmark dataset,
our algorithm significantly outperforms standard genetic algorithms.
| no_new_dataset | 0.951953 |
1411.4304 | Rodrigo Benenson | Rodrigo Benenson, Mohamed Omran, Jan Hosang, Bernt Schiele | Ten Years of Pedestrian Detection, What Have We Learned? | To appear in ECCV 2014 CVRSUAD workshop proceedings | null | null | null | cs.CV | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | Paper-by-paper results make it easy to miss the forest for the trees.We
analyse the remarkable progress of the last decade by discussing the main ideas
explored in the 40+ detectors currently present in the Caltech pedestrian
detection benchmark. We observe that there exist three families of approaches,
all currently reaching similar detection quality. Based on our analysis, we
study the complementarity of the most promising ideas by combining multiple
published strategies. This new decision forest detector achieves the current
best known performance on the challenging Caltech-USA dataset.
| [
{
"version": "v1",
"created": "Sun, 16 Nov 2014 21:25:53 GMT"
}
] | 2014-11-18T00:00:00 | [
[
"Benenson",
"Rodrigo",
""
],
[
"Omran",
"Mohamed",
""
],
[
"Hosang",
"Jan",
""
],
[
"Schiele",
"Bernt",
""
]
] | TITLE: Ten Years of Pedestrian Detection, What Have We Learned?
ABSTRACT: Paper-by-paper results make it easy to miss the forest for the trees.We
analyse the remarkable progress of the last decade by discussing the main ideas
explored in the 40+ detectors currently present in the Caltech pedestrian
detection benchmark. We observe that there exist three families of approaches,
all currently reaching similar detection quality. Based on our analysis, we
study the complementarity of the most promising ideas by combining multiple
published strategies. This new decision forest detector achieves the current
best known performance on the challenging Caltech-USA dataset.
| no_new_dataset | 0.946646 |
1411.4314 | Nikolai Sinitsyn | Benjamin H. Sims, Nikolai Sinitsyn, and Stephan J. Eidenbenz | Hierarchical and Matrix Structures in a Large Organizational Email
Network: Visualization and Modeling Approaches | 15 pages, 9 figures | null | null | null | cs.SI physics.soc-ph stat.AP | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | This paper presents findings from a study of the email network of a large
scientific research organization, focusing on methods for visualizing and
modeling organizational hierarchies within large, complex network datasets. In
the first part of the paper, we find that visualization and interpretation of
complex organizational network data is facilitated by integration of network
data with information on formal organizational divisions and levels. By
aggregating and visualizing email traffic between organizational units at
various levels, we derive several insights into how large subdivisions of the
organization interact with each other and with outside organizations. Our
analysis shows that line and program management interactions in this
organization systematically deviate from the idealized pattern of interaction
prescribed by "matrix management." In the second part of the paper, we propose
a power law model for predicting degree distribution of organizational email
traffic based on hierarchical relationships between managers and employees.
This model considers the influence of global email announcements sent from
managers to all employees under their supervision, and the role support staff
play in generating email traffic, acting as agents for managers. We also
analyze patterns in email traffic volume over the course of a work week.
| [
{
"version": "v1",
"created": "Sun, 16 Nov 2014 22:22:54 GMT"
}
] | 2014-11-18T00:00:00 | [
[
"Sims",
"Benjamin H.",
""
],
[
"Sinitsyn",
"Nikolai",
""
],
[
"Eidenbenz",
"Stephan J.",
""
]
] | TITLE: Hierarchical and Matrix Structures in a Large Organizational Email
Network: Visualization and Modeling Approaches
ABSTRACT: This paper presents findings from a study of the email network of a large
scientific research organization, focusing on methods for visualizing and
modeling organizational hierarchies within large, complex network datasets. In
the first part of the paper, we find that visualization and interpretation of
complex organizational network data is facilitated by integration of network
data with information on formal organizational divisions and levels. By
aggregating and visualizing email traffic between organizational units at
various levels, we derive several insights into how large subdivisions of the
organization interact with each other and with outside organizations. Our
analysis shows that line and program management interactions in this
organization systematically deviate from the idealized pattern of interaction
prescribed by "matrix management." In the second part of the paper, we propose
a power law model for predicting degree distribution of organizational email
traffic based on hierarchical relationships between managers and employees.
This model considers the influence of global email announcements sent from
managers to all employees under their supervision, and the role support staff
play in generating email traffic, acting as agents for managers. We also
analyze patterns in email traffic volume over the course of a work week.
| no_new_dataset | 0.940188 |
1411.4379 | Md Lisul Islam | Md. Lisul Islam, Novia Nurain, Swakkhar Shatabda and M Sohel Rahman | FGPGA: An Efficient Genetic Approach for Producing Feasible Graph
Partitions | Accepted in the 1st International Conference on Networking Systems
and Security 2015 (NSysS 2015) | null | null | null | cs.NE cs.AI cs.DC | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | Graph partitioning, a well studied problem of parallel computing has many
applications in diversified fields such as distributed computing, social
network analysis, data mining and many other domains. In this paper, we
introduce FGPGA, an efficient genetic approach for producing feasible graph
partitions. Our method takes into account the heterogeneity and capacity
constraints of the partitions to ensure balanced partitioning. Such approach
has various applications in mobile cloud computing that include feasible
deployment of software applications on the more resourceful infrastructure in
the cloud instead of mobile hand set. Our proposed approach is light weight and
hence suitable for use in cloud architecture. We ensure feasibility of the
partitions generated by not allowing over-sized partitions to be generated
during the initialization and search. Our proposed method tested on standard
benchmark datasets significantly outperforms the state-of-the-art methods in
terms of quality of partitions and feasibility of the solutions.
| [
{
"version": "v1",
"created": "Mon, 17 Nov 2014 06:51:50 GMT"
}
] | 2014-11-18T00:00:00 | [
[
"Islam",
"Md. Lisul",
""
],
[
"Nurain",
"Novia",
""
],
[
"Shatabda",
"Swakkhar",
""
],
[
"Rahman",
"M Sohel",
""
]
] | TITLE: FGPGA: An Efficient Genetic Approach for Producing Feasible Graph
Partitions
ABSTRACT: Graph partitioning, a well studied problem of parallel computing has many
applications in diversified fields such as distributed computing, social
network analysis, data mining and many other domains. In this paper, we
introduce FGPGA, an efficient genetic approach for producing feasible graph
partitions. Our method takes into account the heterogeneity and capacity
constraints of the partitions to ensure balanced partitioning. Such approach
has various applications in mobile cloud computing that include feasible
deployment of software applications on the more resourceful infrastructure in
the cloud instead of mobile hand set. Our proposed approach is light weight and
hence suitable for use in cloud architecture. We ensure feasibility of the
partitions generated by not allowing over-sized partitions to be generated
during the initialization and search. Our proposed method tested on standard
benchmark datasets significantly outperforms the state-of-the-art methods in
terms of quality of partitions and feasibility of the solutions.
| no_new_dataset | 0.950503 |
1411.4455 | Miao Fan | Miao Fan, Deli Zhao, Qiang Zhou, Zhiyuan Liu, Thomas Fang Zheng,
Edward Y. Chang | Errata: Distant Supervision for Relation Extraction with Matrix
Completion | null | null | null | null | cs.CL cs.LG | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | The essence of distantly supervised relation extraction is that it is an
incomplete multi-label classification problem with sparse and noisy features.
To tackle the sparsity and noise challenges, we propose solving the
classification problem using matrix completion on factorized matrix of
minimized rank. We formulate relation classification as completing the unknown
labels of testing items (entity pairs) in a sparse matrix that concatenates
training and testing textual features with training labels. Our algorithmic
framework is based on the assumption that the rank of item-by-feature and
item-by-label joint matrix is low. We apply two optimization models to recover
the underlying low-rank matrix leveraging the sparsity of feature-label matrix.
The matrix completion problem is then solved by the fixed point continuation
(FPC) algorithm, which can find the global optimum. Experiments on two widely
used datasets with different dimensions of textual features demonstrate that
our low-rank matrix completion approach significantly outperforms the baseline
and the state-of-the-art methods.
| [
{
"version": "v1",
"created": "Mon, 17 Nov 2014 12:43:30 GMT"
}
] | 2014-11-18T00:00:00 | [
[
"Fan",
"Miao",
""
],
[
"Zhao",
"Deli",
""
],
[
"Zhou",
"Qiang",
""
],
[
"Liu",
"Zhiyuan",
""
],
[
"Zheng",
"Thomas Fang",
""
],
[
"Chang",
"Edward Y.",
""
]
] | TITLE: Errata: Distant Supervision for Relation Extraction with Matrix
Completion
ABSTRACT: The essence of distantly supervised relation extraction is that it is an
incomplete multi-label classification problem with sparse and noisy features.
To tackle the sparsity and noise challenges, we propose solving the
classification problem using matrix completion on factorized matrix of
minimized rank. We formulate relation classification as completing the unknown
labels of testing items (entity pairs) in a sparse matrix that concatenates
training and testing textual features with training labels. Our algorithmic
framework is based on the assumption that the rank of item-by-feature and
item-by-label joint matrix is low. We apply two optimization models to recover
the underlying low-rank matrix leveraging the sparsity of feature-label matrix.
The matrix completion problem is then solved by the fixed point continuation
(FPC) algorithm, which can find the global optimum. Experiments on two widely
used datasets with different dimensions of textual features demonstrate that
our low-rank matrix completion approach significantly outperforms the baseline
and the state-of-the-art methods.
| no_new_dataset | 0.939969 |
1411.4464 | Kai Kang | Kai Kang, Xiaogang Wang | Fully Convolutional Neural Networks for Crowd Segmentation | 9 pages,7 figures | null | null | null | cs.CV | http://creativecommons.org/licenses/by-nc-sa/3.0/ | In this paper, we propose a fast fully convolutional neural network (FCNN)
for crowd segmentation. By replacing the fully connected layers in CNN with 1
by 1 convolution kernels, FCNN takes whole images as inputs and directly
outputs segmentation maps by one pass of forward propagation. It has the
property of translation invariance like patch-by-patch scanning but with much
lower computation cost. Once FCNN is learned, it can process input images of
any sizes without warping them to a standard size. These attractive properties
make it extendable to other general image segmentation problems. Based on FCNN,
a multi-stage deep learning is proposed to integrate appearance and motion cues
for crowd segmentation. Both appearance filters and motion filers are
pretrained stage-by-stage and then jointly optimized. Different combination
methods are investigated. The effectiveness of our approach and component-wise
analysis are evaluated on two crowd segmentation datasets created by us, which
include image frames from 235 and 11 scenes, respectively. They are currently
the largest crowd segmentation datasets and will be released to the public.
| [
{
"version": "v1",
"created": "Mon, 17 Nov 2014 13:09:09 GMT"
}
] | 2014-11-18T00:00:00 | [
[
"Kang",
"Kai",
""
],
[
"Wang",
"Xiaogang",
""
]
] | TITLE: Fully Convolutional Neural Networks for Crowd Segmentation
ABSTRACT: In this paper, we propose a fast fully convolutional neural network (FCNN)
for crowd segmentation. By replacing the fully connected layers in CNN with 1
by 1 convolution kernels, FCNN takes whole images as inputs and directly
outputs segmentation maps by one pass of forward propagation. It has the
property of translation invariance like patch-by-patch scanning but with much
lower computation cost. Once FCNN is learned, it can process input images of
any sizes without warping them to a standard size. These attractive properties
make it extendable to other general image segmentation problems. Based on FCNN,
a multi-stage deep learning is proposed to integrate appearance and motion cues
for crowd segmentation. Both appearance filters and motion filers are
pretrained stage-by-stage and then jointly optimized. Different combination
methods are investigated. The effectiveness of our approach and component-wise
analysis are evaluated on two crowd segmentation datasets created by us, which
include image frames from 235 and 11 scenes, respectively. They are currently
the largest crowd segmentation datasets and will be released to the public.
| no_new_dataset | 0.938463 |
1411.4472 | Andrej Gajduk | Andrej Gajduk and Ljupco Kocarev | Opinion mining of text documents written in Macedonian language | In press, MASA proceedings | null | null | null | cs.CL | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | The ability to extract public opinion from web portals such as review sites,
social networks and blogs will enable companies and individuals to form a view,
an attitude and make decisions without having to do lengthy and costly
researches and surveys. In this paper machine learning techniques are used for
determining the polarity of forum posts on kajgana which are written in
Macedonian language. The posts are classified as being positive, negative or
neutral. We test different feature metrics and classifiers and provide detailed
evaluation of their participation in improving the overall performance on a
manually generated dataset. By achieving 92% accuracy, we show that the
performance of systems for automated opinion mining is comparable to a human
evaluator, thus making it a viable option for text data analysis. Finally, we
present a few statistics derived from the forum posts using the developed
system.
| [
{
"version": "v1",
"created": "Mon, 17 Nov 2014 13:36:49 GMT"
}
] | 2014-11-18T00:00:00 | [
[
"Gajduk",
"Andrej",
""
],
[
"Kocarev",
"Ljupco",
""
]
] | TITLE: Opinion mining of text documents written in Macedonian language
ABSTRACT: The ability to extract public opinion from web portals such as review sites,
social networks and blogs will enable companies and individuals to form a view,
an attitude and make decisions without having to do lengthy and costly
researches and surveys. In this paper machine learning techniques are used for
determining the polarity of forum posts on kajgana which are written in
Macedonian language. The posts are classified as being positive, negative or
neutral. We test different feature metrics and classifiers and provide detailed
evaluation of their participation in improving the overall performance on a
manually generated dataset. By achieving 92% accuracy, we show that the
performance of systems for automated opinion mining is comparable to a human
evaluator, thus making it a viable option for text data analysis. Finally, we
present a few statistics derived from the forum posts using the developed
system.
| new_dataset | 0.954774 |
1411.4510 | Kian Hsiang Low | Kian Hsiang Low, Jiangbo Yu, Jie Chen, Patrick Jaillet | Parallel Gaussian Process Regression for Big Data: Low-Rank
Representation Meets Markov Approximation | 29th AAAI Conference on Artificial Intelligence (AAAI 2015), Extended
version with proofs, 10 pages | null | null | null | stat.ML cs.DC cs.LG | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | The expressive power of a Gaussian process (GP) model comes at a cost of poor
scalability in the data size. To improve its scalability, this paper presents a
low-rank-cum-Markov approximation (LMA) of the GP model that is novel in
leveraging the dual computational advantages stemming from complementing a
low-rank approximate representation of the full-rank GP based on a support set
of inputs with a Markov approximation of the resulting residual process; the
latter approximation is guaranteed to be closest in the Kullback-Leibler
distance criterion subject to some constraint and is considerably more refined
than that of existing sparse GP models utilizing low-rank representations due
to its more relaxed conditional independence assumption (especially with larger
data). As a result, our LMA method can trade off between the size of the
support set and the order of the Markov property to (a) incur lower
computational cost than such sparse GP models while achieving predictive
performance comparable to them and (b) accurately represent features/patterns
of any scale. Interestingly, varying the Markov order produces a spectrum of
LMAs with PIC approximation and full-rank GP at the two extremes. An advantage
of our LMA method is that it is amenable to parallelization on multiple
machines/cores, thereby gaining greater scalability. Empirical evaluation on
three real-world datasets in clusters of up to 32 computing nodes shows that
our centralized and parallel LMA methods are significantly more time-efficient
and scalable than state-of-the-art sparse and full-rank GP regression methods
while achieving comparable predictive performances.
| [
{
"version": "v1",
"created": "Mon, 17 Nov 2014 15:31:04 GMT"
}
] | 2014-11-18T00:00:00 | [
[
"Low",
"Kian Hsiang",
""
],
[
"Yu",
"Jiangbo",
""
],
[
"Chen",
"Jie",
""
],
[
"Jaillet",
"Patrick",
""
]
] | TITLE: Parallel Gaussian Process Regression for Big Data: Low-Rank
Representation Meets Markov Approximation
ABSTRACT: The expressive power of a Gaussian process (GP) model comes at a cost of poor
scalability in the data size. To improve its scalability, this paper presents a
low-rank-cum-Markov approximation (LMA) of the GP model that is novel in
leveraging the dual computational advantages stemming from complementing a
low-rank approximate representation of the full-rank GP based on a support set
of inputs with a Markov approximation of the resulting residual process; the
latter approximation is guaranteed to be closest in the Kullback-Leibler
distance criterion subject to some constraint and is considerably more refined
than that of existing sparse GP models utilizing low-rank representations due
to its more relaxed conditional independence assumption (especially with larger
data). As a result, our LMA method can trade off between the size of the
support set and the order of the Markov property to (a) incur lower
computational cost than such sparse GP models while achieving predictive
performance comparable to them and (b) accurately represent features/patterns
of any scale. Interestingly, varying the Markov order produces a spectrum of
LMAs with PIC approximation and full-rank GP at the two extremes. An advantage
of our LMA method is that it is amenable to parallelization on multiple
machines/cores, thereby gaining greater scalability. Empirical evaluation on
three real-world datasets in clusters of up to 32 computing nodes shows that
our centralized and parallel LMA methods are significantly more time-efficient
and scalable than state-of-the-art sparse and full-rank GP regression methods
while achieving comparable predictive performances.
| no_new_dataset | 0.948251 |
physics/0606042 | Suen Hou | J. Antos, M. Babik, D. Benjamin, S. Cabrera, A.W. Chan, Y.C. Chen, M.
Coca, B. Cooper, S. Farrington, K. Genser, K. Hatakeyama, S. Hou, T.L. Hsieh,
B. Jayatilaka, S.Y. Jun, A.V. Kotwal, A.C. Kraan, R. Lysak, I.V.
Mandrichenko, P. Murat, A. Robson, P. Savard, M. Siket, B. Stelzer, J. Syu,
P.K. Teng, S.C. Timm, T. Tomura, E. Vataga, and S.A. Wolbers | Data processing model for the CDF experiment | 12 pages, 10 figures, submitted to IEEE-TNS | IEEE Trans.Nucl.Sci.53:2897-2906,2006 | 10.1109/TNS.2006.881908 | FERMILAB-PUB-06-169-CD-E | physics.ins-det physics.data-an | null | The data processing model for the CDF experiment is described. Data
processing reconstructs events from parallel data streams taken with different
combinations of physics event triggers and further splits the events into
datasets of specialized physics datasets. The design of the processing control
system faces strict requirements on bookkeeping records, which trace the status
of data files and event contents during processing and storage. The computing
architecture was updated to meet the mass data flow of the Run II data
collection, recently upgraded to a maximum rate of 40 MByte/sec. The data
processing facility consists of a large cluster of Linux computers with data
movement managed by the CDF data handling system to a multi-petaByte Enstore
tape library. The latest processing cycle has achieved a stable speed of 35
MByte/sec (3 TByte/day). It can be readily scaled by increasing CPU and
data-handling capacity as required.
| [
{
"version": "v1",
"created": "Mon, 5 Jun 2006 16:37:59 GMT"
},
{
"version": "v2",
"created": "Fri, 9 Jun 2006 06:23:26 GMT"
}
] | 2014-11-18T00:00:00 | [
[
"Antos",
"J.",
""
],
[
"Babik",
"M.",
""
],
[
"Benjamin",
"D.",
""
],
[
"Cabrera",
"S.",
""
],
[
"Chan",
"A. W.",
""
],
[
"Chen",
"Y. C.",
""
],
[
"Coca",
"M.",
""
],
[
"Cooper",
"B.",
""
],
[
"Farrington",
"S.",
""
],
[
"Genser",
"K.",
""
],
[
"Hatakeyama",
"K.",
""
],
[
"Hou",
"S.",
""
],
[
"Hsieh",
"T. L.",
""
],
[
"Jayatilaka",
"B.",
""
],
[
"Jun",
"S. Y.",
""
],
[
"Kotwal",
"A. V.",
""
],
[
"Kraan",
"A. C.",
""
],
[
"Lysak",
"R.",
""
],
[
"Mandrichenko",
"I. V.",
""
],
[
"Murat",
"P.",
""
],
[
"Robson",
"A.",
""
],
[
"Savard",
"P.",
""
],
[
"Siket",
"M.",
""
],
[
"Stelzer",
"B.",
""
],
[
"Syu",
"J.",
""
],
[
"Teng",
"P. K.",
""
],
[
"Timm",
"S. C.",
""
],
[
"Tomura",
"T.",
""
],
[
"Vataga",
"E.",
""
],
[
"Wolbers",
"S. A.",
""
]
] | TITLE: Data processing model for the CDF experiment
ABSTRACT: The data processing model for the CDF experiment is described. Data
processing reconstructs events from parallel data streams taken with different
combinations of physics event triggers and further splits the events into
datasets of specialized physics datasets. The design of the processing control
system faces strict requirements on bookkeeping records, which trace the status
of data files and event contents during processing and storage. The computing
architecture was updated to meet the mass data flow of the Run II data
collection, recently upgraded to a maximum rate of 40 MByte/sec. The data
processing facility consists of a large cluster of Linux computers with data
movement managed by the CDF data handling system to a multi-petaByte Enstore
tape library. The latest processing cycle has achieved a stable speed of 35
MByte/sec (3 TByte/day). It can be readily scaled by increasing CPU and
data-handling capacity as required.
| no_new_dataset | 0.936168 |
1411.3159 | Marcel Simon | Marcel Simon, Erik Rodner, Joachim Denzler | Part Detector Discovery in Deep Convolutional Neural Networks | Accepted for publication on Asian Conference on Computer Vision
(ACCV) 2014 | null | null | null | cs.CV | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | Current fine-grained classification approaches often rely on a robust
localization of object parts to extract localized feature representations
suitable for discrimination. However, part localization is a challenging task
due to the large variation of appearance and pose. In this paper, we show how
pre-trained convolutional neural networks can be used for robust and efficient
object part discovery and localization without the necessity to actually train
the network on the current dataset. Our approach called "part detector
discovery" (PDD) is based on analyzing the gradient maps of the network outputs
and finding activation centers spatially related to annotated semantic parts or
bounding boxes.
This allows us not just to obtain excellent performance on the CUB200-2011
dataset, but in contrast to previous approaches also to perform detection and
bird classification jointly without requiring a given bounding box annotation
during testing and ground-truth parts during training. The code is available at
http://www.inf-cv.uni-jena.de/part_discovery and
https://github.com/cvjena/PartDetectorDisovery.
| [
{
"version": "v1",
"created": "Wed, 12 Nov 2014 12:42:54 GMT"
},
{
"version": "v2",
"created": "Fri, 14 Nov 2014 11:57:27 GMT"
}
] | 2014-11-17T00:00:00 | [
[
"Simon",
"Marcel",
""
],
[
"Rodner",
"Erik",
""
],
[
"Denzler",
"Joachim",
""
]
] | TITLE: Part Detector Discovery in Deep Convolutional Neural Networks
ABSTRACT: Current fine-grained classification approaches often rely on a robust
localization of object parts to extract localized feature representations
suitable for discrimination. However, part localization is a challenging task
due to the large variation of appearance and pose. In this paper, we show how
pre-trained convolutional neural networks can be used for robust and efficient
object part discovery and localization without the necessity to actually train
the network on the current dataset. Our approach called "part detector
discovery" (PDD) is based on analyzing the gradient maps of the network outputs
and finding activation centers spatially related to annotated semantic parts or
bounding boxes.
This allows us not just to obtain excellent performance on the CUB200-2011
dataset, but in contrast to previous approaches also to perform detection and
bird classification jointly without requiring a given bounding box annotation
during testing and ground-truth parts during training. The code is available at
http://www.inf-cv.uni-jena.de/part_discovery and
https://github.com/cvjena/PartDetectorDisovery.
| no_new_dataset | 0.95511 |
1411.3749 | Timothy La Fond | Timothy La Fond, Jennifer Neville, Brian Gallagher | Anomaly Detection in Dynamic Networks of Varying Size | null | null | null | null | cs.SI physics.soc-ph | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | Dynamic networks, also called network streams, are an important data
representation that applies to many real-world domains. Many sets of network
data such as e-mail networks, social networks, or internet traffic networks are
best represented by a dynamic network due to the temporal component of the
data. One important application in the domain of dynamic network analysis is
anomaly detection. Here the task is to identify points in time where the
network exhibits behavior radically different from a typical time, either due
to some event (like the failure of machines in a computer network) or a shift
in the network properties. This problem is made more difficult by the fluid
nature of what is considered "normal" network behavior. The volume of traffic
on a network, for example, can change over the course of a month or even vary
based on the time of the day without being considered unusual. Anomaly
detection tests using traditional network statistics have difficulty in these
scenarios due to their Density Dependence: as the volume of edges changes the
value of the statistics changes as well making it difficult to determine if the
change in signal is due to the traffic volume or due to some fundamental shift
in the behavior of the network. To more accurately detect anomalies in dynamic
networks, we introduce the concept of Density-Consistent network statistics. On
synthetically generated graphs anomaly detectors using these statistics show a
a 20-400% improvement in the recall when distinguishing graphs drawn from
different distributions. When applied to several real datasets
Density-Consistent statistics recover multiple network events which standard
statistics failed to find.
| [
{
"version": "v1",
"created": "Thu, 13 Nov 2014 21:41:55 GMT"
}
] | 2014-11-17T00:00:00 | [
[
"La Fond",
"Timothy",
""
],
[
"Neville",
"Jennifer",
""
],
[
"Gallagher",
"Brian",
""
]
] | TITLE: Anomaly Detection in Dynamic Networks of Varying Size
ABSTRACT: Dynamic networks, also called network streams, are an important data
representation that applies to many real-world domains. Many sets of network
data such as e-mail networks, social networks, or internet traffic networks are
best represented by a dynamic network due to the temporal component of the
data. One important application in the domain of dynamic network analysis is
anomaly detection. Here the task is to identify points in time where the
network exhibits behavior radically different from a typical time, either due
to some event (like the failure of machines in a computer network) or a shift
in the network properties. This problem is made more difficult by the fluid
nature of what is considered "normal" network behavior. The volume of traffic
on a network, for example, can change over the course of a month or even vary
based on the time of the day without being considered unusual. Anomaly
detection tests using traditional network statistics have difficulty in these
scenarios due to their Density Dependence: as the volume of edges changes the
value of the statistics changes as well making it difficult to determine if the
change in signal is due to the traffic volume or due to some fundamental shift
in the behavior of the network. To more accurately detect anomalies in dynamic
networks, we introduce the concept of Density-Consistent network statistics. On
synthetically generated graphs anomaly detectors using these statistics show a
a 20-400% improvement in the recall when distinguishing graphs drawn from
different distributions. When applied to several real datasets
Density-Consistent statistics recover multiple network events which standard
statistics failed to find.
| no_new_dataset | 0.949342 |
1411.3787 | Ping Li | Anshumali Shrivastava, Ping Li | Asymmetric Minwise Hashing | null | null | null | null | stat.ML cs.DB cs.DS cs.IR cs.LG | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | Minwise hashing (Minhash) is a widely popular indexing scheme in practice.
Minhash is designed for estimating set resemblance and is known to be
suboptimal in many applications where the desired measure is set overlap (i.e.,
inner product between binary vectors) or set containment. Minhash has inherent
bias towards smaller sets, which adversely affects its performance in
applications where such a penalization is not desirable. In this paper, we
propose asymmetric minwise hashing (MH-ALSH), to provide a solution to this
problem. The new scheme utilizes asymmetric transformations to cancel the bias
of traditional minhash towards smaller sets, making the final "collision
probability" monotonic in the inner product. Our theoretical comparisons show
that for the task of retrieving with binary inner products asymmetric minhash
is provably better than traditional minhash and other recently proposed hashing
algorithms for general inner products. Thus, we obtain an algorithmic
improvement over existing approaches in the literature. Experimental
evaluations on four publicly available high-dimensional datasets validate our
claims and the proposed scheme outperforms, often significantly, other hashing
algorithms on the task of near neighbor retrieval with set containment. Our
proposal is simple and easy to implement in practice.
| [
{
"version": "v1",
"created": "Fri, 14 Nov 2014 04:18:33 GMT"
}
] | 2014-11-17T00:00:00 | [
[
"Shrivastava",
"Anshumali",
""
],
[
"Li",
"Ping",
""
]
] | TITLE: Asymmetric Minwise Hashing
ABSTRACT: Minwise hashing (Minhash) is a widely popular indexing scheme in practice.
Minhash is designed for estimating set resemblance and is known to be
suboptimal in many applications where the desired measure is set overlap (i.e.,
inner product between binary vectors) or set containment. Minhash has inherent
bias towards smaller sets, which adversely affects its performance in
applications where such a penalization is not desirable. In this paper, we
propose asymmetric minwise hashing (MH-ALSH), to provide a solution to this
problem. The new scheme utilizes asymmetric transformations to cancel the bias
of traditional minhash towards smaller sets, making the final "collision
probability" monotonic in the inner product. Our theoretical comparisons show
that for the task of retrieving with binary inner products asymmetric minhash
is provably better than traditional minhash and other recently proposed hashing
algorithms for general inner products. Thus, we obtain an algorithmic
improvement over existing approaches in the literature. Experimental
evaluations on four publicly available high-dimensional datasets validate our
claims and the proposed scheme outperforms, often significantly, other hashing
algorithms on the task of near neighbor retrieval with set containment. Our
proposal is simple and easy to implement in practice.
| no_new_dataset | 0.949248 |
1411.4006 | Zhongwen Xu | Zhongwen Xu, Yi Yang and Alexander G. Hauptmann | A Discriminative CNN Video Representation for Event Detection | null | null | null | null | cs.CV | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | In this paper, we propose a discriminative video representation for event
detection over a large scale video dataset when only limited hardware resources
are available. The focus of this paper is to effectively leverage deep
Convolutional Neural Networks (CNNs) to advance event detection, where only
frame level static descriptors can be extracted by the existing CNN toolkit.
This paper makes two contributions to the inference of CNN video
representation. First, while average pooling and max pooling have long been the
standard approaches to aggregating frame level static features, we show that
performance can be significantly improved by taking advantage of an appropriate
encoding method. Second, we propose using a set of latent concept descriptors
as the frame descriptor, which enriches visual information while keeping it
computationally affordable. The integration of the two contributions results in
a new state-of-the-art performance in event detection over the largest video
datasets. Compared to improved Dense Trajectories, which has been recognized as
the best video representation for event detection, our new representation
improves the Mean Average Precision (mAP) from 27.6% to 36.8% for the TRECVID
MEDTest 14 dataset and from 34.0% to 44.6% for the TRECVID MEDTest 13 dataset.
This work is the core part of the winning solution of our CMU-Informedia team
in TRECVID MED 2014 competition.
| [
{
"version": "v1",
"created": "Fri, 14 Nov 2014 18:37:31 GMT"
}
] | 2014-11-17T00:00:00 | [
[
"Xu",
"Zhongwen",
""
],
[
"Yang",
"Yi",
""
],
[
"Hauptmann",
"Alexander G.",
""
]
] | TITLE: A Discriminative CNN Video Representation for Event Detection
ABSTRACT: In this paper, we propose a discriminative video representation for event
detection over a large scale video dataset when only limited hardware resources
are available. The focus of this paper is to effectively leverage deep
Convolutional Neural Networks (CNNs) to advance event detection, where only
frame level static descriptors can be extracted by the existing CNN toolkit.
This paper makes two contributions to the inference of CNN video
representation. First, while average pooling and max pooling have long been the
standard approaches to aggregating frame level static features, we show that
performance can be significantly improved by taking advantage of an appropriate
encoding method. Second, we propose using a set of latent concept descriptors
as the frame descriptor, which enriches visual information while keeping it
computationally affordable. The integration of the two contributions results in
a new state-of-the-art performance in event detection over the largest video
datasets. Compared to improved Dense Trajectories, which has been recognized as
the best video representation for event detection, our new representation
improves the Mean Average Precision (mAP) from 27.6% to 36.8% for the TRECVID
MEDTest 14 dataset and from 34.0% to 44.6% for the TRECVID MEDTest 13 dataset.
This work is the core part of the winning solution of our CMU-Informedia team
in TRECVID MED 2014 competition.
| no_new_dataset | 0.948537 |
cs/9508101 | null | Q. Zhao, T. Nishida | Using Qualitative Hypotheses to Identify Inaccurate Data | See http://www.jair.org/ for any accompanying files | Journal of Artificial Intelligence Research, Vol 3, (1995),
119-145 | null | null | cs.AI | null | Identifying inaccurate data has long been regarded as a significant and
difficult problem in AI. In this paper, we present a new method for identifying
inaccurate data on the basis of qualitative correlations among related data.
First, we introduce the definitions of related data and qualitative
correlations among related data. Then we put forward a new concept called
support coefficient function (SCF). SCF can be used to extract, represent, and
calculate qualitative correlations among related data within a dataset. We
propose an approach to determining dynamic shift intervals of inaccurate data,
and an approach to calculating possibility of identifying inaccurate data,
respectively. Both of the approaches are based on SCF. Finally we present an
algorithm for identifying inaccurate data by using qualitative correlations
among related data as confirmatory or disconfirmatory evidence. We have
developed a practical system for interpreting infrared spectra by applying the
method, and have fully tested the system against several hundred real spectra.
The experimental results show that the method is significantly better than the
conventional methods used in many similar systems.
| [
{
"version": "v1",
"created": "Tue, 1 Aug 1995 00:00:00 GMT"
}
] | 2014-11-17T00:00:00 | [
[
"Zhao",
"Q.",
""
],
[
"Nishida",
"T.",
""
]
] | TITLE: Using Qualitative Hypotheses to Identify Inaccurate Data
ABSTRACT: Identifying inaccurate data has long been regarded as a significant and
difficult problem in AI. In this paper, we present a new method for identifying
inaccurate data on the basis of qualitative correlations among related data.
First, we introduce the definitions of related data and qualitative
correlations among related data. Then we put forward a new concept called
support coefficient function (SCF). SCF can be used to extract, represent, and
calculate qualitative correlations among related data within a dataset. We
propose an approach to determining dynamic shift intervals of inaccurate data,
and an approach to calculating possibility of identifying inaccurate data,
respectively. Both of the approaches are based on SCF. Finally we present an
algorithm for identifying inaccurate data by using qualitative correlations
among related data as confirmatory or disconfirmatory evidence. We have
developed a practical system for interpreting infrared spectra by applying the
method, and have fully tested the system against several hundred real spectra.
The experimental results show that the method is significantly better than the
conventional methods used in many similar systems.
| no_new_dataset | 0.948585 |
1411.3374 | Manas Joglekar | Manas Joglekar, Hector Garcia-Molina, Aditya Parameswaran, Christopher
Re | Exploiting Correlations for Expensive Predicate Evaluation | null | null | null | null | cs.DB | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | User Defined Function(UDFs) are used increasingly to augment query languages
with extra, application dependent functionality. Selection queries involving
UDF predicates tend to be expensive, either in terms of monetary cost or
latency. In this paper, we study ways to efficiently evaluate selection queries
with UDF predicates. We provide a family of techniques for processing queries
at low cost while satisfying user-specified precision and recall constraints.
Our techniques are applicable to a variety of scenarios including when
selection probabilities of tuples are available beforehand, when this
information is available but noisy, or when no such prior information is
available. We also generalize our techniques to more complex queries. Finally,
we test our techniques on real datasets, and show that they achieve significant
savings in cost of up to $80\%$, while incurring only a small reduction in
accuracy.
| [
{
"version": "v1",
"created": "Wed, 12 Nov 2014 22:00:08 GMT"
}
] | 2014-11-14T00:00:00 | [
[
"Joglekar",
"Manas",
""
],
[
"Garcia-Molina",
"Hector",
""
],
[
"Parameswaran",
"Aditya",
""
],
[
"Re",
"Christopher",
""
]
] | TITLE: Exploiting Correlations for Expensive Predicate Evaluation
ABSTRACT: User Defined Function(UDFs) are used increasingly to augment query languages
with extra, application dependent functionality. Selection queries involving
UDF predicates tend to be expensive, either in terms of monetary cost or
latency. In this paper, we study ways to efficiently evaluate selection queries
with UDF predicates. We provide a family of techniques for processing queries
at low cost while satisfying user-specified precision and recall constraints.
Our techniques are applicable to a variety of scenarios including when
selection probabilities of tuples are available beforehand, when this
information is available but noisy, or when no such prior information is
available. We also generalize our techniques to more complex queries. Finally,
we test our techniques on real datasets, and show that they achieve significant
savings in cost of up to $80\%$, while incurring only a small reduction in
accuracy.
| no_new_dataset | 0.94699 |
1411.3377 | Manas Joglekar | Manas Joglekar, Hector Garcia-Molina, Aditya Parameswaran | Comprehensive and Reliable Crowd Assessment Algorithms | ICDE 2015 | null | null | null | cs.DB | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | Evaluating workers is a critical aspect of any crowdsourcing system. In this
paper, we devise techniques for evaluating workers by finding confidence
intervals on their error rates. Unlike prior work, we focus on
"conciseness"---that is, giving as tight a confidence interval as possible.
Conciseness is of utmost importance because it allows us to be sure that we
have the best guarantee possible on worker error rate. Also unlike prior work,
we provide techniques that work under very general scenarios, such as when not
all workers have attempted every task (a fairly common scenario in practice),
when tasks have non-boolean responses, and when workers have different biases
for positive and negative tasks. We demonstrate conciseness as well as accuracy
of our confidence intervals by testing them on a variety of conditions and
multiple real-world datasets.
| [
{
"version": "v1",
"created": "Wed, 12 Nov 2014 22:09:17 GMT"
}
] | 2014-11-14T00:00:00 | [
[
"Joglekar",
"Manas",
""
],
[
"Garcia-Molina",
"Hector",
""
],
[
"Parameswaran",
"Aditya",
""
]
] | TITLE: Comprehensive and Reliable Crowd Assessment Algorithms
ABSTRACT: Evaluating workers is a critical aspect of any crowdsourcing system. In this
paper, we devise techniques for evaluating workers by finding confidence
intervals on their error rates. Unlike prior work, we focus on
"conciseness"---that is, giving as tight a confidence interval as possible.
Conciseness is of utmost importance because it allows us to be sure that we
have the best guarantee possible on worker error rate. Also unlike prior work,
we provide techniques that work under very general scenarios, such as when not
all workers have attempted every task (a fairly common scenario in practice),
when tasks have non-boolean responses, and when workers have different biases
for positive and negative tasks. We demonstrate conciseness as well as accuracy
of our confidence intervals by testing them on a variety of conditions and
multiple real-world datasets.
| no_new_dataset | 0.953188 |
1411.3409 | Paul Mineiro | Paul Mineiro, Nikos Karampatziakis | A Randomized Algorithm for CCA | null | null | null | null | stat.ML cs.LG | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | We present RandomizedCCA, a randomized algorithm for computing canonical
analysis, suitable for large datasets stored either out of core or on a
distributed file system. Accurate results can be obtained in as few as two data
passes, which is relevant for distributed processing frameworks in which
iteration is expensive (e.g., Hadoop). The strategy also provides an excellent
initializer for standard iterative solutions.
| [
{
"version": "v1",
"created": "Thu, 13 Nov 2014 00:51:19 GMT"
}
] | 2014-11-14T00:00:00 | [
[
"Mineiro",
"Paul",
""
],
[
"Karampatziakis",
"Nikos",
""
]
] | TITLE: A Randomized Algorithm for CCA
ABSTRACT: We present RandomizedCCA, a randomized algorithm for computing canonical
analysis, suitable for large datasets stored either out of core or on a
distributed file system. Accurate results can be obtained in as few as two data
passes, which is relevant for distributed processing frameworks in which
iteration is expensive (e.g., Hadoop). The strategy also provides an excellent
initializer for standard iterative solutions.
| no_new_dataset | 0.94474 |
1411.3410 | Kwangchol Jang | Kwangchol Jang, Sokmin Han, Insong Kim | Person Re-identification Based on Color Histogram and Spatial
Configuration of Dominant Color Regions | 12 pages, 6 figures | null | null | null | cs.CV | http://creativecommons.org/licenses/publicdomain/ | There is a requirement to determine whether a given person of interest has
already been observed over a network of cameras in video surveillance systems.
A human appearance obtained in one camera is usually different from the ones
obtained in another camera due to difference in illumination, pose and
viewpoint, camera parameters. Being related to appearance-based approaches for
person re-identification, we propose a novel method based on the dominant color
histogram and spatial configuration of dominant color regions on human body
parts. Dominant color histogram and spatial configuration of the dominant color
regions based on dominant color descriptor(DCD) can be considered to be robust
to illumination and pose, viewpoint changes. The proposed method is evaluated
using benchmark video datasets. Experimental results using the cumulative
matching characteristic(CMC) curve demonstrate the effectiveness of our
approach for person re-identification.
| [
{
"version": "v1",
"created": "Thu, 13 Nov 2014 00:55:48 GMT"
}
] | 2014-11-14T00:00:00 | [
[
"Jang",
"Kwangchol",
""
],
[
"Han",
"Sokmin",
""
],
[
"Kim",
"Insong",
""
]
] | TITLE: Person Re-identification Based on Color Histogram and Spatial
Configuration of Dominant Color Regions
ABSTRACT: There is a requirement to determine whether a given person of interest has
already been observed over a network of cameras in video surveillance systems.
A human appearance obtained in one camera is usually different from the ones
obtained in another camera due to difference in illumination, pose and
viewpoint, camera parameters. Being related to appearance-based approaches for
person re-identification, we propose a novel method based on the dominant color
histogram and spatial configuration of dominant color regions on human body
parts. Dominant color histogram and spatial configuration of the dominant color
regions based on dominant color descriptor(DCD) can be considered to be robust
to illumination and pose, viewpoint changes. The proposed method is evaluated
using benchmark video datasets. Experimental results using the cumulative
matching characteristic(CMC) curve demonstrate the effectiveness of our
approach for person re-identification.
| no_new_dataset | 0.951504 |
1406.2199 | Karen Simonyan | Karen Simonyan, Andrew Zisserman | Two-Stream Convolutional Networks for Action Recognition in Videos | null | null | null | null | cs.CV | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | We investigate architectures of discriminatively trained deep Convolutional
Networks (ConvNets) for action recognition in video. The challenge is to
capture the complementary information on appearance from still frames and
motion between frames. We also aim to generalise the best performing
hand-crafted features within a data-driven learning framework.
Our contribution is three-fold. First, we propose a two-stream ConvNet
architecture which incorporates spatial and temporal networks. Second, we
demonstrate that a ConvNet trained on multi-frame dense optical flow is able to
achieve very good performance in spite of limited training data. Finally, we
show that multi-task learning, applied to two different action classification
datasets, can be used to increase the amount of training data and improve the
performance on both.
Our architecture is trained and evaluated on the standard video actions
benchmarks of UCF-101 and HMDB-51, where it is competitive with the state of
the art. It also exceeds by a large margin previous attempts to use deep nets
for video classification.
| [
{
"version": "v1",
"created": "Mon, 9 Jun 2014 14:44:14 GMT"
},
{
"version": "v2",
"created": "Wed, 12 Nov 2014 20:48:33 GMT"
}
] | 2014-11-13T00:00:00 | [
[
"Simonyan",
"Karen",
""
],
[
"Zisserman",
"Andrew",
""
]
] | TITLE: Two-Stream Convolutional Networks for Action Recognition in Videos
ABSTRACT: We investigate architectures of discriminatively trained deep Convolutional
Networks (ConvNets) for action recognition in video. The challenge is to
capture the complementary information on appearance from still frames and
motion between frames. We also aim to generalise the best performing
hand-crafted features within a data-driven learning framework.
Our contribution is three-fold. First, we propose a two-stream ConvNet
architecture which incorporates spatial and temporal networks. Second, we
demonstrate that a ConvNet trained on multi-frame dense optical flow is able to
achieve very good performance in spite of limited training data. Finally, we
show that multi-task learning, applied to two different action classification
datasets, can be used to increase the amount of training data and improve the
performance on both.
Our architecture is trained and evaluated on the standard video actions
benchmarks of UCF-101 and HMDB-51, where it is competitive with the state of
the art. It also exceeds by a large margin previous attempts to use deep nets
for video classification.
| no_new_dataset | 0.946101 |
1411.3041 | Ramakrishna Vedantam | Ramakrishna Vedantam, C. Lawrence Zitnick, and Devi Parikh | Collecting Image Description Datasets using Crowdsourcing | null | null | null | null | cs.CV | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | We describe our two new datasets with images described by humans. Both the
datasets were collected using Amazon Mechanical Turk, a crowdsourcing platform.
The two datasets contain significantly more descriptions per image than other
existing datasets. One is based on a popular image description dataset called
the UIUC Pascal Sentence Dataset, whereas the other is based on the Abstract
Scenes dataset con- taining images made from clipart objects. In this paper we
describe our interfaces, analyze some properties of and show example
descriptions from our two datasets.
| [
{
"version": "v1",
"created": "Wed, 12 Nov 2014 01:34:46 GMT"
}
] | 2014-11-13T00:00:00 | [
[
"Vedantam",
"Ramakrishna",
""
],
[
"Zitnick",
"C. Lawrence",
""
],
[
"Parikh",
"Devi",
""
]
] | TITLE: Collecting Image Description Datasets using Crowdsourcing
ABSTRACT: We describe our two new datasets with images described by humans. Both the
datasets were collected using Amazon Mechanical Turk, a crowdsourcing platform.
The two datasets contain significantly more descriptions per image than other
existing datasets. One is based on a popular image description dataset called
the UIUC Pascal Sentence Dataset, whereas the other is based on the Abstract
Scenes dataset con- taining images made from clipart objects. In this paper we
describe our interfaces, analyze some properties of and show example
descriptions from our two datasets.
| new_dataset | 0.955361 |
1411.3212 | Francesco Lettich | Francesco Lettich, Salvatore Orlando, Claudio Silvestri and Christian
S. Jensen | Manycore processing of repeated range queries over massive moving
objects observations | null | null | null | null | cs.DB cs.DC cs.DS | http://arxiv.org/licenses/nonexclusive-distrib/1.0/ | The ability to timely process significant amounts of continuously updated
spatial data is mandatory for an increasing number of applications. Parallelism
enables such applications to face this data-intensive challenge and allows the
devised systems to feature low latency and high scalability. In this paper we
focus on a specific data-intensive problem, concerning the repeated processing
of huge amounts of range queries over massive sets of moving objects, where the
spatial extents of queries and objects are continuously modified over time. To
tackle this problem and significantly accelerate query processing we devise a
hybrid CPU/GPU pipeline that compresses data output and save query processing
work. The devised system relies on an ad-hoc spatial index leading to a problem
decomposition that results in a set of independent data-parallel tasks. The
index is based on a point-region quadtree space decomposition and allows to
tackle effectively a broad range of spatial object distributions, even those
very skewed. Also, to deal with the architectural peculiarities and limitations
of the GPUs, we adopt non-trivial GPU data structures that avoid the need of
locked memory accesses and favour coalesced memory accesses, thus enhancing the
overall memory throughput. To the best of our knowledge this is the first work
that exploits GPUs to efficiently solve repeated range queries over massive
sets of continuously moving objects, characterized by highly skewed spatial
distributions. In comparison with state-of-the-art CPU-based implementations,
our method highlights significant speedups in the order of 14x-20x, depending
on the datasets, even when considering very cheap GPUs.
| [
{
"version": "v1",
"created": "Wed, 12 Nov 2014 15:46:39 GMT"
}
] | 2014-11-13T00:00:00 | [
[
"Lettich",
"Francesco",
""
],
[
"Orlando",
"Salvatore",
""
],
[
"Silvestri",
"Claudio",
""
],
[
"Jensen",
"Christian S.",
""
]
] | TITLE: Manycore processing of repeated range queries over massive moving
objects observations
ABSTRACT: The ability to timely process significant amounts of continuously updated
spatial data is mandatory for an increasing number of applications. Parallelism
enables such applications to face this data-intensive challenge and allows the
devised systems to feature low latency and high scalability. In this paper we
focus on a specific data-intensive problem, concerning the repeated processing
of huge amounts of range queries over massive sets of moving objects, where the
spatial extents of queries and objects are continuously modified over time. To
tackle this problem and significantly accelerate query processing we devise a
hybrid CPU/GPU pipeline that compresses data output and save query processing
work. The devised system relies on an ad-hoc spatial index leading to a problem
decomposition that results in a set of independent data-parallel tasks. The
index is based on a point-region quadtree space decomposition and allows to
tackle effectively a broad range of spatial object distributions, even those
very skewed. Also, to deal with the architectural peculiarities and limitations
of the GPUs, we adopt non-trivial GPU data structures that avoid the need of
locked memory accesses and favour coalesced memory accesses, thus enhancing the
overall memory throughput. To the best of our knowledge this is the first work
that exploits GPUs to efficiently solve repeated range queries over massive
sets of continuously moving objects, characterized by highly skewed spatial
distributions. In comparison with state-of-the-art CPU-based implementations,
our method highlights significant speedups in the order of 14x-20x, depending
on the datasets, even when considering very cheap GPUs.
| no_new_dataset | 0.945248 |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.